CPU Tests: Encoding

One of the interesting elements on modern processors is encoding performance. This covers two main areas: encryption/decryption for secure data transfer, and video transcoding from one video format to another.

In the encrypt/decrypt scenario, how data is transferred and by what mechanism is pertinent to on-the-fly encryption of sensitive data - a process by which more modern devices are leaning to for software security.

Video transcoding as a tool to adjust the quality, file size and resolution of a video file has boomed in recent years, such as providing the optimum video for devices before consumption, or for game streamers who are wanting to upload the output from their video camera in real-time. As we move into live 3D video, this task will only get more strenuous, and it turns out that the performance of certain algorithms is a function of the input/output of the content.

HandBrake 1.32: Link

Video transcoding (both encode and decode) is a hot topic in performance metrics as more and more content is being created. First consideration is the standard in which the video is encoded, which can be lossless or lossy, trade performance for file-size, trade quality for file-size, or all of the above can increase encoding rates to help accelerate decoding rates. Alongside Google's favorite codecs, VP9 and AV1, there are others that are prominent: H264, the older codec, is practically everywhere and is designed to be optimized for 1080p video, and HEVC (or H.265) that is aimed to provide the same quality as H264 but at a lower file-size (or better quality for the same size). HEVC is important as 4K is streamed over the air, meaning less bits need to be transferred for the same quality content. There are other codecs coming to market designed for specific use cases all the time.

Handbrake is a favored tool for transcoding, with the later versions using copious amounts of newer APIs to take advantage of co-processors, like GPUs. It is available on Windows via an interface or can be accessed through the command-line, with the latter making our testing easier, with a redirection operator for the console output.

We take the compiled version of this 16-minute YouTube video about Russian CPUs at 1080p30 h264 and convert into three different files: (1) 480p30 ‘Discord’, (2) 720p30 ‘YouTube’, and (3) 4K60 HEVC.

(5-1a) Handbrake 1.3.2, 1080p30 H264 to 480p Discord(5-1b) Handbrake 1.3.2, 1080p30 H264 to 720p YouTube(5-1c) Handbrake 1.3.2, 1080p30 H264 to 4K60 HEVC

Threads wins, although the +30W difference to the desktop processor with two fewer cores can't be overhauled. Almost though.

7-Zip 1900: Link

The first compression benchmark tool we use is the open-source 7-zip, which typically offers good scaling across multiple cores. 7-zip is the compression tool most cited by readers as one they would rather see benchmarks on, and the program includes a built-in benchmark tool for both compression and decompression.

The tool can either be run from inside the software or through the command line. We take the latter route as it is easier to automate, obtain results, and put through our process. The command line flags available offer an option for repeated runs, and the output provides the average automatically through the console. We direct this output into a text file and regex the required values for compression, decompression, and a combined score.

(5-2c) 7-Zip 1900 Combined Score

AES Encoding

Algorithms using AES coding have spread far and wide as a ubiquitous tool for encryption. Again, this is another CPU limited test, and modern CPUs have special AES pathways to accelerate their performance. We often see scaling in both frequency and cores with this benchmark. We use the latest version of TrueCrypt and run its benchmark mode over 1GB of in-DRAM data. Results shown are the GB/s average of encryption and decryption.

(5-3) AES Encoding

 

WinRAR 5.90: Link

For the 2020 test suite, we move to the latest version of WinRAR in our compression test. WinRAR in some quarters is more user friendly that 7-Zip, hence its inclusion. Rather than use a benchmark mode as we did with 7-Zip, here we take a set of files representative of a generic stack

  • 33 video files , each 30 seconds, in 1.37 GB,
  • 2834 smaller website files in 370 folders in 150 MB,
  • 100 Beat Saber music tracks and input files, for 451 MB

This is a mixture of compressible and incompressible formats. The results shown are the time taken to encode the file. Due to DRAM caching, we run the test for 20 minutes times and take the average of the last five runs when the benchmark is in a steady state.

For automation, we use AHK’s internal timing tools from initiating the workload until the window closes signifying the end. This means the results are contained within AHK, with an average of the last 5 results being easy enough to calculate.

(5-4) WinRAR 5.90 Test, 3477 files, 1.96 GB

 

CPU Tests: Rendering CPU Tests: Legacy and Web
Comments Locked

229 Comments

View All Comments

  • ozzuneoj86 - Monday, May 17, 2021 - link

    While it is nice that it supports gen 4, realistically you're just getting SSDs that put out more heat, with more power draw, while gaining performance benefits that are only measurable in benchmarks or very specific situations.

    I'm sure file copy performance is much higher, but how fast do you need that to be? Assuming you're copying to the drive itself or maybe to a Thunderbolt 4 external drive, it is the difference between copying 1TB of data in 2 minutes versus 6 minutes. You can (theoretically) completely fill a $400 2TB SSD in 4 minutes with gen4 vs maybe 12 minutes with Gen 3. If someone needs to do that all the time, then sure there's a difference... but that has to be pretty uncommon.

    For smaller amounts of data, any decent nvme drive is fast enough to make the difference between models almost unnoticeable. For the vast majority of users, even a SATA drive is plenty fast enough to provide a smooth and nearly wait-free experience.
  • mode_13h - Monday, May 17, 2021 - link

    > realistically you're just getting SSDs that put out more heat, with more power draw,
    > while gaining performance benefits that are only measurable in benchmarks
    > or very specific situations.

    Exactly. Thank you.
  • mode_13h - Monday, May 17, 2021 - link

    > Assuming you're copying to the drive itself or maybe to a Thunderbolt 4 external drive

    Oops! TB 4 is limited to PCIe 3.0 x4 speeds! So, it'd be little-to-no help there!
  • Calin - Tuesday, May 18, 2021 - link

    Well, you could copy full blast to an external drive and have plenty of remaining performance to do other storage intensive things - that's assuming your external drives is fast enough to suffocate PCIe 3.0 x4, and your internal drive is faster still.
  • mode_13h - Thursday, May 20, 2021 - link

    > Well, you could copy full blast to an external drive and have plenty of remaining performance

    I'm not one to turn down "free" performance, but PCIe 4 uses significantly more power. In a laptop, that's not a minor point.
  • inighthawki - Monday, May 17, 2021 - link

    Sequential read and write speeds are basically just flexing. Very few people actually ever make significant use of such speeds in a way that saves more than a second or two here or there. Most laptop users are not sitting there copying a terabyte of sequential data over and over again.
  • The_Assimilator - Monday, May 17, 2021 - link

    There is no laptop chassis on the market that can adequately handle the excess of 8W of heat that a PCIe 4.0 NVMe SSD can dissipate.
  • Cooe - Monday, May 17, 2021 - link

    You're not getting those kind of speeds sustained in a laptop without RIDICULOUS thermal throttling. PCIe 4.0 in mobile atm is just a marketing checkmark & nothing more.
  • Calin - Tuesday, May 18, 2021 - link

    It allows faster "races to sleep" for the processor. And, since the Core2 architecture, the winning move was "fast and power hungry processor that does what it must and then goes to a very low power state". This gives you very good burst speed and low average power - as soon as you finish, you can throttle everything down (CPU, caches, SSDs, ...)
  • mode_13h - Thursday, May 20, 2021 - link

    > It allows faster "races to sleep" for the processor.

    Are we still talking about PCIe 4? I don't think it works like that.

    > since the Core2 architecture, the winning move was "fast and power hungry processor that does what it must and then goes to a very low power state".

    No, it's more energy-efficient to run at a slower clock speed. There's a huge difference between the amount of energy used in turbo and non-turbo modes. As it's far bigger than the performance difference, there's no way that going to idle a little sooner is going to make up for it.

Log in

Don't have an account? Sign up now