CPU Performance: New Tests!

As part of our ever on-going march towards a better rounded view of the performance of these processors, we have a few new tests for you that we’ve been cooking in the lab. Some of these new benchmarks provide obvious talking points, others are just a bit of fun. Most of them are so new we’ve only run them on a few processors so far. It will be interesting to hear your feedback!

As far as this review goes, we need to perform regression testing of our new benchmarks on the older hardware, and so these resulst are here more for completeness.

NAMD ApoA1

One frequent request over the years has been for some form of molecular dynamics simulation. Molecular dynamics forms the basis of a lot of computational biology and chemistry when modeling specific molecules, enabling researchers to find low energy configurations or potential active binding sites, especially when looking at larger proteins. We’re using the NAMD software here, or Nanoscale Molecular Dynamics, often cited for its parallel efficiency. Unfortunately the version we’re using is limited to 64 threads on Windows, but we can still use it to analyze our processors. We’re simulating the ApoA1 protein for 10 minutes, and reporting back the ‘nanoseconds per day’ that our processor can simulate. Molecular dynamics is so complex that yes, you can spend a day simply calculating a nanosecond of molecular movement.

NAMD 2.31 Molecular Dynamics (ApoA1)

 

Crysis CPU Render

One of the most oft used memes in computer gaming is ‘Can It Run Crysis?’. The original 2007 game, built in the Crytek engine by Crytek, was heralded as a computationally complex title for the hardware at the time and several years after, suggesting that a user needed graphics hardware from the future in order to run it. Fast forward over a decade, and the game runs fairly easily on modern GPUs, but we can also apply the same concept to pure CPU rendering – can the CPU render Crysis? Since 64 core processors entered the market, one can dream. We built a benchmark to see whether the hardware can.

For this test, we’re running Crysis’ own GPU benchmark, but in CPU render mode. This is a 2000 frame test, which we run over a series of resolutions from 800x600 up to 1920x1080. Here we have the 1920x1080 results, with the rest being in our Benchmark database.

Crysis CPU Render: (6) 1920x1080

Dwarf Fortress

Another long standing request for our benchmark suite has been Dwarf Fortress, a popular management/roguelike indie video game, first launched in 2006. Emulating the ASCII interfaces of old, this title is a rather complex beast, which can generate environments subject to millennia of rule, famous faces, peasants, and key historical figures and events. The further you get into the game, depending on the size of the world, the slower it becomes.

DFMark is a benchmark built by vorsgren on the Bay12Forums that gives two different modes built on DFHack: world generation and embark. These tests can be configured, but range anywhere from 3 minutes to several hours. I’ve barely scratched the surface here, but after analyzing the test, we ended up going for three different world generation sizes.

Dwarf Fortress (Small) 65x65 World, 250 YearsDwarf Fortress (Medium) 129x129 World, 550 YearsDwarf Fortress (Big) 257x257 World, 550 Years

AI Benchmark

One of the longest time requests we’ve had for our benchmark suite is AI-related benchmark, and the folks over at ETH have moved their popular AI Benchmark from mobile over PC. Using Intel’s MKL and Tensorflow 2.1.0, we use version 0.1.2 of the benchmark which tests both training and inference over a variety of different models. You can read the full scope of the benchmark here.

AI Benchmark (ETH) CombinedAI Benchmark (ETH) InferenceAI Benchmark (ETH) Training

 

V-Ray

We have a couple of renderers and ray tracers in our suite already, however V-Ray’s benchmark came through for a requested benchmark enough for us to roll it into our suite. We run the standard standalone benchmark application, but in an automated fashion to pull out the result in the form of kilosamples/second. We run the test six times and take an average of the valid results.

V-Ray Renderer

CPU Performance: Synthetic, Web and Legacy Tests Gaming: World of Tanks enCore
Comments Locked

114 Comments

View All Comments

  • vortmax2 - Monday, May 18, 2020 - link

    Anyone know why the 3300X is at the top of the Digicortex 1.20 bench?
  • gouthamravee - Monday, May 18, 2020 - link

    I'm guessing here, but the 3300X has all its cores on a single CCX and if Digicortex is one of those benches that's highly dependent on latency that could explain why the 3300X is at the top of the list here.

    I checked the previous 3300x article and it seems to be the same story there.
  • wolfesteinabhi - Monday, May 18, 2020 - link

    Thanks for a great Article Ian and AT.

    the main problem with mid/lower range CPU (review) like this Ryzen 3600/X and even i5/i3's is that their reviews are almost always focused on "Gaming" (for some reason everything budget oriented is just gaming) ... no one talks about AI workloads or MATLABs, Tensorflows,etc many people and developers dont want to shell out monies for 2080Ti and Ryzen 9 3950X or even TR's .... they have to make do with lower end or say "reasonable" CPU's ... and products like these Ryzen 5 that makes sensible choice in this segment ... a developer/learner on budget.

    a lot of people would appreciate if there are some more pages dedicated to such development workflows (AI,Tensor,compile, etc) even for such mid range CPU's.
  • DanNeely - Monday, May 18, 2020 - link

    Ian periodically tweets requests for scriptable benchmarks for those categories and for anyone with connections at commercial vendors in those spaces who can provide evaluation licenses for commercial products. He's gotten minimal uptake on the former and doesn't have time to learn enough about $industry to create a reasonable benchmark from scratch using their FOSS tools. On the commercial side, the various engineering software companies don't care about reviews from sites like this one and their PR contacts can't/won't give out licenses.
  • webdoctors - Monday, May 18, 2020 - link

    Because office tasks don't require any computation, and gaming is what's most mainstream that actually requires computation.

    Scientific stuff like MATLAB, Folding@Home needs computation but if that's useful you'd just buy the higher end parts. Price diff between 3600x and 3700x (6 vs 8core) is $100, $200 vs $300 at retail prices. For someone working, $100 is nothing for improving your commercial or academic output. These are parts you use for 5+ years.

    I agree a TR doesnt make sense if you can get the consumer version like a 3800x much cheaper.
  • Impetuous - Monday, May 18, 2020 - link

    Logged in to second this. I think a lot of students and professionals like me who do research on-the-side (and are on pretty tight Grants/allowances) would appreciate a MATLAB benchmark. This looks like a great option for a grad student workstation!
  • brucethemoose - Monday, May 18, 2020 - link

    I think one MKL TF benchmark is enough, as you'd have to be crazy to buy a 3600 over a cheap GPU for AI training training. If money is that tight, you're probably not buying a new system and/or using Google Colab.

    +1 for more compilation benchmarking. I'd like a Python benchmark too, if theres any demand for such a thing.
  • PeachNCream - Monday, May 18, 2020 - link

    A lot of people don't have money to throw away at hardware, moreso now than ever before so we are going to make older equipment work for longer or buy less compute at a lower price. It's important to get hardware out of its comfort zone because these general purpose processors will be used in all sorts of ways beyond a narrow set of games and unzipping a huge archive file. After all, if you want to play games, buying as much GPU as you can afford and then feeding it enough power solves the problem for the most part. That answer has been the case for years so we really don't need more text and time spent on telling us that. Say it once for each new generation and then get to reviewing hardware more relevant to how people actually use their computers.
  • jabber - Tuesday, May 19, 2020 - link

    Plus most of us don't upgrade hardware as much as we used to. back in the day (single core days) I was upgrading my CPU every 6-8 months. Each upgrade pushed the graphics from 28FPS to 32FPS to 36FPS which made a difference. Now with modest setups pushing past 60FPS...why bother. I upgrade my CPU every 6 years or so now.
  • wolfesteinabhi - Tuesday, May 19, 2020 - link

    as i said in one of the replies below... maybe TF is not a good example ..but its not like it will be purely on a CPU for TF work, but some benchmark around it ...and similar other work/development related tasks.

    Most of us have to depend on these gaming only benchmarks to guesstimate how good/bad a cpu will be for dev work. maybe a fewer core cpu might have been better with extra cache and extra clocks or vice versa ... but almost no reviews tell that kind of story for mid/low range CPU's.... having said that..i dont expect that kind of analysis from dual cores and such CPU ..but higherup there are a lot of CPU that can be made to do a lot of good job even beyond gaming (even if it needs to pair up with some GPU)

Log in

Don't have an account? Sign up now