Gaming: Civilization 6 (DX12)

Originally penned by Sid Meier and his team, the Civ series of turn-based strategy games are a cult classic, and many an excuse for an all-nighter trying to get Gandhi to declare war on you due to an integer overflow. Truth be told I never actually played the first version, but every edition from the second to the sixth, including the fourth as voiced by the late Leonard Nimoy, it a game that is easy to pick up, but hard to master.

Benchmarking Civilization has always been somewhat of an oxymoron – for a turn based strategy game, the frame rate is not necessarily the important thing here and even in the right mood, something as low as 5 frames per second can be enough. With Civilization 6 however, Firaxis went hardcore on visual fidelity, trying to pull you into the game. As a result, Civilization can taxing on graphics and CPUs as we crank up the details, especially in DirectX 12.

Perhaps a more poignant benchmark would be during the late game, when in the older versions of Civilization it could take 20 minutes to cycle around the AI players before the human regained control. The new version of Civilization has an integrated ‘AI Benchmark’, although it is not currently part of our benchmark portfolio yet, due to technical reasons which we are trying to solve. Instead, we run the graphics test, which provides an example of a mid-game setup at our settings.

AnandTech CPU Gaming 2019 Game List
Game Genre Release Date API IGP Low Med High
Civilization VI RTS Oct
2016
DX12 1080p
Ultra
4K
Ultra
8K
Ultra
16K
Low

All of our benchmark results can also be found in our benchmark engine, Bench.

Civilization 6 IGP Low Medium High
Average FPS
95th Percentile

Continuing the theme we’ve seen thus far, Civilization 6 is another game where the 9900K does provide some benefits, but not under all circumstances. The game is not particularly GPU-intensive to begin with, so at just 4K Ultra we’re still not entirely GPU limited; but past a Ryzen 7 2700X or so, all the CPUs start running together. We have to drop to 1080p Ultra to really pull the CPUs off of the dogpile, at which point the 9900K comes out in the lead.

This is another game that doesn’t seem to care about core counts so much as it does frequencies. So the 9900K has the strongest position here, while the 9700K brings up second place. But neither are very far from the 8700K, with Intel’s latest coming in at just 12% faster than their former flagship even at these CPU benchmarking sympathetic settings.

Curiously we also see the 9900K fall behind the 9700K at 4K and higher. The difference is easily close enough to be noise, but it might be a very slight impact of the lower-tier chips not having to share their cores with hyper-threading.

Gaming: Shadow of War Gaming: Ashes Classic (DX12)
Comments Locked

274 Comments

View All Comments

  • Targon - Friday, October 19, 2018 - link

    TSMC will do the job for AMD, and in March/April, we should be seeing AMD release the 3700X and/or 3800X that will be hitting the same clock speeds as the 9900k, but with a better IPC.
  • BurntMyBacon - Friday, October 19, 2018 - link

    I am certainly happy that AMD regained competitiveness. I grabbed an R7 1700X early on for thread heavy tasks while retaining use of my i7-6700K in a gaming PC. That said, I can't credit them with everything good that comes out of Intel. To say that Intel would not have released an 8 core processor without AMD is probably inaccurate. They haven't released a new architecture since Skylake and they are still on a 14nm class process. They had to come up with some reason for customers to buy new processors rather than sit on older models. Clock speeds kinda worked for Kaby Lake, but they need more for Coffee Lake. Small, fixed function add-ons that only affect a small portion of the market probably weren't enough. A six core chip on the mainstream platform may have been inevitable. Going yet another round without a major architecture update or new process node, it is entirely possible that the 8-core processor on the mainstream platform was also inevitable. I give AMD credit for speeding up the release schedule, though.

    As to claims that the GF manufacturing is responsible for the entire 1GHz+ frequency deficit, that is only partially true. It is very likely that some inferior characteristics of the node are reducing the potential maximum frequency achievable. However, much of the limitations on frequency also depends on how AMD layed out the nodes. More capacitance on a node makes switching slower. More logic between flip-flops require more switches to resolve before the final result is presented to the flip-flops. There is a trade-off between the number of buffers you can put on a transmission line as reducing input to output capacitance ratios will speed up individual switch speeds, but they will also increase the number of switches that need to occur. Adding more flip-flops increases the depth of the pipeline (think pentium 4) and increases the penalty for branch misses as well as making clock distribution more complicated. These are just a few of the most basic design considerations that can affect maximum attainable frequency that AMD can control.

    Consequently, there is no guarantee that AMD will be able to match Intel's clock speeds even on TSMC's 7nm process. Also, given that AMD's current IPC is more similar to Haswell and still behind Skylake, it is not certain that they next processors will have better IPC than Intel either. I very much hope one or the other ends up true, but unrealistic expectations won't help the situation. I'd rather be pleasantly surprised than disappointed. As such, I expect that AMD will remain competitive. I expect that they will close the gaming performance gap until Intel releases a new architecture. I expect that regardless of how AMD's 7nm processors stack against Intel's best performance-wise, I expect that AMD likely bring better value at least until Intel gets their 10nm node fully online.
  • Spunjji - Monday, October 22, 2018 - link

    "To say that Intel would not have released an 8 core processor without AMD is probably inaccurate."
    It's technically inaccurate to say they would have never made any kind of 8-core processor, sure, but nobody's saying that. That's a straw man. What they are saying is that Intel showed no signs whatsoever of being willing to do it until Ryzen landed at their doorstep.

    To be clear, the evidence is years of Intel making physically smaller and smaller quad-core chips for the mainstream market and pocketing the profit margins, followed by a sudden and hastily-rescheduled grab for the "HEDT" desktop market the second Ryzen came out, followed by a rapid succession of "new" CPU lines with ever-increasing core counts.

    You're also wrong about AMD's IPC, which is very clearly ahead of Haswell. The evidence is here in this very article where you can see the difference in performance between AMD and Intel is mostly a function of the clock speeds they attain. Ryzen was already above Haswell for the 1000 series (more like Broadwell) and the 2000 series brought surprisingly significant steps.
  • khanikun - Tuesday, October 23, 2018 - link

    " What they are saying is that Intel showed no signs whatsoever of being willing to do it until Ryzen landed at their doorstep."

    Intel released an 8 core what? 3 years before Ryzen. Sure, it was one of their super expensive Extreme procs, but they still did it. They were slowly ramping up cores for the HEDT market, while slowly bringing them to more normal consumer prices. 3 years before Ryzen, you could get a 6 core i7 for $400 or less. A year before that it was like $550-600. A 1-2 years before that, a 6 core would be $1000+. 8 cores were slowly coming.

    What Ryzen did was speed up Intel's timeframe. They would have came and came at a price point that normal consumers would be purchasing them. If I had to guess, we're probably 2-3 years ahead of what Intel probably wanted to do.

    Now would Ryzen exist, if not for Intel? Core for core, AMD has nothing that can compete with Intel. So...ramp up the core count. We really don't see Intel going away from a unified die design, so that's the best way AMD has to fight Intel. I'm personally surprised AMD didn't push their MCM design years ago. Maybe they didn't want to cannibalize Opteron sales, bad yields, I don't know. Must have been some reason.
  • Cooe - Friday, October 19, 2018 - link

    Rofl, delusional poster is delusional. And anyone who bought a 2700X sure as shit doesn't need to do anything to "defend their purchase" to themselves hahaha.
  • evernessince - Saturday, October 20, 2018 - link

    Got on my level newb. The 9900K is a pittance compared to my Xeon 8176. I hope you realized that was sarcasm and how stupid it is to put people down for wanting value.
  • JoeyJoJo123 - Friday, October 19, 2018 - link

    >I think far too much emphasis has been placed on 'value'.

    Then buy the most expensive thing. There's no real need to read reviews at that point either. You just want the best, money is no object to you, and you don't care, cool. Just go down the line and put the most expensive part for each part of the PC build as you browse through Newegg/Amazon/whatever, and you'll have the best of the best.

    For everyone else, where money is a fixed and limited resource, reading reviews MATTERS because we can't afford to buy into something that doesn't perform adequately for the cost investment.

    So yes, Anandtech, keep making reviews to be value-oriented. The fools will be departed with their money either way, value-oriented review or not.
  • Arbie - Friday, October 19, 2018 - link

    They'll be parted, yes - and we can hope for departed.
  • GreenReaper - Saturday, October 20, 2018 - link

    Don't be *too* harsh. They're paying the premium to cover lower-level chips which may be barely making back the cost of manufacturing, thus making them a good deal. (Of course, that also helps preserve the monopoly/duopoly by making it harder for others to break in...)
  • Spunjji - Monday, October 22, 2018 - link

    Yeah, to be honest the negatives of idiots buying overpriced "prestige" products tend to outweigh the "trickle down" positives for everyone else. See the product history of nVidia for the past 5 years for reference :/

Log in

Don't have an account? Sign up now