Total War: Attila

The second strategy game in our benchmark suite, Total War: Attila is the latest game in the Total War franchise. Total War games have traditionally been a mix of CPU and GPU bottlenecks, so it takes a good system on both ends of the equation to do well here. In this case the game comes with a built-in benchmark that plays out over a large area with a fortress in the middle, making it a good GPU stress test.

Total War: Attila - 3840x2160 - Max Quality + Perf Shadows

Total War: Attila - 3840x2160 - Quality + Perf Shadows

Total War: Attila - 2560x1440 - Max Quality + Perf Shadows

Attila is the third win in a row for AMD at 4K. Here the R9 Fury X beats the GTX 980 Ti by 5% at the Max quality setting. However as this benchmark is very forward looking (read: ridiculously GPU intensive), the actual performance at 4K Max isn’t very good. No single GPU card can average 30fps here, and framerates will easily dip below 20fps. Since this is a strategy game we don’t have the same high bar for performance requirements, but sub-30fps still won’t cut it.

In which case we have to either compromise on quality or resolution, and in either case AMD’s lead dissolves. At 4K Quality and 1440p Max, the R9 Fury X trails the GTX 980 Ti by 8% and 3% respectively. And actually the 1440p results are still a good showing, but given AMD’s push for 4K, to lose to the GTX 980 Ti by more at the resolution they favor is a bit embarrassing.

Meanwhile, Atilla has always seemed to love pushing shaders more than anything else, so it comes as no great surprise that this game is a strong showing for the R9 Fury X relative to its predecessor. The performance gains at 4K are a consistent 52%, right at the top-end of our performance expectation window, and a bit smaller (but still impressive) 43% at 1440p.

Far Cry 4 GRID Autosport
Comments Locked

458 Comments

View All Comments

  • chizow - Thursday, July 2, 2015 - link

    What about geometry Ryan? ROPs are often used interchangeably with Geometry/Set-up engine, there is definitely something going on with Fury X at lower resolutions, in instances where SP performance is no problem, it just can't draw/fill pixels fast enough and performs VERY similarly to previous gen or weaker cards (290X/390X and 980). TechReport actually has quite a few theoreticals that show this, where their pixel fill is way behind GM200 and much closer to Hawaii/GM204.
  • extide - Thursday, July 2, 2015 - link

    Yeah my bet is on Geometry. Check out the Synthetics page. It own the Pixel and Texel fillrate tests, but loses on the Tessellation test which has a large dependency on geometry. nVidia has also been historically very strong with geometry.
  • CajunArson - Thursday, July 2, 2015 - link

    Thanks for the review! While the conclusions aren't really any different than all the other reputable review sites on the Interwebs, you were very thorough and brought an interesting perspective to the table too. Better late than never!
  • NikosD - Thursday, July 2, 2015 - link

    You must use the latest nightly build of LAV filters, in order to be able to use the 4K H.264 DXVA decoder of AMD cards.
    All previous builds fall back to SW mode.
  • tynopik - Thursday, July 2, 2015 - link

    "today’s launch of the Fiji GPU"
  • andychow - Thursday, July 2, 2015 - link

    Best review ever. Worth the wait. Get sick more often!
  • tynopik - Thursday, July 2, 2015 - link

    pg 2 - compression taking palce
  • limitedaccess - Thursday, July 2, 2015 - link

    Ryan, regarding Mantle performance back in the R9 285 review (http://www.anandtech.com/show/8460/amd-radeon-r9-2... you wrote that AMD stated the issue with performance regression was that developers had not yet optimized for Tonga's newer architecture. While here you state that the performance regression is due to AMD having not optimized on the driver side. What is the actual case? What is the actual weighting given these three categories? -
    Hardware Driver
    API
    Software/Game

    What I'm wondering is if we make an assumption that upcoming low level APIs will have similar behavior as Mantle what will happen going forward as more GPU architectures are introduced and newer games are introduced? If the onus shifts especially heavily towards the software side it it seems more realistic in practice that developers will have much more narrower scope in which optimize for.

    I'm wondering if Anandtech could possibly look more indept into this issue as to how it pertains to the move towards low level APIs used in the future as it could have large implications in terms of the software/hardware support relationship going forward.
  • Ryan Smith - Thursday, July 2, 2015 - link

    "What is the actual case? What is the actual weighting given these three categories? -"

    Right now the ball appears to be solidly in AMD's court. They are taking responsibility for the poor performance of certain Mantle titles on R9 Fury X.

    As it stands I hesitate to read into this too much for DX12/Vulkan. Those are going to be finalized, widely supported APIs, unlike Mantle which has gone from production to retirement in the span of just over a year.
  • limitedaccess - Thursday, July 2, 2015 - link

    Thanks for the response. I guess we will see more for certain as time moves on.

    My concern is if lower level APIs require more architecture specific optimizations and the burden is shifted to developers in practice that will cause some rather "interesting" implications.

    Also what would be of interest is how much of reviewers test suites will still look at DX11 performance as a possible fallback should this become a possible issue.

Log in

Don't have an account? Sign up now