AMD’s Catalyst 11.1a Hotfix

If the first 2 legs of AMD’s GTX 560 Ti counter-offensive were the 6950 1GB and factory overclocked 6870s, then the final leg of that offensive are the drivers. With barely a month between the launch of the 6900 series and today, AMD has only delivered their original launch drivers for the 6900 series. Meanwhile the 6800 series launched with the Catalyst 10.10 hotfix drivers, which due to how AMD organizes their driver branches are little different than the 10.12 drivers currently posted. So in spite of the nearly 2 month gap between the launches of these two card families, AMD is effectively providing the first real driver update for both.

Launching tomorrow will be the Catalyst 11.1a Hotfix drivers. As far as performance goes they contain the usual mix of game-specific performance improvements, with AMD specifically targeting Call of Duty: Black Ops, BattleForge, Metro 2033, and Aliens vs. Predators performance among other games. Having tested these drivers, overall we’re not seeing any significant performance impact in our benchmark suite, even with games that area on AMD’s list. In fact the only area where we are seeing a major change is with our SmallLuxGPU compute benchmark, which looks to be on the receiving end of some shader compiler optimizations by AMD. SLG performance on the 6900 series is up 25%-30%, providing some validity to AMD’s earlier claims that their VLIW4 shader compiler still has some room to grow as AMD learns how to optimize it like they did with the VLIW5 compiler in the past.

The bigger news is what AMD is doing to their control panel, and what it means to you.

Let me first introduce you to a new section of AMD’s 3D Application Settings control panel called Tessellation. With this new control panel feature AMD is implementing a tessellation factor override in to their drivers, allowing AMD and/or the user to clamp down on games and applications using high tessellation factors. The purpose of this feature is to deal with games such as HAWX 2, which uses very high tessellation factors and offers no tessellation configuration besides turning the feature on and off.

As we’ve already well established, NVIDIA has much better tessellation performance than AMD at high tessellation factors even with the 6900 series. This position leaves AMD on the defensive much of the time (“they’re overpowered” doesn’t have the same ring as “they’re underpowered”), but more so than that games like HAWX 2 are particularly damaging to AMD; they don’t just make AMD’s hardware underperform, but they leave users with only the option to accept poor tessellation performance or to turn tessellation off altogether.

The crux of AMD’s argument – and a point that we agree with – is that tessellation is supposed to be easily scalable. It is in fact the basis of tessellation, that a developer can use it to easily scale up a model based on the available hardware, using a combination of mip-chained displacement maps and an appropriate tessellation factor. The end-game of this scenario would be that a game would use low amounts of tessellation on low-end hardware (e.g. APUs), and large amounts of tessellation on high-end hardware such as GeForce GTX 580s and Radeon HD 6970s. But for that to happen game developers need to take advantage of the flexibility of tessellation by having their games and engines use multiple tessellation factors and displacement maps.

Ultimately games like HAWX2 that do not implement these kinds of controls are not easily scalable. This is the choice of the developer, but in long standing tradition both AMD and NVIDIA will override developer wishes in their drivers when they see fit. In this case AMD believes they are helping their customers by having their drivers cap the tessellation factor in some situations, so that their customers can use tessellation without very high tessellation factors bogging down performance.

And while we agree with AMD’s argument, AMD’s implementation leaves us uneasy. Having this feature available is great, just as is the ability to override v-sync, purposely use poor texture filtering quality for speed purposes, and clamping LOD biases .The bit that makes us uneasy is where the default will lie. AMD offers 3 different “modes”: AMD Optimized, which uses an AMD chosen tessellation factor, user control, and Use Application Settings. AMD intends to make the default “AMD Optimized”, which is to say that in the future all games would use the tessellation factor AMD chooses.

We sincerely believe AMD is doing what they think is best for their users even if they also stand to gain in benchmarks, however we find ourselves in disagreement with their choice. While the actions of games like HAWX2 are unfortunate for users, tessellation is well defined in the DirectX 11 specification. We’re more than willing to entertain creative interpretations of matters like texture filtering where the standard doesn’t dictate a single filtering algorithm, but DX11 doesn’t leave any ambiguity here. As such there’s little room in our opinion for drivers to override a game’s request by default. Drivers should not automatically be substituting a lower tessellation factor on their own – this is a power that should be reserved for a user.


Tessellation in action

Admittedly this is a minefield – modern GPUs are all about taking shortcuts, as these are necessary to get reasonable performance with the kind of complexity modern games are shooting for. But it’s our opinion that there’s no better time to take a stand than before an optimization like this is implemented, as once it’s done then it’s almost impossible to change course, or even to have a meaningful discourse about the issue.

At this time AMD has not defined any tessellation factors in their profiles, and as a result the AMD Optimized setting is no different than the Use Application Settings setting. At some point this will change. We would like to see AMD build this feature in to their drivers and leave the choice up to the user, but only time will tell how they proceed.

On that note, tessellation factors are not the only minefield AMD is dabbling in. With the Catalyst 10.10 drivers AMD began playing with their texture filtering quality at different levels. Previously at High Quality (previously known as Catalyst AI Off) AMD would disable all optimizations, while at the default setting of Quality (Catalyst AI Standard) AMD would use a small set of optimizations that had little-to-any impact on image quality, and at Performance (Catalyst AI Advanced) they would use a number of optimizations to improve performance. Texture filtering optimizations are nothing new (having been around practically as long as the 3D accelerator), but in a 2-player market any changes will make a wave.

In the case of AMD’s optimizations, for Quality mode they picked a new set of optimizations that marginally improved performance but at the same time marginally changed the resulting image quality. Many tomes about the issue have already been written, and there’s very little I believe we can add to the subject – meaningful discourse is difficult to have when you believe there’s room for optimizations while at the same time believing there is a point where one can go too far.


AMD Radeon HD 6870, Catalyst 10.10e

In any case while we have found very little to add to the subject, this has not been the case elsewhere on the internet. As such after 3 months AMD is largely reverting their changes to texture filtering, and will be returning it to similar quality levels as what we saw with the Catalyst 10.9 drivers – which is to say they’re now once again shooting for a level of texture filtering quality similar to NVIDIA.

As we have little to add beyond this, here are AMD’s full notes on the matter:

The Quality setting has now been improved to match the HQ setting in all respects except for one – it enables an optimization that limits trilinear anisotropic filtering to areas surrounding texture mipmap level transitions, while doing bilinear anisotropic filtering elsewhere.  Sometimes referred to as “brilinear” filtering, it offers a way to improve filtering performance without visibly affecting image quality.  It has no impact on texture sharpness or shimmering, and this can be verified by comparing it visually with the High Quality setting.

We continue to recommend the Quality setting as the best one to use for competitive testing for the following reasons:

  • It should be visually indistinguishable from the High Quality setting for real textures (with the exception of special test patterns using colored mip levels)
  • Visual quality should now be equal to the default setting used on HD 5800 series GPUs with Catalyst 10.9 and earlier drivers, or better when used on HD 6800/6900 series GPUs due to other hardware filtering improvements
  • It matches the default texture filtering quality setting currently implemented on our competitor’s GPUs, which make use of the same trilinear filtering optimization
Meet The Radeon HD 6950 1GB and XFX Radeon HD 6870 Black Edition The Test & Gaming Performance
Comments Locked

111 Comments

View All Comments

  • Ryan Smith - Wednesday, January 26, 2011 - link

    Indeed CPU bottlenecking is a concern, and we always try to remove it as much as possible. Replacing the CPU means throwing out our entire body of work, so as important as it is to avoid being CPU bottlenecked, we can't do it frequently.

    The issue for us right now is that SNB-E isn't due until late this year, and that's the obvious upgrade path for our GPU testbed since SNB has a limited amount of PCIe bandwidth.
  • 7upMan - Wednesday, January 26, 2011 - link

    RYAN: Hi Ryan, while I usually find AnandTech articles quite entertaining and informative, I always wonder why the f*ck professional editors won't get it into their head to test 2GB cards in areas where they belong to. Meaning: a 2GB vs. 1 GB card test should be about graphically overly intensive games and game mods, like the Half-Life 2 Fake Factory mod, or the STALKER Complete mod (Oblivion too has such mods). There are a number of other mods that put massive numbers of huge textures into the graphics RAM, and I think they should be the ones you need to test the cards with. After all, you can't expect games that were written with 1GB VRAM in mind to utilize the full power of double VRAM.

    So please, please run some tests with the above mentioned mods. Thanks in advance.
  • IceDread - Wednesday, January 26, 2011 - link

    Good review

    As a sidenote, sort of fun to see my one year old card 5970 is still the best when looking at single cards.
  • IceDread - Wednesday, January 26, 2011 - link

    There was not much talk about SLI and crossfire btw, there the value in AMD is higher today day with a fair bit than with SLi solutions comparing AMD 6xxx series with new nvidia cards.
  • Ryan Smith - Wednesday, January 26, 2011 - link

    We had to leave our SLI & CF for this article because of all the driver changes from both parties - as it stands I need to rerun most of those numbers.

    You'll be seeing a lot more on SLI and CF in a week or two; we have a trio of 580s and 6970s in house for some tri-SLI/CF testing.
  • Scootiep7 - Wednesday, January 26, 2011 - link

    Not going to add much to the discourse at this point. I just want to say that I really liked this article and thank you for your time and due diligence in writing it.
  • blackshard - Wednesday, January 26, 2011 - link

    Why they are so different from previous articles? NVIDIA numbers have grown about three times and some AMD numbers are grown too.

    Previous article about HD6950 and HD6970 showed this:
    http://www.anandtech.com/show/4061/amds-radeon-hd-...
  • Ryan Smith - Wednesday, January 26, 2011 - link

    In the GTX 560 Ti article I explained what was going on.

    http://www.anandtech.com/show/4135/nvidias-geforce...

    "Small Lux GPU is the other test in our suite where NVIDIA’s drivers significantly revised our numbers. Where this test previously favored raw theoretical performance, giving the vector-based Radeons an advantage, NVIDIA has now shot well ahead. Given the rough state of both AMD and NVIDIA’s OpenCL drivers, we’re attributing this to bug fixes or possibly enhancements in NVIDIA’s OpenCL driver, with the former seeming particularly likely. However NVIDIA is not alone when it comes to driver fixes, and AMD has seem a similar uptick against the newly released 6900 series. It’s not nearly the leap NVIDIA saw, but it’s good for around 25%-30% more rays/second under SLG. This appears to be accountable to further refinement of AMD’s VLIW4 shader compiler, which as we have previously mentioned stands to gain a good deal of performance as AMD works on optimizing it."
  • blackshard - Wednesday, January 26, 2011 - link

    Ok, got it. I have not read the gtx560 review. Thanks ;)
  • ibudic1 - Wednesday, January 26, 2011 - link

    Don't forget that you can unlock 6950 to 6970, at which point Nvidia is just NOT competitive.

    http://www.techpowerup.com/137140/AMD-Radeon-HD-69...

Log in

Don't have an account? Sign up now