DIRT 2, Mass Effect 2, Wolfenstein, & Compute Performance

With DIRT 2 our benchmarks once again swing back in favor of NVIDIA. At 1920 the GTX 560-448 is 14% of the 6970, never mind the 6950. The advantage over the GTX 560 Ti is even greater, with the GTX 560-448 coming in at nearly 20% faster. Throw on Zotac’s overclock and now the GTX 560-448 is as fast as the GTX 570. If you still don’t believe the GTX 560-448 is really a slower GTX 570, then this should be rather convincing proof.

And we swing back the other way with Mass Effect 2. The GTX 560-448 falls short of the 6950 by 5%, and compared to the GTX 560 Ti it only leads by 6%, highlighting the fact that GF110 and GF104 can be very different architectures at times. Furthermore as Mass Effect 2 responds well to memory bandwidth, even with Zotac’s overclock the GTX 560-448 still comes up a bit short of the GTX 570. This is one of the few cases where the GTX 570 has a distinct benefit even with its limited advantage on paper.

On our final game, the OpenGL based Wolfenstein, AMD and NVIDIA once more swap places with the 6950 squeaking out a small lead over the GTX 560-448. As with other cases where the GTX 560-448 fails to gain on AMD’s cards, it also fails to differentiate itself much from the GTX 560 Ti, this time only taking a meager 4% lead.

Our final performance benchmark is a quick look at compute performance using the Civilization V texture decompression test. In spite of this not benefiting from NVIDIA’s multithreading optimizations, NVIDIA’s drivers and architecture still give the GTX 560-448 a solid lead over AMD’s cards. The benefit over the GTX 560 Ti is also more pronounced at 16%. Interestingly in spite of the GTX 570’s compute performance advantage on paper, the GTX 560-448 can tie it at stock clocks and overtake it with Zotac’s overclock, showcasing that this benchmark isn’t exclusively tied to compute performance.

HAWX, Civ V, Battlefield BC2, & STALKER Power, Temperature, & Noise
Comments Locked

80 Comments

View All Comments

  • Transmitthis14 - Tuesday, November 29, 2011 - link

    Many would ignore "Value" like that.
    I use Premier and Pshop, (cuda Acceleration) and also like the option of Physx with my Games when coded for.
    So A 6950 then becomes a none option
  • Exodite - Wednesday, November 30, 2011 - link

    Fringe benefits like those swing both ways though.

    I need three simultaneous display outputs, which render Nvidia options redundant in much the same way.

    Looking strictly at price/performance, which is the metric most consumers would use in this case, the 2GB 6950 is indeed an excellent option.
  • grammaton.feather - Wednesday, November 30, 2011 - link

    Oh really?

    http://www.overclock.net/t/909086/6950-shaders-unl...

    I also agree with the CUDA argument. I have used Nvidia for years because of stereoscopic 3d support, more recently CUDA and PhysX and because I don't have driver crashes with Nvidia.

    ATI may be better value for money at the cheap end of the spectrum, so long as u don't mind driver crashes.

    A friend who used his high end PC for business had 170 plus ATI driver crashes logged by windows. I chose an Nvidia GPU for him and the crashing stopped.
  • rum - Wednesday, November 30, 2011 - link

    I have had nothing but ATI cards since the 3870, and have NOT experienced these driver crashes you are speaking of.

    I use my computers for gaming and computer programming, and sometimes I do actually stress my system. :)

    I am in the process of evaluating my video needs, as I do need a new video card for a new PC. I won't blindly go with ATI, I will EVALUATE all aspects and get the best card for the MONEY.

    I don't need cutting edge, neither do most people, and we do evaluate things on most bang for the buck.
  • VoraciousGorak - Wednesday, November 30, 2011 - link

    Funny, since I Fold on a GTX275, GTX460, and 9800GT on two different systems, and pretty often have NVIDIA's drivers lose their minds when finishing a WU, shutting down the system, or while simply processing a WU. I've tried multiple WHQL and beta drivers and settled with these being the least of my problems (the worst being total system unresponsiveness while Folding and a hard reset after a client stopped.) However, when I folded on my 6950 (granted, different client, but I Folded on it with the Beta client too) it never had a driver freakout, screen blank, or even a flicker when turning on or shutting off WUs.

    It seems to me that people that have major driver issues with AMD (not ATI anymore, by the way) cards are outliers, relative novices, and/or are using very outdated drivers or are recounting stories back from when the ATI 9700 series were new. I'm probably an outlier as well with my NVIDIA gripes, but characterizing a major modern GPU manufacturer with constant widespread driver issues or stating that any GPU is guaranteed to have problems is just silly.
  • silverblue - Thursday, December 1, 2011 - link

    Oddly, those people who slate AMD are the ones who never ever seem to have an NVIDIA crash, or at least won't admit to one.
  • Gorghor - Thursday, December 1, 2011 - link

    I know this has no technical relevance, but I thought I'd share since my first reaction when I saw the article title, was "did I make a mistake buying an ATI6950?"

    For some reason I have been sticking to nVidia cards since the days of the Geforce 2. I made an exception with the ATI 9800 and although the performance was good back then, I had many problems with the drivers and really didn't approve of the catalyst control center.

    When I upgraded my mainboard to the Sandy-bridge generation, I left my "old" GTX285 in the system and had several daily crashes related to the nVidia drivers (even after several months and driver updates).

    About a month ago I decided it was time for me to upgrade the graphics. I wanted to stick to nVidia albeit the problems with my GTX285, but the price/performance ratio was just unacceptable when compared to AMD cards. I chose the ATI6950 2GB considering the same money would barely get me a GTX560.

    I have to say I'm impressed. The performance is truly exceptional, even with no shader unlock, the catalyst interface has really matured over the years and my crashes are completely gone. So as far as I'm concerned, blindly going for nVidia is not an option anymore.
  • Golgatha - Tuesday, November 29, 2011 - link

    what I take away from this review is the 560-448 is pretty much on par with a stock 570, but it will only be manufactured for roughly 2 months, and then be discontinued. Why would any gamer buy this card when they can get a superior card within $20 of this one and pretty much be guaranteed availability if they want to run SLI? Also, what's with the power connectors on the rear of the card. All of these long cards should situate those connectors on the top of the card.
  • jigglywiggly - Tuesday, November 29, 2011 - link

    I'd avoid it, you will probably be screwd with drivers in the long run.
  • Samus - Tuesday, November 29, 2011 - link

    As far as drivers and modern games go, people with ATI hardware have been having substantially more problems in Battlefield 3. Then the issue of microstutter arises which barely affects nVidia hardware, for what can only be presumed to be a ATI driver problem. Lastly, many people have experienced PowerPlay problems with Radeon 4000-5000 series cards running newer drivers due to clockspeed changes in-game, sometimes causing BSOD's, but always causing performance problems.

    nVidia has been, and always will be, the kind of drivers. The days of Mach64 still plague ATI's driver developement team. nVidia invented unified drivers and changed the game for the industry. They update their SLI profiles more than twice as often as ATI does for Crossfire, which alone shows their loyalty to their SLI users.

    ATI has the hardware engineering team. That's clear. They can produce faster cards that use less power at lower prices. But you will NOT get superior drivers by a longshot.

Log in

Don't have an account? Sign up now