Final Words

Traditionally dual-GPU cards have been a mixed bag. More often than not they have to sacrifice a significant amount of single-GPU performance in order to put two GPUs on a single card, and in the rare occasions where that tradeoff doesn’t happen there’s some other tradeoff such as a loud cooler or immense power consumption. NVIDIA told us that they could break this tradition and put two full GTX 680s on a single card, and that they could do that while making it quieter and less power consuming than a dual video card SLI setup. After going through our benchmarking process we can safely say that NVIDIA has met their goals.

From a gaming performance perspective we haven’t seen a dual-GPU card reach the performance of a pair of high-end cards in SLI/CF since the Radeon HD 4870X2 in 2008, so it’s quite refreshing to see someone get so close again 4 years later. The GTX 690 doesn’t quite reach the performance of the GTX 680 SLI, but it’s very, very close. Based on our benchmarks we’re looking at 95% of the performance of the GTX 680 SLI at 5760x1200 and 96% of the performance at 2560x1600. These are measurable differences, but only just. For all practical purposes the GTX 690 is a single card GTX 680 SLI – a single card GTX 680 SLI that consumes noticeably less power under load and is at least marginally quieter too.

With that said, this would typically be the part of the review where we would inject a well-placed recap of the potential downsides of multi-GPU technology; but in this case there’s really no need. Unlike the GTX 590 and unlike the GTX 295 NVIDIA is not making a performance tradeoff here compared to their single-GPU flagship card. When SLI works the GTX 690 is the fastest card out there, and when SLI doesn’t work the GTX 690 is still the fastest card out there. For the first time in a long time using a dual-GPU card doesn’t mean sacrificing single-GPU performance, and that’s a game changer.

At this point in time NVIDIA offers two different but compelling solutions for ultra-enthusiast performance; the GTX 690 and GTX 680 SLI, and they complement each other well. For most situations the GTX 690 is going to be the way to go thanks to its lower power consumption and lower noise levels, but for cases that need fully exhausting video cards the GTX 680 SLI can offer the same gaming performance at the same price. Unfortunately we’re going to have to put AMD out of the running here; as we’ve seen in games like Crysis and Metro the 7970 in Crossfire has a great deal of potential, but as it stands Crossfire is simply too broken overall to recommend.

The only real question I suppose is simply this: is the GTX 690 worthy of its $999 price tag? I don’t believe there’s any argument to be had with respect to whether the GTX 690 is worth getting over the GTX 680 SLI, as we’ve clearly answered that above. As a $999 card it doesn’t double the performance of the $499 GTX 680, but SLI has never offered quite that much of a performance boost. However at the same time SLI has almost always been good enough to justify the cost of another GPU if you must have performance better than what the fastest single GPU can provide, and this is one of those times.

Is $999 expensive? Absolutely. Is it worth it? If you’re gaming at 2560x1600 or 5760x1200, the GTX 690 is at least worth the consideration. You can certainly get by on less, but if you want 60fps or better and you want it with the same kind of ultra high quality single GPU cards can already deliver at 1920x1080, then you can’t do any better than the GTX 690.

Wrapping things up, there is one question left I feel like we still don’t have a good answer to: how much RAM a $999 card should have. NVIDIA went with a true equal for the GTX 680 SLI, right down to the 2GB of VRAM per GPU. Looking back at what happened to the Radeon HD 5970 and its 1GB of VRAM per GPU – we can’t even run our 5760x1200 benchmarks on it, let alone a couple of 2560x1600 benchmarks – I’m left uneasy. None of our benchmarks today seem to require more than 2GB of VRAM, but that much VRAM has been common in high-end cards since late 2010; the day will come when 2GB isn’t enough, and I'm left to wonder when. A GTX 690 with 4GB of VRAM per GPU would be practically future-proof, but with 2GB of VRAM NVIDIA is going to be cutting it close.

Overclocked: Gaming Performance
Comments Locked

200 Comments

View All Comments

  • bobsmith1492 - Thursday, May 3, 2012 - link

    It's not that rare; I got a fairly inexpensive 24" 1920x1200 HP monitor from Newegg a year ago. There weren't many options but it was there and it's great.
  • a5cent - Thursday, May 3, 2012 - link

    You are right that the average Joe doesn't have a 1920x1200 monitor, but this is an enthusiast web-site! Not a single enthusiast I know owns a 1080 display. 1920x1200 monitors aren't hard to find, but you will need to spend a tad more.
  • CeriseCogburn - Saturday, May 5, 2012 - link

    Nope, 242 vs 16 is availability, you lose miserably. You all didn't suddenly have one along with your "friends" you suddenly acquired and have memorized their monitor sizes instantly as well.
    ROFL - the lies are innumerable at this point.
  • UltraTech79 - Thursday, May 3, 2012 - link

    They make up about 10% stock. I wouldn't call that very rare. Newegg and other places have a couple dozen+ to choose from.

    Maybe YOU dont buy very much.
  • CeriseCogburn - Tuesday, May 8, 2012 - link

    Closer to 5% than it is to 10%, and they cost a lot more for all the moaning penny pinchers who've suddenly become flush.
  • Digimonkey - Thursday, May 3, 2012 - link

    It's either 1920x1200 @ 60hz, or 1920x1080 @ 120hz. I prefer smoother gameplay over 120 pixels. Also I know quite a few gamers that like using their TV for their PC gaming, so this would also be limited to 1080p.
  • CeriseCogburn - Friday, May 4, 2012 - link

    No one here is limited, they all said, so no one uses their big screens, they all want it @ 1200P now because amd loses not so badly there...
    ROFL
  • Dracusis - Thursday, May 3, 2012 - link

    I'm be more worried about AMD's performance going down in certain games due to Crossfire than something as trival as this. As a 4870X2 owner I can tell you this is not at all uncommon for AMD. I still have to disable 1 GPU in most games, including BF3, because AMDs drivers for any card more than 12 months old are just terrible. As you can see even the 6990 is being beat by a 6970 in games as modern as Skyrim - their drivers are just full of fail.
  • Galidou - Thursday, May 3, 2012 - link

    A much higher percentage?!? that's 7% more... nothing extraordinary...Let's just say a higher percentage, when you say much, it makes us beleive Nvidia's paying you.
  • CeriseCogburn - Saturday, May 5, 2012 - link

    10% you might be able to ignore, 17% you cannot. It's much higher, it changes several of the games here as to who wins in the article in the accumulated benches.
    It's a big difference.

Log in

Don't have an account? Sign up now