Earlier this week NVIDIA announced their new top-end single-GPU consumer card, the GeForce GTX Titan. Built on NVIDIA’s GK110 and named after the same supercomputer that GK110 first powered, the GTX Titan is in many ways the apex of the Kepler family of GPUs first introduced nearly one year ago. With anywhere between 25% and 50% more resources than NVIDIA’s GeForce GTX 680, Titan is intended to be the ultimate single-GPU card for this generation.

Meanwhile with the launch of Titan NVIDIA has repositioned their traditional video card lineup to change who the ultimate video card will be chasing. With a price of $999 Titan is decidedly out of the price/performance race; Titan will be a luxury product, geared towards a mix of low-end compute customers and ultra-enthusiasts who can justify buying a luxury product to get their hands on a GK110 video card. So in many ways this is a different kind of launch than any other high performance consumer card that has come before it.

So where does that leave us? On Tuesday we could talk about Titan’s specifications, construction, architecture, and features. But the all-important performance data would be withheld another two days until today. So with Thursday finally upon us, let’s finish our look at Titan with our collected performance data and our analysis.

Titan: A Performance Summary

  GTX Titan GTX 690 GTX 680 GTX 580
Stream Processors 2688 2 x 1536 1536 512
Texture Units 224 2 x 128 128 64
ROPs 48 2 x 32 32 48
Core Clock 837MHz 915MHz 1006MHz 772MHz
Shader Clock N/A N/A N/A 1544MHz
Boost Clock 876Mhz 1019MHz 1058MHz N/A
Memory Clock 6.008GHz GDDR5 6.008GHz GDDR5 6.008GHz GDDR5 4.008GHz GDDR5
Memory Bus Width 384-bit 2 x 256-bit 256-bit 384-bit
VRAM 6GB 2 x 2GB 2GB 1.5GB
FP64 1/3 FP32 1/24 FP32 1/24 FP32 1/8 FP32
TDP 250W 300W 195W 244W
Transistor Count 7.1B 2 x 3.5B 3.5B 3B
Manufacturing Process TSMC 28nm TSMC 28nm TSMC 28nm TSMC 40nm
Launch Price $999 $999 $499 $499

On paper, compared to GTX 680, Titan offers anywhere between a 25% and 50% increase in resource. At the starting end, Titan comes with 25% more ROP throughput, a combination of Titan’s 50% increase in ROP count and simultaneous decrease in clockspeeds relative to GTX 680. Shading and texturing performance meanwhile benefits even more from the expansion of the number of SMXes, from 8 to 14. And finally, Titan has a full 50% more memory bandwidth than GTX 680.

Setting aside the unique scenario of compute for a moment, this means that Titan will be between 25% and 50% faster than GTX 680 in GPU limited situations, depending on the game/application and its mix of resource usage. For an industry and userbase still trying to come to terms with the loss of nearly annual half-node jumps, this kind of performance jump on the same node is quite remarkable. At the same time it also sets expectations for how future products may unfold; one way to compensate for the loss of the rapid cadence in manufacturing nodes is to spread out the gains from a new node over multiple years, and this is essentially what we’ve seen with the Kepler family by launching GK104, and a year later GK110.

In any case, while Titan can improve gaming performance by up to 50%, NVIDIA has decided to release Titan as a luxury product with a price roughly 120% higher than the GTX 680. This means that Titan will not be positioned to push the price of NVIDIA’s current cards down, and in fact it’s priced right off the currently hyper-competitive price-performance curve that the GTX 680/670 and Radeon HD 7970GE/7970 currently occupy.

February 2013 GPU Pricing Comparison
AMD Price NVIDIA
  $1000 GeForce Titan/GTX 690
(Unofficial) Radeon HD 7990 $900  
Radeon HD 7970 GHz Edition $450 GeForce GTX 680
Radeon HD 7970 $390  
  $350 GeForce GTX 670
Radeon HD 7950 $300  

This setup isn’t unprecedented – the GTX 690 more or less created this precedent last May – but it means Titan is a very straightforward case of paying 120% more for 50% more performance; the last 10% always costs more. What this means is that the vast majority of gamers will simply be shut out from Titan at this price, but for those who can afford Titan’s $999 price tag NVIDIA believes they have put together a powerful card and a convincing case to pay for luxury.

So what can potential Titan buyers look forward to on the performance front? As always we’ll do a complete breakdown of performance in the following pages, but we wanted to open up this article with a quick summary of performance. So with that said, let’s take a look at some numbers.

GeForce GTX Titan Performance Summary (2560x1440)
  vs. GTX 680 vs. GTX 690 vs. R7970GE vs. R7990
Average +47% -15% 34% -19%
Dirt: Showdown 47% -5% 3% -38%
Total War: Shogun 2 50% -15% 62% 1%
Hitman: Absolution 34% -15% 18% -15%
Sleeping Dogs 49% -15% 17% -30%
Crysis 54% -13% 21% -25%
Far Cry 3 35% -23% 37% -15%
Battlefield 3 48% -18% 52% -11%
Civilization V 59% -9% 60% 0

Looking first at NVIDIA’s product line, Titan is anywhere between 33% and 54% faster than the GTX 680. In fact with the exception of Hitman: Absolution, a somewhat CPU-bound benchmark, Titan’s performance relative to the GTX 680 is actually very consistent at a narrow 45%-55% range. Titan and GTX 680 are of course based on the same fundamental Kepler architecture, so there haven’t been any fundamental architecture changes between the two; Titan is exactly what you’d expect out of a bigger Kepler GPU. At the same time this is made all the more interesting due to the fact that Titan’s real-world performance advantage of 45%-55% is so close to its peak theoretical performance advantage of 50%, indicating that Titan doesn’t lose much (if anything) in efficiency when scaled up, and that the games we’re testing today favor memory bandwidth and shader/texturing performance over ROP throughput.

Moving on, while Titan offers a very consistent performance advantage over the architecturally similar GTX 680, it’s quite a different story when compared to AMD’s fastest single-GPU product, the Radeon HD 7970 GHz Edition. As we’ve seen time and time again this generation, the difference in performance between AMD and NVIDIA GPUs not only varies with the test and settings, but dramatically so. As a result Titan is anywhere between being merely equal to the 7970GE to being nearly a generation ahead of it.

At the low-end of the scale we have DiRT: Showdown, where Titan’s lead is less than 3%. At the other end is Total War: Shogun 2, where Titan is a good 62% faster than the 7970GE. The average gain over the 7970GE is almost right in the middle at 34%, reflecting a mix of games where the two are close, the two are far, and the two are anywhere in between. With recent driver advancements having helped the 7970GE pull ahead of the GTX 680, NVIDIA had to work harder to take back their lead and to do so in an concrete manner.

Titan’s final competition are the dual-GPU cards of this generation, the GK104 based GTX 690, and the officially unofficial Tahiti based HD 7990 cards, which vary in specs but generally have just shy of the performance of a pair of 7970s. As we’ve seen in past generations, when it comes to raw performance one big GPU is no match for two smaller GPUs, and the same is true with Titan. For frames per second and nothing else, Titan cannot compete with those cards. But as we’ll see there are still some very good reasons for Titan’s existence, and areas Titan excels at that even two lesser GPUs cannot match.

None of this of course accounts for compute. Simply put, Titan stands alone in the compute world. As the first consumer GK110 GPU based video card there’s nothing quite like it. We’ll see why that is in our look at compute performance, but as far as the competitive landscape is concerned there’s not a lot to discuss here.

The Final Word On Overclocking
Comments Locked

337 Comments

View All Comments

  • varg14 - Thursday, February 21, 2013 - link

    I will hang on to my SLI 560 tis for a while longer. Since i game at 1080p they perform very well.
  • mayankleoboy1 - Thursday, February 21, 2013 - link

    Some video conversion benchmarks please.
  • mayankleoboy1 - Thursday, February 21, 2013 - link

    Ohh, and the effect of PCIE2.0 VS PCIE3.0 also. Lets see how much is the Titan gimped by PCIE2.0
  • Ryan Smith - Thursday, February 21, 2013 - link

    This isn't something we can do at this second, but it's definitely something we can follow up on once things slow down a bit.
  • mayankleoboy1 - Thursday, February 21, 2013 - link

    Sure. I am looking forward to a part three of the Titan review
  • Hrel - Thursday, February 21, 2013 - link

    The problem with that reasoning, that they're raising here, is that the 7970 is almost as fast and costs a lot less. The Titan is competing, based on performance, with the 7970. Based on that comparison it's a shitty deal.

    http://www.newegg.com/Product/Product.aspx?Item=N8...

    $430. So based on that I'd say the highest price you can justify for this card is $560. We'll round up to $600.

    Nvidia shipping this, at this price, and just saying "it's a luxury product" is bullshit. It's not a luxury product, it's their version of a 7970GHE. But they want to try and get a ridiculous profit to support their PhysX and CUDA projects.

    Nvidia just lost me as a customer. This is the last straw. This card should be pushing the pricing down on the rest of their lineup. They SHOULD be introducing it to compete with the 7970GHE. Even at my price range, compare the GTX660 to the 7870GHE, or better yet the sub $200 7850. They just aren't competitive anymore. I'll admit, I was a bit of a Nvidia fan boy. Loved their products. Was disappointed by older ATI cards and issues I had with them. (stability, screen fitting properly, audio issues) But ATI has become AMD and they've improved quality a lot and Nvidia is USING their customers loyalty; that's just wrong.

    I'm done with Nvidia on the desktop. By the time I need a new laptop AMD will probably have the graphics switching all sorted; so I'm probably done with Nvidia on laptops too.
  • CeriseCogburn - Saturday, February 23, 2013 - link

    LOL - be done, and buy the alternative crap - amd.

    You'll be sorry, and when you have to hold back admitting it, I'll be laughing the whole time.

    Poor baby can't pony up the grand, so he's boycotting the whole line.
    You know you people are the sickest freaks the world has ever seen, and frankly I don't believe you, and consider you insane.

    You're all little crybaby socialist activists. ROFL You're all pathetic.

    nVidia won't listen to you, so go blow on your amd crash monkey, you and two other people will do it before amd disappears into bankruptcy, and then we can laugh at your driver less video cards.

    I never seen bigger crybaby two year olders in my entire life. You all live in your crybaby world together, in solidarity - ROFL

    No one cares if you lying turds claim you aren't buying nVidia - they have billions and are DESTROYING amd because you cheapskate losers cost amd money - LOL

    YOU ARE A BURDEN AND CANNOT PAY FOR THE PRODUCTION OF A VIDEO CARD !

    Enjoy your false amd ghetto loser lifestyle.
  • Soulnibbler - Thursday, February 21, 2013 - link

    Hey, I'm excited about the fp64 performance but I'm not going to have any time to write code for a bit so I'll ask the question that would let me justify buying a card like this:

    How much acceleration should I expect using this card with Capture One as compared to AMD/software rendering. I've heard anecdotal evidence that the openCL code paths in version 7 make everything much faster, but I'd like a metric before I give up my current setup (windows in VMware) and dual-boot to get openCL support.

    I know openCL is not yet ready on this card but when you revisit it could we see a little Capture One action?

    Preferably the benchmark sets would be high resolution images at both high and low iso.
  • Ryan Smith - Monday, February 25, 2013 - link

    I'm afraid I don't know anything about Capture One. Though if you could describe it, that would be helpful.
  • Soulnibbler - Monday, February 25, 2013 - link

    Capture One is a raw developer for digital cameras.
    http://www.phaseone.com/en/Imaging-Software.aspx
    notably for medium format digital backs but also for 35mm and aps sensors. It could be considered a competitor to Adobe's Lightroom and ACR software but the medium format camera support and workflow are the major differentiators.

    The last two releases have had openCL support for both previews and exporting which I've heard has lead to reductions in time to get an image through post.

    I'd imagine that one could benchmark on a large library of photos and determine if this card as a compute card is any improvement over standard gaming cards in this use scenario.

    I'd imagine this is part of the market that NVIDIA is aiming at as I know at least one user who switched to an ATI W7000 for openCL support with Capture One.

Log in

Don't have an account? Sign up now