NVIDIA’s GeForce GTX Titan Review, Part 2: Titan's Performance Unveiled
by Ryan Smith & Rahul Garg on February 21, 2013 9:00 AM ESTEarlier this week NVIDIA announced their new top-end single-GPU consumer card, the GeForce GTX Titan. Built on NVIDIA’s GK110 and named after the same supercomputer that GK110 first powered, the GTX Titan is in many ways the apex of the Kepler family of GPUs first introduced nearly one year ago. With anywhere between 25% and 50% more resources than NVIDIA’s GeForce GTX 680, Titan is intended to be the ultimate single-GPU card for this generation.
Meanwhile with the launch of Titan NVIDIA has repositioned their traditional video card lineup to change who the ultimate video card will be chasing. With a price of $999 Titan is decidedly out of the price/performance race; Titan will be a luxury product, geared towards a mix of low-end compute customers and ultra-enthusiasts who can justify buying a luxury product to get their hands on a GK110 video card. So in many ways this is a different kind of launch than any other high performance consumer card that has come before it.
So where does that leave us? On Tuesday we could talk about Titan’s specifications, construction, architecture, and features. But the all-important performance data would be withheld another two days until today. So with Thursday finally upon us, let’s finish our look at Titan with our collected performance data and our analysis.
Titan: A Performance Summary
GTX Titan | GTX 690 | GTX 680 | GTX 580 | |
Stream Processors | 2688 | 2 x 1536 | 1536 | 512 |
Texture Units | 224 | 2 x 128 | 128 | 64 |
ROPs | 48 | 2 x 32 | 32 | 48 |
Core Clock | 837MHz | 915MHz | 1006MHz | 772MHz |
Shader Clock | N/A | N/A | N/A | 1544MHz |
Boost Clock | 876Mhz | 1019MHz | 1058MHz | N/A |
Memory Clock | 6.008GHz GDDR5 | 6.008GHz GDDR5 | 6.008GHz GDDR5 | 4.008GHz GDDR5 |
Memory Bus Width | 384-bit | 2 x 256-bit | 256-bit | 384-bit |
VRAM | 6GB | 2 x 2GB | 2GB | 1.5GB |
FP64 | 1/3 FP32 | 1/24 FP32 | 1/24 FP32 | 1/8 FP32 |
TDP | 250W | 300W | 195W | 244W |
Transistor Count | 7.1B | 2 x 3.5B | 3.5B | 3B |
Manufacturing Process | TSMC 28nm | TSMC 28nm | TSMC 28nm | TSMC 40nm |
Launch Price | $999 | $999 | $499 | $499 |
On paper, compared to GTX 680, Titan offers anywhere between a 25% and 50% increase in resource. At the starting end, Titan comes with 25% more ROP throughput, a combination of Titan’s 50% increase in ROP count and simultaneous decrease in clockspeeds relative to GTX 680. Shading and texturing performance meanwhile benefits even more from the expansion of the number of SMXes, from 8 to 14. And finally, Titan has a full 50% more memory bandwidth than GTX 680.
Setting aside the unique scenario of compute for a moment, this means that Titan will be between 25% and 50% faster than GTX 680 in GPU limited situations, depending on the game/application and its mix of resource usage. For an industry and userbase still trying to come to terms with the loss of nearly annual half-node jumps, this kind of performance jump on the same node is quite remarkable. At the same time it also sets expectations for how future products may unfold; one way to compensate for the loss of the rapid cadence in manufacturing nodes is to spread out the gains from a new node over multiple years, and this is essentially what we’ve seen with the Kepler family by launching GK104, and a year later GK110.
In any case, while Titan can improve gaming performance by up to 50%, NVIDIA has decided to release Titan as a luxury product with a price roughly 120% higher than the GTX 680. This means that Titan will not be positioned to push the price of NVIDIA’s current cards down, and in fact it’s priced right off the currently hyper-competitive price-performance curve that the GTX 680/670 and Radeon HD 7970GE/7970 currently occupy.
February 2013 GPU Pricing Comparison | |||||
AMD | Price | NVIDIA | |||
$1000 | GeForce Titan/GTX 690 | ||||
(Unofficial) Radeon HD 7990 | $900 | ||||
Radeon HD 7970 GHz Edition | $450 | GeForce GTX 680 | |||
Radeon HD 7970 | $390 | ||||
$350 | GeForce GTX 670 | ||||
Radeon HD 7950 | $300 |
This setup isn’t unprecedented – the GTX 690 more or less created this precedent last May – but it means Titan is a very straightforward case of paying 120% more for 50% more performance; the last 10% always costs more. What this means is that the vast majority of gamers will simply be shut out from Titan at this price, but for those who can afford Titan’s $999 price tag NVIDIA believes they have put together a powerful card and a convincing case to pay for luxury.
So what can potential Titan buyers look forward to on the performance front? As always we’ll do a complete breakdown of performance in the following pages, but we wanted to open up this article with a quick summary of performance. So with that said, let’s take a look at some numbers.
GeForce GTX Titan Performance Summary (2560x1440) | ||||||
vs. GTX 680 | vs. GTX 690 | vs. R7970GE | vs. R7990 | |||
Average | +47% | -15% | 34% | -19% | ||
Dirt: Showdown | 47% | -5% | 3% | -38% | ||
Total War: Shogun 2 | 50% | -15% | 62% | 1% | ||
Hitman: Absolution | 34% | -15% | 18% | -15% | ||
Sleeping Dogs | 49% | -15% | 17% | -30% | ||
Crysis | 54% | -13% | 21% | -25% | ||
Far Cry 3 | 35% | -23% | 37% | -15% | ||
Battlefield 3 | 48% | -18% | 52% | -11% | ||
Civilization V | 59% | -9% | 60% | 0 |
Looking first at NVIDIA’s product line, Titan is anywhere between 33% and 54% faster than the GTX 680. In fact with the exception of Hitman: Absolution, a somewhat CPU-bound benchmark, Titan’s performance relative to the GTX 680 is actually very consistent at a narrow 45%-55% range. Titan and GTX 680 are of course based on the same fundamental Kepler architecture, so there haven’t been any fundamental architecture changes between the two; Titan is exactly what you’d expect out of a bigger Kepler GPU. At the same time this is made all the more interesting due to the fact that Titan’s real-world performance advantage of 45%-55% is so close to its peak theoretical performance advantage of 50%, indicating that Titan doesn’t lose much (if anything) in efficiency when scaled up, and that the games we’re testing today favor memory bandwidth and shader/texturing performance over ROP throughput.
Moving on, while Titan offers a very consistent performance advantage over the architecturally similar GTX 680, it’s quite a different story when compared to AMD’s fastest single-GPU product, the Radeon HD 7970 GHz Edition. As we’ve seen time and time again this generation, the difference in performance between AMD and NVIDIA GPUs not only varies with the test and settings, but dramatically so. As a result Titan is anywhere between being merely equal to the 7970GE to being nearly a generation ahead of it.
At the low-end of the scale we have DiRT: Showdown, where Titan’s lead is less than 3%. At the other end is Total War: Shogun 2, where Titan is a good 62% faster than the 7970GE. The average gain over the 7970GE is almost right in the middle at 34%, reflecting a mix of games where the two are close, the two are far, and the two are anywhere in between. With recent driver advancements having helped the 7970GE pull ahead of the GTX 680, NVIDIA had to work harder to take back their lead and to do so in an concrete manner.
Titan’s final competition are the dual-GPU cards of this generation, the GK104 based GTX 690, and the officially unofficial Tahiti based HD 7990 cards, which vary in specs but generally have just shy of the performance of a pair of 7970s. As we’ve seen in past generations, when it comes to raw performance one big GPU is no match for two smaller GPUs, and the same is true with Titan. For frames per second and nothing else, Titan cannot compete with those cards. But as we’ll see there are still some very good reasons for Titan’s existence, and areas Titan excels at that even two lesser GPUs cannot match.
None of this of course accounts for compute. Simply put, Titan stands alone in the compute world. As the first consumer GK110 GPU based video card there’s nothing quite like it. We’ll see why that is in our look at compute performance, but as far as the competitive landscape is concerned there’s not a lot to discuss here.
337 Comments
View All Comments
veppers - Saturday, February 23, 2013 - link
Grow up man.CeriseCogburn - Saturday, February 23, 2013 - link
Adults like myself face reality. Lying fanboys act like spoiled brats and cannot stand to hear or see the truth.You're a child.
Alucard291 - Sunday, February 24, 2013 - link
Yup and you sound just like the spoilt brat in question. This is not engadget mate. Go away.CeriseCogburn - Sunday, February 24, 2013 - link
So far you've posted 3 attacks against me, and added exactly NOTHING to any discussion here.It's clear you're the whining troll with nothing to say, so you are one that needs to go away, right ? Right.
chizow - Monday, February 25, 2013 - link
Oh the irony, you are crying about posting personal attacks and adding nothing to any discussion here? That's what every single one of your posts boils down to.Alucard291 - Friday, March 8, 2013 - link
The discussion? You spew random offensive insulting nonsense against anyone who dares to point out that slower + more expensive is worse than faster and cheaper (be it amd or nv).You then proceed to attack people (on a very personal level I might add) for whatever other reason and go on to say that AMD (did anyone except you even mention amd? - well I'm sure some did but mostly due to your constant stream of butthurt) is terrible.
Cool don't use them. Calm down, relax, take a breather go for a walk.
Or of course you can continue whiteknighting some random product that you are unlikely (given the odds) to ever buy for yourself. Who cares. Just get off the neophyte train when you do it. Ok?
CeriseCogburn - Tuesday, March 12, 2013 - link
You have no clue on any odds.Like I said, you people are 3rd worlder crybabies.
Between bragging hardcore upper end users frequent Anandtech, you whine and cry about 1/3rd the price of a decent PC.
You're all full of it, and all act like you're budget is personal malaria in sub saharan Africa, except of course when you're rig bragging.
This is the USA, except of course wherever you penniless paupers reside.
RussianSensation - Thursday, February 21, 2013 - link
Yes, yes. Keep eating NV's marketing. 36-38% faster than a $430 HD7970GE for $1000???!!http://www.computerbase.de/artikel/grafikkarten/20...
Heck, you can buy a $500 Asus Matrix Platinum 7970 and those overclock to 1300mhz, which makes the Matrix 30% faster than the GTX680. Do the math and see where that ends up relative to the Titan.
http://www.hardwarecanucks.com/forum/hardware-canu...
This is really a $699 card max.
CeriseCogburn - Saturday, February 23, 2013 - link
Why buy a crashing piece of crap amd cannot even write drivers for ?Forget it.
AMD is almost gone too, so "future proof" is nowhere except in nVidia's pocket.
Now and in the future, nVidia wins period.
Idiots buy amd.
Hrel - Thursday, February 21, 2013 - link
The problem with that reasoning, that they're raising here, is that the 7970 is almost as fast and costs a lot less. The Titan is competing, based on performance, with the 7970. Based on that comparison it's a shitty deal.http://www.newegg.com/Product/Product.aspx?Item=N8...
$430. So based on that I'd say the highest price you can justify for this card is $560. We'll round up to $600.
Nvidia shipping this, at this price, and just saying "it's a luxury product" is bullshit. It's not a luxury product, it's their version of a 7970GHE. But they want to try and get a ridiculous profit to support their PhysX and CUDA projects.
Nvidia just lost me as a customer. This is the last straw. This card should be pushing the pricing down on the rest of their lineup. They SHOULD be introducing it to compete with the 7970GHE. Even at my price range, compare the GTX660 to the 7870GHE, or better yet the sub $200 7850. They just aren't competitive anymore. I'll admit, I was a bit of a Nvidia fan boy. Loved their products. Was disappointed by older ATI cards and issues I had with them. (stability, screen fitting properly, audio issues) But ATI has become AMD and they've improved quality a lot and Nvidia is USING their customers loyalty; that's just wrong.
I'm done with Nvidia on the desktop. By the time I need a new laptop AMD will probably have the graphics switching all sorted; so I'm probably done with Nvidia on laptops too.