Earlier this week NVIDIA announced their new top-end single-GPU consumer card, the GeForce GTX Titan. Built on NVIDIA’s GK110 and named after the same supercomputer that GK110 first powered, the GTX Titan is in many ways the apex of the Kepler family of GPUs first introduced nearly one year ago. With anywhere between 25% and 50% more resources than NVIDIA’s GeForce GTX 680, Titan is intended to be the ultimate single-GPU card for this generation.

Meanwhile with the launch of Titan NVIDIA has repositioned their traditional video card lineup to change who the ultimate video card will be chasing. With a price of $999 Titan is decidedly out of the price/performance race; Titan will be a luxury product, geared towards a mix of low-end compute customers and ultra-enthusiasts who can justify buying a luxury product to get their hands on a GK110 video card. So in many ways this is a different kind of launch than any other high performance consumer card that has come before it.

So where does that leave us? On Tuesday we could talk about Titan’s specifications, construction, architecture, and features. But the all-important performance data would be withheld another two days until today. So with Thursday finally upon us, let’s finish our look at Titan with our collected performance data and our analysis.

Titan: A Performance Summary

  GTX Titan GTX 690 GTX 680 GTX 580
Stream Processors 2688 2 x 1536 1536 512
Texture Units 224 2 x 128 128 64
ROPs 48 2 x 32 32 48
Core Clock 837MHz 915MHz 1006MHz 772MHz
Shader Clock N/A N/A N/A 1544MHz
Boost Clock 876Mhz 1019MHz 1058MHz N/A
Memory Clock 6.008GHz GDDR5 6.008GHz GDDR5 6.008GHz GDDR5 4.008GHz GDDR5
Memory Bus Width 384-bit 2 x 256-bit 256-bit 384-bit
VRAM 6GB 2 x 2GB 2GB 1.5GB
FP64 1/3 FP32 1/24 FP32 1/24 FP32 1/8 FP32
TDP 250W 300W 195W 244W
Transistor Count 7.1B 2 x 3.5B 3.5B 3B
Manufacturing Process TSMC 28nm TSMC 28nm TSMC 28nm TSMC 40nm
Launch Price $999 $999 $499 $499

On paper, compared to GTX 680, Titan offers anywhere between a 25% and 50% increase in resource. At the starting end, Titan comes with 25% more ROP throughput, a combination of Titan’s 50% increase in ROP count and simultaneous decrease in clockspeeds relative to GTX 680. Shading and texturing performance meanwhile benefits even more from the expansion of the number of SMXes, from 8 to 14. And finally, Titan has a full 50% more memory bandwidth than GTX 680.

Setting aside the unique scenario of compute for a moment, this means that Titan will be between 25% and 50% faster than GTX 680 in GPU limited situations, depending on the game/application and its mix of resource usage. For an industry and userbase still trying to come to terms with the loss of nearly annual half-node jumps, this kind of performance jump on the same node is quite remarkable. At the same time it also sets expectations for how future products may unfold; one way to compensate for the loss of the rapid cadence in manufacturing nodes is to spread out the gains from a new node over multiple years, and this is essentially what we’ve seen with the Kepler family by launching GK104, and a year later GK110.

In any case, while Titan can improve gaming performance by up to 50%, NVIDIA has decided to release Titan as a luxury product with a price roughly 120% higher than the GTX 680. This means that Titan will not be positioned to push the price of NVIDIA’s current cards down, and in fact it’s priced right off the currently hyper-competitive price-performance curve that the GTX 680/670 and Radeon HD 7970GE/7970 currently occupy.

February 2013 GPU Pricing Comparison
AMD Price NVIDIA
  $1000 GeForce Titan/GTX 690
(Unofficial) Radeon HD 7990 $900  
Radeon HD 7970 GHz Edition $450 GeForce GTX 680
Radeon HD 7970 $390  
  $350 GeForce GTX 670
Radeon HD 7950 $300  

This setup isn’t unprecedented – the GTX 690 more or less created this precedent last May – but it means Titan is a very straightforward case of paying 120% more for 50% more performance; the last 10% always costs more. What this means is that the vast majority of gamers will simply be shut out from Titan at this price, but for those who can afford Titan’s $999 price tag NVIDIA believes they have put together a powerful card and a convincing case to pay for luxury.

So what can potential Titan buyers look forward to on the performance front? As always we’ll do a complete breakdown of performance in the following pages, but we wanted to open up this article with a quick summary of performance. So with that said, let’s take a look at some numbers.

GeForce GTX Titan Performance Summary (2560x1440)
  vs. GTX 680 vs. GTX 690 vs. R7970GE vs. R7990
Average +47% -15% 34% -19%
Dirt: Showdown 47% -5% 3% -38%
Total War: Shogun 2 50% -15% 62% 1%
Hitman: Absolution 34% -15% 18% -15%
Sleeping Dogs 49% -15% 17% -30%
Crysis 54% -13% 21% -25%
Far Cry 3 35% -23% 37% -15%
Battlefield 3 48% -18% 52% -11%
Civilization V 59% -9% 60% 0

Looking first at NVIDIA’s product line, Titan is anywhere between 33% and 54% faster than the GTX 680. In fact with the exception of Hitman: Absolution, a somewhat CPU-bound benchmark, Titan’s performance relative to the GTX 680 is actually very consistent at a narrow 45%-55% range. Titan and GTX 680 are of course based on the same fundamental Kepler architecture, so there haven’t been any fundamental architecture changes between the two; Titan is exactly what you’d expect out of a bigger Kepler GPU. At the same time this is made all the more interesting due to the fact that Titan’s real-world performance advantage of 45%-55% is so close to its peak theoretical performance advantage of 50%, indicating that Titan doesn’t lose much (if anything) in efficiency when scaled up, and that the games we’re testing today favor memory bandwidth and shader/texturing performance over ROP throughput.

Moving on, while Titan offers a very consistent performance advantage over the architecturally similar GTX 680, it’s quite a different story when compared to AMD’s fastest single-GPU product, the Radeon HD 7970 GHz Edition. As we’ve seen time and time again this generation, the difference in performance between AMD and NVIDIA GPUs not only varies with the test and settings, but dramatically so. As a result Titan is anywhere between being merely equal to the 7970GE to being nearly a generation ahead of it.

At the low-end of the scale we have DiRT: Showdown, where Titan’s lead is less than 3%. At the other end is Total War: Shogun 2, where Titan is a good 62% faster than the 7970GE. The average gain over the 7970GE is almost right in the middle at 34%, reflecting a mix of games where the two are close, the two are far, and the two are anywhere in between. With recent driver advancements having helped the 7970GE pull ahead of the GTX 680, NVIDIA had to work harder to take back their lead and to do so in an concrete manner.

Titan’s final competition are the dual-GPU cards of this generation, the GK104 based GTX 690, and the officially unofficial Tahiti based HD 7990 cards, which vary in specs but generally have just shy of the performance of a pair of 7970s. As we’ve seen in past generations, when it comes to raw performance one big GPU is no match for two smaller GPUs, and the same is true with Titan. For frames per second and nothing else, Titan cannot compete with those cards. But as we’ll see there are still some very good reasons for Titan’s existence, and areas Titan excels at that even two lesser GPUs cannot match.

None of this of course accounts for compute. Simply put, Titan stands alone in the compute world. As the first consumer GK110 GPU based video card there’s nothing quite like it. We’ll see why that is in our look at compute performance, but as far as the competitive landscape is concerned there’s not a lot to discuss here.

The Final Word On Overclocking
Comments Locked

337 Comments

View All Comments

  • CeriseCogburn - Tuesday, February 26, 2013 - link

    I really don't understand that mentality you have. I'm surrounded by thousands of dollars of computer parts and I certainly don't consider myself some sort of hardware enthusiast or addicted overclocker, or insane gamer.

    Yet this card is easily a consideration, since several other systems have far more than a thousand dollars in them on just the basics. It's very easy to spend a couple thousand even being careful.

    I don't get what the big deal is. The current crop of top end cards before this are starkly inadequate at common monitor resolutions.
    One must nearly ALWAYS turn down features in the popular benched games to be able to play.

    People just don't seem to understand that I guess. I have untold thousands of dollars in many computers and the only thing that will make them really gaming capable at cheap monitor resolutions is a card like this.

    Cripes my smartphone cost a lot more than the former top two cards just below Titan.

    This is the one area that comes to mind ( the only one that exists as far as I can tell) where the user is left with "my modern computer can't do it" - and that means, take any current taxing game (lots of those - let's say 50% of those reviewed as a rough thumb) and you're stuck unable to crank it up.

    Now 120hz monitors are becoming common, so this issue is increased.
    As you may have noticed, another poster exclaimed:
    " Finally ! 1920x1080 a card that can do it ! "

    There's the flat out closest to the truth, and I agree with that entirely, at least for this moment, as I stated here before the 7970 didn't do it when it was released and doesn't now and won't ever. (neither does the 680)

    I'm trying to deny it, but really it is already clear that the Titan doesn't cut it for everything at the above rez either, not really, and not at higher refresh rates.

    More is still needed, and this is the spot that is lacking for gamers, the video card.

    This card is the card to have, and it's not about bragging, it's about firing up your games and not being confronted with the depressing "turn off the eyecandy" and check the performance again... see if that is playable...

    I mean ****, that apparently does not bother any of you, and I do not know why.
    Everything else in your system is capable...
    This is an IMPORTANT PART that actually completes the package, where the end user isn't compromising.
  • HighTech4US - Thursday, February 21, 2013 - link

    If it does could we see a new story on performance using NVENC across the entire Kepler line along with any FREEware/PAYware software that utilizes it. I have an older Intel Q8300 that is used as my HTPC/Living Room Gaming System and encoding videos take a long time just using the CPU cores.

    If getting a Kepler GPU and using NVENC can speed up encoding significantly I would like to know. As that would be the lowest cost upgrade along with getting a Gaming card upgrade.

    Thanks
  • Ryan Smith - Thursday, February 21, 2013 - link

    Yes, NVEnc is present.
  • lkuzmanov - Thursday, February 21, 2013 - link

    excellent! now make it 30-40% cheaper and I'm on board.
  • Zink - Thursday, February 21, 2013 - link

    Rahul Garg picked the lowest HD 7970 scores in both cases from the Matsumoto et al. paper. The other higher GFLOPS scores represent performance using alternate kernels performing the same calculation on the same hardware as far as I can tell. Rahul needs to justify choosing only the lowest HD 7970 numbers in his report or I can only assume he is tilting the numbers in favor of Titan.
  • JarredWalton - Thursday, February 21, 2013 - link

    Picking the highest scoring results that are using optimized cores and running on different hardware in the first place (e.g. not the standard test bed) would be tilting the results very far in AMD's favor. A default run is basically what Titan gets to do, so the same for 7970 would make sense.
  • codedivine - Thursday, February 21, 2013 - link

    The different algorithms are actually not performing the exact same calculation. There are differences in matrix layouts and memory allocations. We chose the ones that are closest to the layouts and allocations we were testing on the Titan.

    In the future, we intend to test with AMD's official OpenCL BLAS. While Matsumoto's numbers are good for illustrative purposes. We would prefer running our own benchmarks on our own testbeds, and on real-world code which will typically use AMD's BLAS for AMD cards. AMD's OpenCL BLAS performance is actually a little bit lower than Matsumoto's numbers so I don't think we tilted the numbers in AMD's favour. If anything, we gave AMD a bit of benefit-of-the-doubt here.

    In the same vein, faster results than Nvidia's CUBLAS have been demonstrated on Nvidia hardware. However, we chose to test only using CUBLAS as all production code will typically use CUBLAS due to its reliability and support from Nvidia.

    AMD's OpenCL BLAS is a bit complicated to setup correctly and in my research, I have had problems with stability with it on Windows. Thus, we avoided it in this particular review but we will likely look at it in the future.
  • Zink - Thursday, February 21, 2013 - link

    Thanks, shouldn't have doubted you :)
  • Nfarce - Thursday, February 21, 2013 - link

    ...about my 680 purchase last April (nearly a year ago already, wow). Was so worried I made the wrong decision replacing two 570s knowing the Kepler was less than a year away. The news on this card has firmed up my decision to lock in with a second 680 now for moving up to a 2560x1440 monitor.

    Very *very* disappointing, Nvidia.
  • CeriseCogburn - Thursday, February 21, 2013 - link

    The new top card has been near the same as two of the former cards FOREVER.

    You people are nothing short of stupid nut jobs.

    There are not enough tampons at Johnson and Johnson warehouses for this thread.

    THE VERY SAME RATIO has occurred every time for all the prior launches.

Log in

Don't have an account? Sign up now