Power, Temperature, & Noise

As always, last but not least is our look at power, temperature, and noise. Next to price and performance of course, these are some of the most important aspects of a GPU, due in large part to the impact of noise. All things considered, a loud card is undesirable unless there’s a sufficiently good reason – or sufficiently good performance – to ignore the noise.

As the GM200 flagship card, GTX Titan X gets the pick of the litter as far as GM200 GPUs go. GTX Titan X needed fully-functional GM200 GPUs, and even then needed GPUs that were good enough to meet NVIDIA’s power requirements. GTX 980 Ti on the other hand, as a cut-down/salvage card, gets second pick. So we expect to see these chips be just a bit worse; to have either functional units that came out of the fab damaged, or have functional units that have been turned off due to power reasons.

GeForce GTX Titan X/980 Voltages
GTX Titan X Boost Voltage GTX 980 Ti Boost Voltage GTX 980 Boost Voltage
1.162v 1.187v 1.225v

Looking at voltages, we can see just that in our samples. GTX 980 Ti has a slightly higher boost voltage – 1.187v – than our GTX Titan X. NVIDIA sometimes bins their second-tier cards for lower voltage, but this isn’t something we’re seeing here. Nor is there necessarily a need to bin in such a manner since the 250W TDP is unchanged from GTX Titan X.

GeForce GTX 980 Ti Average Clockspeeds
Game GTX 980 Ti GTX Titan X
Max Boost Clock 1202MHz 1215MHz
Battlefield 4
1139MHz
1088MHz
Crysis 3
1177MHz
1113MHz
Mordor
1151MHz
1126MHz
Civilization: BE
1101MHz
1088MHz
Dragon Age
1189MHz
1189MHz
Talos Principle
1177MHz
1126MHz
Far Cry 4
1139MHz
1101MHz
Total War: Attila
1139MHz
1088MHz
GRID Autosport
1164MHz
1151MHz
Grand Theft Auto V
1189MHz
1189MHz

The far more interesting story here is GTX 980 Ti’s clockspeeds. As we have pointed out time and time again, GTX 980 Ti’s gaming performance trails GTX Titan X by just a few percent, this despite the fact that GTX 980 Ti is down by 2 SMMs and is clocked identically. On paper there is a 9% performance difference that in the real world we’re not seeing. So what’s going on?

The answer to that is that what GTX 980 Ti lacks in SMMs it’s making up in clockspeeds. The card’s average clockspeeds are frequently two or more bins ahead of GTX Titan X, topping out at a 64MHz advantage under Crysis 3. All of this comes despite the fact that GTX 980 Ti has a lower maximum boost clock than GTX Titan X, topping out one bin lower at 1202MHz to GTX Titan X’s 1215MHz.

Ultimately the higher clockspeeds are a result of the increased power and thermal headroom the GTX 980 Ti picks up from halving the number of VRAM chips along with disabling two SMMs. With those components no longer consuming power or generating heat, and yet the TDP staying at 250W, GTX 980 Ti can spend its power savings to boost just a bit higher. This in turn compresses the performance gap between the two cards (despite what the specs say), which coupled with the fact that performance doesn't scale lineraly with SMM count or clockspeed (you rarely lose the full theoretical performance amount when shedding frequency or functional units) leads to the GTX 980 Ti trailing the GTX Titan X by an average of just 3%.

Idle Power Consumption

Starting off with idle power consumption, there's nothing new to report here. GTX 980 Ti performs just like the GTX Titan X, which at 74W is second only to the GTX 980 by a single watt.

Load Power Consumption - Crysis 3

Load Power Consumption - FurMark

Meanwhile load power consumption is also practically identical to the GTX Titan X. With the same GPU on the same board operating at the same TDP, GTX 980 Ti ends up right where we expect it, next to GTX Titan X. GTX Titan X did very well as far as energy efficiency is concerned – setting a new bar for 250W cards – and GTX 980 Ti in turn does just as well.

Idle GPU Temperature

Load GPU Temperature - Crysis 3

Load GPU Temperature - FurMark

As was the case with power consumption, video card temperatures are similarly unchanged. NVIDIA’s metal cooler does a great job here, keeping temperatures low at idle while NVIDIA’s GPU Boost mechanism keeps temperatures from exceeding 83C under full load.

Idle Noise Levels

Load Noise Levels - Crysis 3

Load Noise Levels - FurMark

Finally for noise, the situation is much the same. Unexpected but not all that surprising, the GTX 980 Ti ends up doing a hair worse than the GTX Titan X here. NVIDIA has not changed the fan curves or TDP, so this ultimately comes down to manufacturing variability in NVIDIA’s metal cooler, with our GTX 980 Ti faring ever so slightly worse than the Titan. Which is to say that it's still right at the sweet spot for noise versus power consumption, dissipating 250W at no more than 53dB, and once again proving the mettle of NVIDIA's metal cooler.

Compute Overclocking
Comments Locked

290 Comments

View All Comments

  • kyuu - Monday, June 1, 2015 - link

    Witcher 3 runs just fine on my single 290. Is it just the xfire profile? Do you have the new driver and latest patches? Also, have you turned down tesselation or turned off hairworks?
  • PEJUman - Monday, June 1, 2015 - link

    4K... was hoping my U28D590D will have freesync, but alas... no such luck. I am very sensitive to stutter, it gives me motion sickness, to the point I have to stop playing :(

    limiting hairworks to 8x does help, but I really dislike the hair without it. I rather wait for 15.5.1 or 15.6. I have other games to keep me busy for a while.

    I can get 45 avg if I drop to 21:9 ratio using 2840 x 1646, but even then I still get motion sickness from the occasional drops.
  • chizow - Monday, June 1, 2015 - link

    Yes CrossFire support of TW3 is broken from Day1, its a well-known issue. AMD hastily released a driver last week with a CF profile, but its virtually unusable as it introduces a number of other issues with AA and flickering icons.
  • PEJUman - Monday, June 1, 2015 - link

    15.5 no longer flickers with or without AA. still slow though.
  • chizow - Monday, June 1, 2015 - link

    Are you sure? Did they release a follow-up to the 15.5 Beta? Because the notes and independent user feedback stated there was still flickering:

    *The Witcher 3: Wild Hunt - To enable the best performance and experience in Crossfire, users must disable Anti-Aliasing from the games video-post processing options. Some random flickering may occur when using Crossfire. If the issue is affecting the game experience, as a work around we suggest disabling Crossfire while we continue to work with CD Projekt Red to resolve this issue
  • Peichen - Monday, June 1, 2015 - link

    295X2 is indeed faster but it also uses twice as much power. You have to take the 1000W PSU into account as well as one or two additional 120mm fans that's needed to get the heat out the case. When you add up all the extra cost for PSU, fans, electricity, noise and stutter against an overclocked 980Ti (last few pages of review), the slight speed advantage aren't going to be worth it.

    Also, Maxwell 2 supports DirectX 12, I am not so sure about any of the current AMD/ATI cards since they were designed in 2013.
  • xthetenth - Monday, June 1, 2015 - link

    You don't have to buy a new PSU every time you buy a high TDP card, but otherwise a valid point. Going multi-GPU for the same performance requires a much bigger price difference to be worth it vs. a single card.
  • Kutark - Monday, June 1, 2015 - link

    Basically you're gonna spend an extra $5/mo on electricity with that card, or $60/yr vs a 980ti. thats actually pretty huge. Thats at 4hrs/day of gaming, at an average of 12c/kwh. If you game 6 or 7 hours a day, its even worse.

    These high power cards are a little ridiculous. 600w just for one video card?!!
  • Daroller - Monday, June 1, 2015 - link

    I had a GTX690, and I run SLI TITAN X. Dual GPU IS a hindrance. You'd have to be blind, stupid, or a rabid fanboy to claim otherwise. The 295x2 isn't exempt from that just because you dislike NV and harbor a not so secret love for AMD.
  • Daroller - Monday, June 1, 2015 - link

    I had a GTX690, and I run SLI TITAN X. Dual GPU IS a hindrance. You'd have to be blind, stupid, or a rabid fanboy to claim otherwise. The 295x2 isn't exempt from that just because you dislike NV and harbor a not so secret love for AMD.

Log in

Don't have an account? Sign up now