Power, Temperature, & Noise

As always, last but not least is our look at power, temperature, and noise. Next to price and performance of course, these are some of the most important aspects of a GPU, due in large part to the impact of noise. All things considered, a loud card is undesirable unless there’s a sufficiently good reason – or sufficiently good performance – to ignore the noise.

GeForce GTX 780 Series Voltages
GTX 780 Ti Boost Voltage GTX 780 Boost Voltage GTX 780 Ti Base Voltage
1.187v 1.1625v 1.012v

Taking a quick look at voltages, we find that our GTX 780 Ti operates at a slightly higher voltage at its maximum boost bin than the original GTX 780 did. The difference is minor, but the additional voltage may be necessary to hit the slightly higher clockspeeds GTX 780 Ti operates at relative to GTX Titan and GTX 780.

GeForce GTX 780 Ti Average Clockspeeds
Max Boost Clock 1020MHz
Metro: LL
1000MHz
CoH2
997MHz
Bioshock
954MHz
Battlefield 3
980MHz
Crysis 3
980MHz
Crysis: Warhead
1000MHz
TW: Rome 2
950MHz
Hitman
993MHz
GRID 2
967MHz
Furmark
823MHz

Moving on to clockspeeds, we find that the GTX 780 Ti does very well when it comes to boosting. With a maximum boost clock of 1020MHz, we have 2 benchmarks averaging 1000MHz, and another 4 averaging 980MHz or better.

With all of our GK110 cards sharing a common design, at idle there’s very little to differentiate them. Other than GTX Titan’s extra 3GB of VRAM, we’re essentially looking at identical cards when idling.

Moving on to load power, we can see the power/heat ramifications of the slight clockspeed increase coupled with the activation of the 15th SMX. Even with the further optimizations NVIDIA has put into the new revision of GK110, power consumption has gone up in accordance with the higher performance of the card, just as we’d expect. Since NVIDIA doesn’t notably alter their power efficiency here, that increased performance has to come at the cost of increased power consumption. Though in this benchmark it’s worth pointing out that we’re measuring from the wall and that GTX 780 Ti outperforms GTX Titan by 8%, so some of that 29W power difference will come from the higher CPU load caused by the increased framerates.

As for the GTX 780 Ti SLI, here we see power level off at 556W, 20W more than the GTX 780 SLI. Some (if not most) of that is going to be explained by the increased CPU power consumption from the GTX 780 Ti SLI’s higher framerates. Coupled with that is the fact that in SLI setups these cards get hotter, and hence have to downclock a bit more to maintain equilibrium, which helps to offset the increased power requirements of GTX 780 Ti and keep the SLI results so close to the GTX 780 SLI results.

Switching over to FurMark, we find that power consumption is also up, but only slightly. With GPU Boost 2.0 clamping down on power consumption all of our GK110 cards should be clamped at 250W here, and with a difference between GTX 780 and GTX 780 Ti of under 10W, that’s exactly what appears to be happening here.

On a side note, it’s interesting to note here that under FurMark we’re seeing the GTX 780 Ti draw more power than the Radeon R9 290X. Despite the fact that the 290X has a higher rated TDP, in the card’s default quiet mode the card can’t actually dissipate as much heat (and thereby consume as much power) as the GTX 780 Ti can.

For idle temperatures we’re once again looking at cards that are for all intents and purposes identical. At 30C the GTX 780 Ti easily stays nice and cool.

As we mentioned in our look at the GTX 780 Ti hardware, NVIDIA has increased their default temperature throttle point from 80C on the GTX Titan/780 to 83C on the GTX 780 Ti. The end result is that in all of our temperature limited tests the GTX 780 Ti will peak at 83C-84C, whereas the older GK110 cards will peak at 80C-81C.

FurMark reiterates what we saw with Crysis 3. The temps are up a bit across the board, while the GK110 cards are holding near their throttle points. The SLI setups meanwhile approach the upper-80s at 88C, reflecting the fact that even with blowers, there’s some impact on neighboring cards in high load situations.

Our last idle scenario, we once again see all of our GK110 cards performing similarly, with idle noise levels in the 38dB-39dB range.

Moving on to our gaming load noise results, we can see the full repercussions of the GTX 780 Ti’s higher average power consumption coupled with the card’s higher temperature throttle point. Moving the throttle point along the same curve has the end result of moving higher the equilibrium point and thereby the card’s operating noise levels. As the fastest single-GPU card on this card, the GTX 780 Ti is still doing very well for itself and for a blower based design at 51.7dB, though at 1.5dB louder than GTX Titan and 4.2dB louder than GTX 780 the noise tradeoff for the card’s higher performance is very clear. Meanwhile the fact that it’s tied with the GTX 780 SLI comes with its own bit of irony.

Speaking of the GTX 780 SLI, we can see the noise impact of SLI configurations too. The GTX 780 Ti SLI levels out at 53.7dB, 2dB louder than our single-card configuration and 2dB louder than the GTX 780 SLI. At this point it’s just a bit louder than the 290X and quieter than a number of other 290 series setups.

Finally with load noise levels under FurMark we can see where our various cards will peak at for noise levels. The GTX 780 Ti creeps up to 52.3dB, essentially tying with the GTX 780 and GTX Titan. Otherwise it comes in just behind the 290X, and the start of the pack for our multi-GPU setups.

As for the GTX 780 Ti SLI, like our single-card comparison points its up slightly as compared to the GTX 780 SLI.

Overall, our look at power, temperatures, and noise has been a rather straightforward validation of our earlier suspicions. GTX 780 Ti’s higher performance leads to higher power consumption, and will all other factors being held equal – including the cooler – power, temps, and noise levels all rise a bit as compared to GTX Titan and GTX 780. There’s no such thing as a free lunch here, and while GPU Boost 2.0 will keep the maximum levels suitably in check, on average GTX 780 Ti is going to be a bit worse than the other GK110 cards due to those factors. Though even with the increased noise levels in particular, GTX 780 Ti is still able to outperform 290X on noise while also delivering better gaming performance, which makes this another tidy victory for NVIDIA.

Compute Overclocking
Comments Locked

302 Comments

View All Comments

  • 1Angelreloaded - Thursday, November 7, 2013 - link

    Physically the 780 and 780 TI are literally the same unit, minus minor things. The difference is the neutered chip, and OC'd VRAM, Which means your paying for the same unit at 2 completely different prices, in fact, How much does it cost to disable the SMX texture? So shouldn't the overhead on the original unit be higher with more work having to be done? Or like AMD tricore are we paying for defective chips again ?
  • TheJian - Thursday, November 7, 2013 - link

    Wrong. They have been saving DEFECT FREE GK110 units for months just to be able to launch this with good quantity (probably only started having better results at B1, which all of these are). I doubt that there are many 780's that have fully working units that are disabled. They are failed Tesla chips (you can say DP is disabled on purpose, but not the SMX's). Do you really think 550mm chips have a ZERO defect rate?...LOL. I would be surprised if the first runs of Titan had any more working SMX's too as they were directly failed Tesla's. Sure there are probably a few cards with some working that are disabled but Yields and history say with chips this big there just has to be a pretty high defect rate vs. 100% working chips. It is pretty much the largest chip TSMC can make. That's not easy. Both AMD and NV do this to salvage failed chips (heck everybody does). You come with a flagship, then anything that fails you mark as a lower model (many models). It allows you to increase your yield and chips that can be sold. You should be thankful they have the tech to do this or we'd all be paying FAR higher prices do to chucking chips by the millions in the trash.
  • TheJian - Thursday, November 7, 2013 - link

    http://www.tomshardware.com/reviews/geforce-gtx-78...
    "Not to be caught off-guard, Nvidia was already binning its GK110B GPUs, which have been shipping since this summer on GeForce GTX 780 and Titan cards. The company won’t get specific about what it was looking for, but we have to imagine it set aside flawless processors with the lowest power leakage to create a spiritual successor for GeForce GTX 580. Today, those fully-functional GPUs drop into Nvidia’s GeForce GTX 780 Ti."

    There, don't have to believe me...confirmed I guess ;)
  • beck2448 - Thursday, November 7, 2013 - link

    Great job Nvidia! I think the partners with custom cooling will get another 15 to 20 % performance out of it with lower temps and less noise, and that is insane for a single GPU. Can't wait to see the Lightning and Windforce editions.
  • aznjoka - Thursday, November 7, 2013 - link

    The crossfire scaling on the 290x is much better than the 780ti. If you are running a dual card set up, getting a 290x is pretty much a no brainer.
  • beck2448 - Thursday, November 7, 2013 - link

    from Benchmark reviews: In conclusion, GeForce GTX 780 Ti is the gamer’s version of GTX TITAN with a powerful lead ahead of Radeon R9 290X. Even if it were possible for the competition to overclock and reach similar frame rate performance, temperatures and noise would still heavily favor the GTX 780 Ti design. I was shocked at how loud AMD’s R9 290X would roar once it began to heat up midway through a benchmark test, creating a bit of sadness for gamers trying to play with open speakers instead of an insulated headset. There is a modest price difference between them, but quite frankly, the competition doesn’t belong in the same class.
    Read more at http://benchmarkreviews.com/8468/nvidia-geforce-gt...
  • 1Angelreloaded - Thursday, November 7, 2013 - link

    TBH the stock Nvidia cooler isn't that much better either they tend to run a lot hotter than ACX/Twin Frozer, and other such solutions, so both have cooling headroom, Hawaii though is just plain ridiculous.
  • deedubs - Thursday, November 7, 2013 - link

    Noticed the graph for shadowplay performance has its labels reversed. It makes it look like SP increases performance instead of decreasing.
  • Ryan Smith - Thursday, November 7, 2013 - link

    Whoops. Thanks for that. The multi-series graph tool is a bit picky...
  • Filiprino - Thursday, November 7, 2013 - link

    I don't see NVIDIA as a real winner here, really. Their margin is very tight, and AMD drivers still have to mature, and when you talk about crossfire, AMD is doing clearly better, for $200 less and 6dB more.

Log in

Don't have an account? Sign up now