Power, Temperature, & Noise

Wrapping up our look at the stock performance characteristics of the GTX 680 Classified, we have what’s perhaps our most important results section: power, temperature, & noise.

When it comes to custom cards these results can often make or break a card, depending on the niche it’s going for. The GTX 680 Classified is a high performing card, but it’s not a card shooting for performance at any cost, so we expect to find some kind of reasonable balance here.

GeForce GTX 600 Series Voltages
EVGA GTX 680C Boost Load Ref GTX 680 Boost Load EVGA GTX 680C Idle
1.175v 1.175v 0.987v

Note that at its stock settings EVGA must follow NVIDIA’s defaults, so the GTX 680 Classified has the same 1.175v load voltage as the reference GTX 680. This is where binning comes into play, as EVGA needs to identify GK104 GPUs that can reach 1100MHz+ at 1.175v in order to make the GTX 680 Classified a viable product.

EVGA GTX 680 Classified Average Clockspeed
Game Clockspeed
Crysis: Warhead 1188MHz
Metro 2033 1211MHz
DiRT 3 1211MHz
Shogun2 1162MHz
Batman: Arkham City 1188MHz
Portal 2 1211MHz
Battlefield 3 1188MHz
Starcraft II 1188MHz
Skyrim 1211MHz
Civilization V 1188MHz

Meanwhile, looking at the GTX 680 Classified’s median clockspeed during our benchmarks, we can see that our earlier speculation about the impact of such a high power target was true. The maximum boost bin on our card is 1201MHz (versus 1110MHz for our reference card), and with the exception of Shogun 2, the GTX 680 Classified is almost constantly at 1201MHz or the temperature-induced 1188MHz during our tests. With Shogun 2 that drops down to 1162MHz. With such a high power target the resulting clockspeeds are very consistent for a GTX 680 card.

Moving on to power consumption, there are no great surprises here for the GTX 680 Classified when it comes to idle power consumption. With the additional VRM phases and more importantly 8 extra GDDR5 memory chips, idle power consumption has to go up. The impact is that idle power consumption rises by 8-9W, which would push the total idle power consumption of the card up to around 25W. Note however that this is still better than the GTX 580.

Load power on the other hand looks very good. In fact it’s much better than we were expecting. Despite the additional memory chips and the factory overclock, under Metro power consumption only rises 10W at the wall. This is still less than the7970, let alone the 7970GE.

The story with OCCT is much the same. Here we can see that power at the wall increases by all of 12W, which is roughly the remaining difference between the GTX 680 Classified and the 7970. Ultimately the binning process necessary for EVGA to assign GK104 GPUs to their various products has pushed the very best GK104 GPUs into the Classified. The end result is that even at NVIDIA’s standard voltage of 1.175v, the GK104 GPU in our card ends up consuming less power than a more typical GK104 GPU, which allows the GTX 680 Classified to partially offset the higher power consumption of the additional RAM, additional VRM phases, and the factory overclock.

The final result is that the GTX 680 Classified still consumes a bit more power than the reference GTX 680, but not immensely so. Furthermore depending on whether we’re looking at Metro or OCCT, the GTX 680 Classified still draws 30-50W less than the next-nearest competitor, the 7970GE.

Moving on to temperatures, since EVGA has equipped the GTX 680 Classified with what’s functionally a larger version of the GTX 680’s blower, there are no grand surprises here. The GTX 680 Classified idles at 33C, which so far appears to be typical for these latest iterations of high-performance video cards.

Load GPU temperatures on the other hand are clearly benefitting not only from the larger cooler of the GTX 680 Classified, but also its lower GPU power consumption. With both OCCT and Metro the GTX 680 Classified is 6C cooler than the reference GTX 680. Blowers aren’t known as the most effective coolers, so to reach the low 70s like this is definitely impressive.

Last but not least we have our look at noise levels. Starting with idle noise, the GTX 680 Classified makes no significant tradeoff for its size. Idle noise ends up being a hair higher than the reference GTX 680, but not noticeably so.

With load noise levels on the other hand we find ourselves once again impressed. The larger cooler and larger fan means that the GTX 680 Classified generates less noise than even the reference GTX 680, which was already a fairly quiet card for its performance level. We had expected EVGA to make some kind of temperature/noise tradeoff here but there’s none to be found. They have managed to lower both at the same time.

Ultimately it’s clear that while EVGA has engineered the GTX 680 Classified for heavy overclocking, they haven’t sacrificed reference performance to get there. When it comes to both temperatures and noise, the GTX 680 Classified is suitably superior to the reference GTX 680. The fact that EVGA did all of this with a blower makes it all the more impressive, since we typically only see gains like these by switching to an open air cooler. Granted these improvements are largely a consequence of the need to overbuild the card for the kind of heavy overclocking it was designed for, but if you do run at stock there are definitely benefits to be realized.

Portal 2, Battlefield 3, Starcraft II, Skyrim, Civilization V, & Compute Overclocked: Power, Temperature, & Noise
Comments Locked

75 Comments

View All Comments

  • HisDivineOrder - Tuesday, July 24, 2012 - link

    To be fair, AMD started the gouging with the 7970 series, its moderate boost over the 580 series, and its modest mark-up over that line.

    When nVidia saw what AMD had launched, they must have laughed and rubbed their hands together with glee. Because their mainstream part was beating it and it cost them a LOT less to make. So they COULD have passed those savings onto the customer and launched at nominal pricing, pressuring AMD with extremely low prices that AMD could probably not afford to match...

    ...or they could join with the gouging. They joined with the gouging. They knocked the price down by $50 and AMD's pricing (besides the 78xx series) has been in a freefall ever since.
  • CeriseCogburn - Tuesday, July 24, 2012 - link

    You people are using way too much tin foil, it's already impinged bloodflow to the brain from it's weight crimping that toothpick neck... at least the amd housefire heatrays won't further cook the egg under said foil hat.

    Since nVidia just barely has recently gotten a few 680's and 670's in stock on the shelves, how pray tell, would they produce a gigantic 7 billion transistor chip that it appears no forge, even the one in Asgard, could possibly currently have produced on time for any launch, up to and perhaps past even today ?

    See that's what I find so interesting. Forget reality, the Charlie D semi accurate brain fart smell is a fine delicacy for so many, that they will never stop inhaling.

    Wow.

    I'll ask again - at what price exactly was the 680 "midrange" chip supposed to land at ? Recall the GTX580 was still $499+ when amd launched - let's just say since nVidia was holding back according to the 20lbs of tinfoil you guys have lofted, they could have released GTX680 midrange when their GTX580 was still $499+ - right when AMD launched... so what price exactly was GTX680 supposed to be, and where would that put the rest of the lineups on down the price alley ?

    Has one of you wanderers EVER comtemplated that ? Where are you going to put the card lineups with GTX680 at the $250-$299 midrange in January ? Heck ... even right now, you absolute geniuses ?
  • natsume - Sunday, July 22, 2012 - link

    For that price, I prefer rather the Sapphire HD 7970 Toxic 6GB @ 1200Mhz
  • CeriseCogburn - Tuesday, July 24, 2012 - link

    Currently unavailable it appears.

    And amd fan boys have told us 7970 overclocks so well to (1300mhz they claim) so who cares.

    Toxic starts at 1100, and no amd fan boy would admit the run of the mill 7970 can't do that out of the box, as it's all we've heard now since January.

    It's nice seeing 6GB on a card though that cannot use even 3GB an maintain a playable frame rate at any resolution or settings, including 100+ Skyrim mods at once attempts.
  • CeriseCogburn - Tuesday, July 24, 2012 - link

    Sad how it loses so often to a reference GTX680 in 1920 and at triple monitor resolutions.

    http://www.overclockersclub.com/reviews/sapphire__...
  • Sabresiberian - Sunday, July 22, 2012 - link

    One good reason not to have it is the fact that software overclocking can sometimes be rather wonky. I can see Nvidia erring on the cautious side to protect their customers from untidy programs.

    EVGA is a company I want to love, but they are, in my opinion, one that "almost" goes the extra mile. This card is a good example, I think. Their customers expressed a desire for unlocked voltage and 4GB cards (or "more than 2GB"), and they made it for us.

    But they leave the little things out. Where do you go to find out what those little letters mean on the EVBot display? I'll tell you where I went - to this article. I looked in the EVBot manual, looked up the manual online to see if it was updated - it wasn't; scoured the website and forums, and no where could I find a breakdown of what the list of voltage settings meant from EVGA!

    I'm not regretting my purchase of this card; it is a very nice piece of hardware. It just doesn't have the 100% commitment to it a piece of hardware like this should.

    But then, EVGA, in my opinion, does at least as good as anybody, in my opinion. MSI is an excellent company, but they released their Lightning that was supposed to be over-voltable without a way to do it. Asus makes some of the best stuff in the business - if their manufacturing doesn't bungle the job and leave film that needs to be removed between heatsinks and what they should be attached to.

    Cards like this are necessarily problematic. To make them worth their money in a strict results sense, EVGA would have to guarantee they overclock to something like 1400MHz. If they bin to that strict of a standard, why don't they just factory overclock to 1400 to begin with?

    And, what's going to be the cost of a chip guaranteed to overclock that high? I don't know; I don't know what EVGA's current standards are for a "binning for the Classified" pass, but my guess is it would drive the price up, so that cost value target will be missed again.

    No, you can judge these cards strictly by value for yourself, that's quite a reasonable thing to do, but to be fair you must understand that some people are interested in getting value from something other than better frame rates in the games they are playing. For this card, that means the hours spent overclocking - not just the results, the results are almost beside the point, but the time spent itself. In the OC world that often means people will be disappointed in the final results, and it's too bad companies can't guarantee better success - but if they could, really what would be the point for the hard-core overclocker? They would be running a fixed race, and for people like that it would make the race not worth running.

    These cards aren't meant for the general-population overclocker that wants a guaranteed more "bang for the buck" out of his purchase. Great OCing CPUs like Nehalem and Sandy Bridge bring a lot of people into the overclocking world that expect to get great results easily, that don't understand the game it is for those who are actually responsible for discovering those great overclocking items, and that kind of person talks down a card like this.

    Bottom line - if you want a GTX 680 with a guaranteed value equivalent to a stock card, then don't buy this card! It's no more meant for you than a Mack truck is meant to be a family car. However, if you are a serious overclocker that likes to tinker and wants the best starting point, this may be exactly what you want.

    ;)
  • Oxford Guy - Sunday, July 22, 2012 - link

    Nvidia wasn't happy with the partners' designs, eh? Oh please. We all remember the GTX 480. That was Nvidia's doing, including the reference card and cooler. Their partners, the ones who didn't use the awful reference design, did Nvidia a favor by putting three fans on it and such.

    Then there's the lack of mention of Big Kepler on the first page of this review, even though it's very important for framing since this card is being presented as "monstrous". It's not so impressive when compared to Big Kepler.

    And there's the lack of mention that the regular 680's cooler doesn't use a vapor chamber like the previous generation card (580). That's not the 680 being a "jack of all trades and a master of none". That's Nvidia making an inferior cooler in comparison with the previous generation.
  • CeriseCogburn - Tuesday, July 24, 2012 - link

    I, for one, find the 3rd to the last paragraph of the 1st review page a sad joke.

    Let's take this sentence for isntance, and keep in mind the nVidia reference cooler does everything better than the amd reference:
    " Even just replacing the cooler while maintaining the reference board – what we call a semi-custom card – can have a big impact on noise, temperatures, and can improve overclocking. "

    One wonders why amd epic failure in comparison never gets that kind of treatment.

    If nVidia doesn't find that sentence I mentioned a ridiculous insult, I'd be surprised, because just before that, they got treated to this one: " NVIDIA’s reference design is a jack of all trades but master of none "

    I guess I wouldn't mind one bit if the statements were accompanied by flat out remarks that despite the attitude presented, amd's mock up is a freaking embarrassingly hot and loud disaster in every metric of comparison...

    I do wonder where all these people store all their mind bending twisted hate for nVidia, I really do.

    The 480 cooler was awesome because one could simply remove the gpu sink and still have a metal covered front of the pcb card and thus a better gpu HS would solve OC limits, which were already 10-15% faster than 5870 at stock and gaining more from OC than the 5870.

    Speaking of that, we're supposed to sill love the 5870, this sight claimed the 5850 that lost to 480 and 470 was the best card to buy, and to this day our amd fans proclaim the 5870 a king, compare it to their new best bang 6870 and 6850 that were derided for lack of performance when they came out, and now 6870 CF is some wonderkin for the fan boys.

    I'm pretty sick of it. nVidia spanked the 5000 series with their 400 series, then slammed the GTX460 down their throats to boot - the card all amd fans never mention now - pretending it never existed and still doesn't exist...
    It's amazing to me. All the blabbing stupid praise about amd cards and either don't mention nVidia cards or just cut them down and attack, since amd always loses, that must be why.
  • Oxford Guy - Tuesday, July 24, 2012 - link

    Nvidia cheaped out and didn't use a vapor chamber for the 680 as it did with the 580. AMD is irrelevant to that fact.

    The GF100 has far worse performance per watt, according to techpowerup's calculations than anything AMD released in 40nm. The 480 was very hot and very loud, regardless of whether AMD even existed in the market.

    AMD may have a history of using loud inefficient cooling, but my beef with the article is that Nvidia developed a more efficient cooler (580's vapor chamber) and then didn't bother to us it for the 680, probably to save a little money.
  • CeriseCogburn - Wednesday, July 25, 2012 - link

    The 680 is cooler and quieter and higher performing all at the same time than anything we've seen in a long time, hence "your beef" is a big pile of STUPID dung, and you should know it, but of course, idiocy never ends here with people like you.

    Let me put it another way for the faux educated OWS corporate "profit watcher" jack***: " It DOESN'T NEED A VAPOR CHAMBER YOU M*R*N ! "

    Hopefully that penetrates the inbred "Oxford" stupidity.

    Thank so much for being such a doof. I really appreciate it. I love encountering complete stupidity and utter idiocy all the time.

Log in

Don't have an account? Sign up now