OC: Power, Temperature, & Noise

Before wrapping things up, we wanted to quickly take a look at the overclocking potential of the GTX 660. As the first GK106 product GTX 660 should give us some idea as to how capable GK106 is at overclocking, though like GK104 we’re eventually at the mercy of NVIDIA’s locked voltages and limited power target control.

In its rawest form, GTX 660 will have two things going against it for overclocking. First and foremost, as the highest clocked GK106 part it’s already starting out at a fairly high clockspeed – 980MHz for reference cards, and upwards of 1050MHz for factory overclocked cards – so there may not be a great deal of overclocking headroom left to exploit. Furthermore because NVIDIA is keeping the power consumption of the card low (it needs to stay under 150W max), the maximum power target is the lowest we’ve seen for any GTX 600 card yet: it’s a mere 110%. As a result even if we can hit a large GPU clock offset, there may not be enough power headroom available to let the GPU regularly reach those speeds.

Memory overclocking on the other hand looks much better. With the same memory controllers and the same spec’d RAM as on the other high-end GTX 600 cards, there’s no reason to believe that the GTX 660 shouldn’t be able to hit equally high memory clocks, which means 6.5GHz+ is a reasonable goal.

GeForce GTX 660 Overclocking
  Ref GTX 660 EVGA GTX 660 SC Zotac GTX 660 Gigabyte GTX 660 OC
Shipping Core Clock 980MHz 1046MHz 993MHz 1033MHz
Shipping Max Boost Clock 1084MHz 1123MHz 1110MHz 1123MHz
Shipping Memory Clock 6GHz 6GHz 6GHz 6GHz
Shipping Max Boost Voltage 1.175v 1.175v 1.162v 1.175v
         
Overclock Core Clock 1080MHz 1096MHz 1093MHz 1083MHz
Overclock Max Boost Clock 1185MHz 1174MHz 1215MHz 1174MHz
Overclock Memory Clock 6.7GHz 6.9GHz 6.7GHz 6.5GHz
Overclock Max Boost Voltage 1.175v 1.175v 1.162v 1.175v

Throwing in our factory overclocked cards from our companion roundup, our core overclocking experience was remarkably consistent. The difference in the max boost clock between the slowest and fastest card was a mere 41MHz, with the Zotac card being a clear outlier compared to the rest of our cards. This comes as no great surprise since all of these launch cards are using the NVIDIA reference PCB, so there’s little room at this moment for overclocking innovation.

Memory overclocking is as volatile as ever, with a 400MHz spread between our best and worst cards. Again with the use of the reference PCB (and the same Samsung RAM), memory overclocking is entirely the luck of the draw.

For the moment at least GTX 660 overclocking looks to be on a level playing field due to all partners using the same PCB. For overclockers the choice of a card will come down to pricing, what cooler they prefer, and any preference in vendor.

The end result of all of this is that at best we’re seeing 100MHz overclocks (going by the max boost clock), which represents roughly a 10% overclock. Coupling this with a good memory overclock and the 10% increase in the power target will result in around a 10% increase in performance, which isn’t shabby but also is the same kind of shallow overclocking potential that we’ve seen on cards like the GTX 670 and GTX 660 Ti. All told the GTX 660 isn’t a poor overclocker – 10% more performance for free is nothing to sneeze at – but it’s also not going to enamor itself with hardware overclockers who like to chase 20% or more.

Moving on to our performance charts, we’re going to once again start with power, temperature, and noise, before moving on to gaming performance. Due to popular demand we’ll also be including overclocking results with just a 110% power target so that you can see the impact of adjusting the power target separately from the clock offsets.

With a 110% power target we should be seeing an 11W-14W increase in power consumption, which is indeed roughly what we’re seeing at the wall after accounting for PSU inefficiencies. In Metro this is just enough of a difference to erase most of the GTX 660’s power consumption advantage over the GTX 660 Ti, though the GTX 660 still draws marginally less power than the stock 7870. Meanwhile under OCCT the GTX 660 now draws more power than the 7870, but still is still drawing over 20W less than the stock GTX 660 Ti.

Our increased power consumption pushes temperatures up by another 2-3C. This is nothing a blower can’t handle, let alone an open-air cooler.

Interestingly enough, despite the increase in power consumption and temperatures, overclocking has almost no impact on noise. In the worst case scenario our GTX 660’s increased its fan speed by all of 2%, which increases noise by less than 1dB. As a result the amount of noise generated by the overclocked GTX 660 is practically identical to that generated by the stock GTX 660, and still below the reference 7870.

Power, Temperature, & Noise OC: Gaming Performance
POST A COMMENT

147 Comments

View All Comments

  • chizow - Sunday, September 16, 2012 - link

    But SB had no impact on 980X pricing, Intel is very deliberate in their pricing and EOL schedules so these parts do not lose much value before they gracefully go EOL. Otherwise, 980X still offered benefit over SB with 6 cores, something that was not replaced until SB-E over 1 year later. Even then, there was plenty of indication before Intel launched their SB-E platform to mitigate any sense of buyer's remorse.

    As for the German site you're critical of, you need to read German to be able to understand numbers and bar graphs? Not to mention Computerbase is internationally acclaimed as one of the best resources for PC related topics. I linked their sites because they were one of the first to use such easy to read performance summaries and even break them down by resolution and settings.

    If you prefer since you listed it, TechPowerUp has similar listings, they only copied the performance summaries of course after ground-breaking sites like Computerbase were using them for some years.

    http://www.techpowerup.com/reviews/NVIDIA/GeForce_...

    As you can see, even at your lower 1680x1050 resolution, the GTX 280 still handily outclasses the 4870 512MB by ~21%, so I guess there goes your theory? I've seen more recent benchmarks in games like Skyrim or any title with 4xAA or high-res textures where the gap widens as the 4870 chokes on the size of the requisite framebuffer.

    As for your own situation and you're wife's graphics card, TigerDirect has the same GTX 670 GC card I bought for $350 after rebate. Not as good as the deal I got but again, there are certainly new ones out there to be had for cheap. I'm personally going to wait another week or two to see if the GTX 660 price and its impact on AMD prices (another round of cuts expected next week, LOL) forces Nvidia to drop their prices as I also need to buy another GPU for the gf.
    Reply
  • Galidou - Sunday, September 16, 2012 - link

    And then once again you're off the subject, you send me relative performance of the gtx 280 to the 4870 2 years after their launch...... We were speaking of the rebate issued in 2008 for the performance it had back then. All the links I sent you were from 4 years ago and there's a reason to it and they're showing the 280 on average 10% difference in performance, and sometimes loosing big time.

    Sure the 1gb ram of the gtx 280 in the long run paid off. But we're speaking of 2008 situation that forced them to issue rebates..... which is 1 month after it's launch there's a part that performs ''similarly'' that costs less than HALF of it's price. computerbase is a good website, not the first time I see it, but it's the only one(and I never use only one website to base on the REAL average) that shows 20% difference in performance for a part that did still cost 115% more than the radeon 4870... 115 FREAKING % within one month!! nothing else to say.

    From a % point of view the 7970 and gtx 680 was a REALLY different fight.... and it was 3 month separating them which is something we commonly see in video card industry. while 115% more pricey parts that performs let's say 15% for your pleasure average from all websites than another part.....

    ''980X still offered benefit over SB with 6 cores''

    Never said it didn't, that's why I precised: ''A gamer buying an i7 980x 1 week prior to the sandy bridge launch''. For the gamer there was NO benefit at ALL, it even lost to the sandy bridge in games.

    But that's true, the 980x was out for a while unlike the gtx 280 who didn't have almost any time to keep it's amazing lead from last gen parts. The reason why they HAD to issue rebates, and the reason why I switched back to ATI from my 7800gt that died.

    The gtx 660 ti is a fine card, I'm just worried about the ROPs for future proofing, she'll keep the card a VERY long time so I regret not buying the 670 in the first place.
    Reply
  • chizow - Sunday, September 16, 2012 - link

    No I'm not off the subject, you're obviously basing your performance differences based on a specific low-resolution setting that was important to you at launch while I'm showing performance numbers of all resolutions that have only increased over time. The TPU link I provided was from a later review because as I already stated, Computerbase was one of the first sites to use these aggregate performance numbers, only later did other sites like TPU follow suit. The GTX 280 was always the best choice for enthusiasts running higher resolutions and more demanding AA and those differences only increased over time, just as I stated.

    Nvidia didn't feel they needed to drop the price any more than the initial cut on the 280 because after the cut they had a "similar" part to compete with the 4870 with their own GTX 260. Once again, AMD charged too little for their effort, but that has no bearing on the fact that the 280's launch price was *JUSTIFIED* based on relative performance to last-gen parts, unlike the situation with AMD's 28nm launch prices.

    As for the 680 and 7970? It just started driving home the fact the 7970 was grossly overpriced, as it offered 10-15% *more* performance than the 7970 at *10%* less price which began the tumble on AMD prices we see today. I've also been critical of the GTX 680 though, as it only offers ~35-40% increase in performance over GTX 580 at 100% of the price, which is still the worst increase in the last 10 years for Nvidia, but still obviously better than the joke AMD launched with Tahiti. 115% performance for 110% of the price compared to last-gen after 18 months is an absolute debacle.

    As for the 980X and SB, again the whole tangent is irrelevant. What would make it applicable would be if AMD launched a bulldozer variant that offered 90% of 980X performance at $400 price point and forced Intel to drop prices and issue rebates, but that obviously didn't happen. You're comparing factors that Intel has complete control over where in the case of the GTX 280, Nvidia obviously had no control over what AMD decided to do with the 4870.
    Reply
  • chizow - Sunday, September 16, 2012 - link

    There were numerous other important resolutions that took advantage of the 280's larger frame buffer, 1600x1200, 1920x1080 and 1920x1200. While they were obviously not as prevalent as they are now, they were certainly not uncommon for anyone shopping for a $300+ or $500 video card.

    As for the 7970 asking price, are you kidding? I had 10x as many AMD fanboys saying the 7970 price was justified at launch (not just Rarson), and where do you get 70% more perf? Its 50% being generous.

    So you got 150% performance for 150% of last-gen AMD price compared to 6970, how is that a good deal? Or similarly, you got 120% more performance for 110% the price compared to GTX 580, both last-gen parts.

    What you *SHOULD* expect is 150-200% performance for 100% of last-gen price, which is what the GTX 280 offered relative to 8800GTX, which is why I stated its pricing was justified.

    We've already covered the RV770, AMD could've easily priced it higher, even matched the GTX 260 price at $400 and still won, but they admittedly chose to go after market/mindshare instead after being beaten so badly by Nvidia since R600. Ever since then, they have clearly admitted their pricing mistake and have done everything in their power to slowly creep those prices upwards, culminating in the HUGE price increase we saw with Tahiti (see 150% price increase from 6970).
    Reply
  • Galidou - Sunday, September 16, 2012 - link

    They went for price related to the size of the Die. The radeon 4870 was more than half the size of the gtx 280 thus costing less than half to produce then justifying the value of the chip by the size and not the perdformance.

    We all know why this is happening now, AMD was battling to get back for competition against the top because they left this idea by building smaller die with the HD 3xxx, leaving the higher end to double chip boards, end parts below 400$ prices. So I knew it had to happen one day or another.

    So if you're really into making a wikipedia about pricing scheme for video cards and developping about it, go on. But in my opinion, as long as they do not sell us something worse than last gen for a higher price, I'll leave it for people to discern what they need. With all the competition in the market, it's hard to settle for anything that's a real winner, it's mostly based on personnal usage and money someone is willing to spend.

    If someone want to upgrade his video card for xxx$, only thing he have to do is look at the benchmark for the game(s) he plays. Not looking at the price of the last generation of video cards to see if the price is relevant to the price he pays now. Usually a good 30 minutes looking at 3-4 different web sites looking at the graphs reading a little will give you good indication without sending you in the dust by speaking and arguing about last gen stuff compared to what's out now...

    You speak like you're trying to justify what people should buy now because of how things were priced in the past..... Not working. You take your X bucks check out benchies, go out there and buy the card you want end of the freaking line. No video card is interesting you now, wait until something does. Stop living in the past and get to another chapter ffs...
    Reply
  • chizow - Sunday, September 16, 2012 - link

    Heh Wikipedia page? Obviously its necessary to set the record straight as revisionists like yourself are only going to emphasize the lowlights rather than the highlights.

    What you don't seem to understand is that transactions in a free market are not conducted in a vacuum, so the purchases of others do directly impact you if you are in the same market for these goods.

    Its important for reviewers to emphasize such important factors like historical prices and changes in performance, otherwise it reinforces and encourages poor pricing practices like we've seen from AMD. It just sets a bad precedent.

    Obviously the market has reacted by rejecting AMD's pricing scheme, and as a result we see the huge price drops we've seen over the last few months on their 28nm parts. All that's left is all the ill will from the AMD early adopters. You think all those people who got burned are OK with all the price drops, and that AMD won't have to deal with those repercussions later?

    You want to get dismissive and condescending, if all it took was a good 30 minutes looking at 3-4 different websites to get it, why haven't people like you and rarson gotten it yet?
    Reply
  • rarson - Tuesday, September 18, 2012 - link

    Yeah, it was rhetorical. It was also pointless and off-topic.

    "And even after the launch of 28nm, they still held their prices because there was no incentive or need to drop in price based on relative price and performance?"

    Your problem is that you just don't pay any goddamn attention. You have the attention span of a fruit fly. Let me refresh your memory. Kepler "launched" way back in March. All throughout April, the approximate availability of Kepler was zero. AMD didn't drop prices immediately because Kepler only existed in a few thousand parts. They dropped prices sometime around early May, when Kepler finally started appearing in decent quantities, because by then, the cards had already been on the market for FOUR MONTHS. See, even simple math escapes you.

    "You have no idea what you're talking about, stop typing."

    You're projecting and need to take your own advice.
    Reply
  • CeriseCogburn - Thursday, November 29, 2012 - link

    AMD's 7970/7950 series supply finally became "available" on average a few days before Kepler launched.
    LOL
    You said something about amnesia ?
    You rarson, are a sad joke.
    Reply
  • CeriseCogburn - Thursday, November 29, 2012 - link

    I love it when the penny pinching amd fanboy whose been whining about 5 bucks in amd "big win!" pricing loses their mind and their cool and starts yapping about new technology is expensive, achieving the highest amd price apologist marks one could hope for.
    LOL
    It's awesome seeing amd fanboys with zero cred and zero morality.
    The GTX570 made the 7850 and 7870 the morons choice from the very first date of release.
    You cannot expect the truth from the amd fans. It never happens. If there's any exception to that hard and fast rule, it's a mistake, soon to be corrected, with a vengeance, as the brainwashing and emotional baggage is all powerful.
    Reply
  • Galidou - Monday, September 17, 2012 - link

    While speaking about all that, pricing of the 4870 and 7970 do you really know everything around that, because it seems not when you are arguing, you just seem to put everything on the shoulder of a company not knowing any of the background.

    Do you know the price of the 4870 was already decided and it was in correlation with Nvidia's 9000 series performance. That the 4870 was supposed to compete against 400$ cards and not win and the 4850 supposed to compete against 300$ series card and not win. You heard right, the 9k series, not the GTX 2xx.

    The results even just before the coming out of the cards were already ''known''. The real things were quite different with the final product and last drivers enhancements. The performance of the card was actually a surprise, AMD never thought it was supposed to compete against the gtx 280, because they already knew the performance of the latter and that it was ''unnaittanable'' considering the size of the thing. Life is full of surprise you know.

    Do you know that after that, Nvidia sued AMD/ATI for price fixing asking for more communications between launch and less ''surprises''. Yes, they SUED them because they had a nice surprise... AMD couldn't play with prices too much because they were already published by the media and it was not supposed to compete against gtx2xx series. They had hoped that at 300$ it would ''compete'' against the gtx260 and not win against i thus justifying the price of the things at launch. And here you are saying it's a mistake launching insults at me, telling me I have a low intelligence and showing you're a know it all....

    Do you know that this price fixing obligation is the result of the pricing of the 7970, I bet AMD would of loved to price the latter at 400$ and could do it but it would of resulted in another war and more suing from Nvidia that wanted to price it's gtx 680 500$ 3 month after so to not break their consumers joy, they communicate A LOT more than before so everyone is happy, except now it hurts AMD because you compare to last gen and it makes things seems less of a deal. But with things back to normal we will be able to compare last gen after the refreshed radeon 7xxx parts and new gen after that.

    Nvidia the ''giant'' suing companies on the limit of ''extinction'', nice image indeed. Imagine the rich bankers starting to sue people in the streets, and they are the one you defend so vigorously. If they are that rich, do you rightly think the gtx 280 was well priced even considering it was double the last generation..

    It just means one thing, they could sell their card for less money but instead they sue the other company to take more money from our pockets, nice image.... very nice..... But that doesn't mean I won't buy an Nvidia card, I just won't defend them as vigorously as you do.... For every Goliath, we need a David, and I prefer David over Goliath.... even if I admire the strenght of the latter....
    Reply

Log in

Don't have an account? Sign up now