In the Words of Jeremy Clarkson: POWEEEERRRRR

As with all the major processor launches in the past few years, performance is nothing without a good efficiency to go with it. Doing more work for less power is a design mantra across all semiconductor firms, and teaching silicon designers to build for power has been a tough job (they all want performance first, naturally). Of course there might be other tradeoffs, such as design complexity or die area, but no-one ever said designing a CPU through to silicon was easy. Most semiconductor companies that ship processors do so with a Thermal Design Power, which has caused some arguments recently based on performance presentations.

Yes, technically the TDP rating is not the power draw. It’s a number given by the manufacturer to the OEM/system designer to ensure that the appropriate thermal cooling mechanism is employed: if you have a 65W TDP piece of silicon, the thermal solution must support at least 65W without going into heat soak.  Both Intel and AMD also have different ways of rating TDP, either as a function of peak output running all the instructions at once, or as an indication of a ‘real-world peak’ rather than a power virus. This is a contentious issue, especially when I’m going to say that while TDP isn’t power, it’s still a pretty good metric of what you should expect to see in terms of power draw in prosumer style scenarios.

So for our power analysis, we do the following: in a system using one reasonable sized memory stick per channel at JEDEC specifications, a good cooler with a single fan, and a GTX 770 installed, we look at the long idle in-Windows power draw, and a mixed AVX power draw given by OCCT (a tool used for stability testing). The difference between the two, with a good power supply that is nice and efficient in the intended range (85%+ from 50W and up), we get a good qualitative comparison between processors. I say qualitative as these numbers aren’t absolute, as these are at-wall VA numbers based on power you are charged for, rather than consumption. I am working with our PSU reviewer, E.Fylladikatis, in order to find the best way to do the latter, especially when working at scale.

Nonetheless, here are our recent results for Kaby Lake at stock frequencies:

Power Delta (Long Idle to OCCT)

What amazes me, if anything, is how close the Core i7 and Core i3 parts are to their TDP in our measurements. Previously, such as with the Core i7-6700K and Core i7-4790K, we saw +20W on our system compared to TDP, but the Core i7-7700K is pretty much bang on at 90W (for a 91W rated part). Similarly, the Core i3-7350K is rated at 60W and we measured it at 55W. The Core i5-7600K is a bit different due to no hyperthreading meaning the AVX units aren’t loaded as much, but more on that in that review.

To clarify, our tests were performed on retail units. No engineering sample trickery here.

With power on the money, this perhaps mean that Intel is getting the voltages of each CPU to where they should be based on the quality of the silicon. In previous generations, Intel would over estimate the voltage needed in order to capture more CPUs within a given yield – however AMD has been demonstrating of late that it is possible to tailor the silicon more based on internal metrics. Either our samples are flukes, or Intel is doing something similar here.

With power consumption in mind, let’s move on to Overclocking, and watch some sand burn a hole in a PCB (hopefully not).


At this point I’ll assume that as an AnandTech reader, you are au fait with the core concepts of overclocking, the reason why people do it, and potentially how to do it yourself. The core enthusiast community always loves something for nothing, so Intel has put its high-end SKUs up as unlocked for people to play with. As a result, we still see a lot of users running a Sandy Bridge i7-2600K heavily overclocked for a daily system, as the performance they get from it is still highly competitive.

Despite that, the i7-7700K has somewhat of an uphill battle. As a part with a 4.5 GHz turbo frequency, if users are expecting a 20-30% increase for a daily system then we will be pushing 5.4-5.8 GHz, which for daily use with recent processors has not happened.

There’s also a new feature worth mentioning before we get into the meat: AVX Offset. We go into this more in our bigger overclocking piece, but the crux is that AVX instructions are power hungry and hurt stability when overclocked. The new Kaby Lake processors come with BIOS options to implement an offset for these instructions in the form of a negative multiplier. As a result, a user can stick on a high main overclock with a reduced AVX frequency for when the odd instruction comes along that would have previously caused the system to crash.

Because of this, we took our overclocking methods in two ways. First, we left the AVX Offset alone, meaning our OCCT mixed-AVX stability test got the full brunt of AVX power and the increased temperature/power reading there in.  We then applied a second set of overclocks with a -10 offset, meaning that at 4.5 GHz the AVX instructions were at 3.5 GHz. This did screw up some of our usual numbers that rely on the AVX part to measure power, but here are our results:

At stock, our Core i7-7700K ran at 1.248 volts at load, drawing 90 watts (the column marked ‘delta’), and saw a temperature of 79C using our 2kg copper cooling.

After this, we put the CPU on a 20x multiplier, set it to 1.000 volt (which didn’t work, so 1.100 volts instead), gave the load-line calibration setting to Level 1 (constant voltage on ASRock boards), and slowly went up in 100 MHz jumps. Every time the POV-Ray/OCCT stability tests failed, the voltage was raised 0.025V.

This gives a few interesting metrics. For a long time, we did not need a voltage increase: 1.100 volts worked as a setting all the way up to 4.2 GHz, which is what we’ve been expecting for a 14nm processor at such a frequency. From there the voltage starts increasing, but at 4.5 GHz we needed more voltage in a manual overclock to achieve stability than the CPU gave itself at stock frequency. So much for overclocking! It wasn’t until that 4.3-4.5 GHz set of results that the CPU started to get warm, as shown by the OCCT temperature values.

At 4.8 GHz, the Core i7-7700K passed POV-Ray with ease, however the 1.400 volts needed at that point were pushing the processor up to 95C during OCCT and its mixed AVX workload. At that point I decided to call an end to it, where the CPU was now drawing 122W from idle to load. The fact that it is only 122W is surprisingly low – I would have thought we would be nearing 160W at this point, other i7 overclockable processors at this level in the past.

The second set of results is with the AVX offset. This afforded stability at 4.8 GHz and 4.9 GHz, however at 5.0 GHz and 1.425 volts the CPU was clearly going into thermal recovery modes, as given by the lower scores in POV-Ray.

Based on what we’ve heard out in the ether, our CPU sample is somewhat average to poor in terms of overclocking performance. Some colleagues at the motherboard manufacturers are seeing 5.0 GHz at 1.3 volts (with AVX offset) although I’m sure they’re not talking in terms of a serious reasonable stability. 

Gaming: Shadow of Mordor Conclusions: The New Champion


View All Comments

  • Michael Bay - Saturday, January 14, 2017 - link

    Doom is not very CPU-intensive, and Battlefield is just drek. Reply
  • cpy - Wednesday, January 04, 2017 - link

    OMG AVX offset? What took them so long! I need AVX offset for my 4770K! That thing really mess with my OC. It feels like everytime AVX is used i need +0.1V to CPU. Reply
  • thekdub - Wednesday, January 04, 2017 - link

    Why is the 4790K "highly overclocked" in these benchmarks, while everything else appears to be running at stock speeds? I was hoping to get some insight as to where my stock 4790K stacks against this new chip as I'm on the fence about upgrading this year or waiting another generation. Instead, I'm stuck making inferences and trying to guess the difference in speed between a known variable and a vague statement of "highly overclocked". Reply
  • Achaios - Wednesday, January 04, 2017 - link

    @OP: It is a mistake that you did not explicitly state the OC of the chips you compared.

    The only thing I am interested right now is the release pf the 1080TI, so please ANANDTECH do your best to bring us some news (preview) regarding.

    CPU releases are completely uninteresting and by-and-large meaningless to anyone on a K Sandy Bridge CPU.

    I am on a Haswell, and don't see myself upgrading before 2020. Conversely, up to 2020, I would have upgraded my GPU twice.
  • Achaios - Wednesday, January 04, 2017 - link

    "by-and-large meaningless to anyone on a K Sandy Bridge CPU OR LATER". Please consider adding an EDIT function too. :-) Reply
  • just4U - Wednesday, January 04, 2017 - link

    Yeah.. there isn't really much of a jump is there? Buying a 7700K for people already on High End I7's made in the past few years is more about ... just wanting something newer, with maybe taking advantage of what some of the newest MBs offer.. I think anyway. Speed? Pfft... you already got that..

    Now if they released a mainstream I7 that had 6 cores/12 threads in the same package, well then.. that might be something of a different story. (..shrug) intels loss.
  • Badelhas - Wednesday, January 04, 2017 - link

    "Given that Intel has no competition, it is perhaps easy to roll out a new mainstream performance champion".

    This frase sums it up. And this is why I still own a SandyBridge i5 2500K overclocked to 4.4Ghz (a 33% increase) and find no need to upgrade. Intel is not even trying. It´s just SAD.
  • Thatguy97 - Wednesday, January 04, 2017 - link

    When ivy bridge came out I saw this coming Reply
  • willis936 - Wednesday, January 04, 2017 - link

    Wow I can't believe it: an entire generation of CPUs that change absolutely nothing about the microarchitecture design or transistor size. Why even bother rolling out new SKUs? Reply
  • willis936 - Wednesday, January 04, 2017 - link

    This would have been the perfect generation to add eDRAM to the K series. If only AMD would put out. Reply

Log in

Don't have an account? Sign up now