In the Words of Jeremy Clarkson: POWEEEERRRRR

As with all the major processor launches in the past few years, performance is nothing without a good efficiency to go with it. Doing more work for less power is a design mantra across all semiconductor firms, and teaching silicon designers to build for power has been a tough job (they all want performance first, naturally). Of course there might be other tradeoffs, such as design complexity or die area, but no-one ever said designing a CPU through to silicon was easy. Most semiconductor companies that ship processors do so with a Thermal Design Power, which has caused some arguments recently based on performance presentations.

Yes, technically the TDP rating is not the power draw. It’s a number given by the manufacturer to the OEM/system designer to ensure that the appropriate thermal cooling mechanism is employed: if you have a 65W TDP piece of silicon, the thermal solution must support at least 65W without going into heat soak.  Both Intel and AMD also have different ways of rating TDP, either as a function of peak output running all the instructions at once, or as an indication of a ‘real-world peak’ rather than a power virus. This is a contentious issue, especially when I’m going to say that while TDP isn’t power, it’s still a pretty good metric of what you should expect to see in terms of power draw in prosumer style scenarios.

So for our power analysis, we do the following: in a system using one reasonable sized memory stick per channel at JEDEC specifications, a good cooler with a single fan, and a GTX 770 installed, we look at the long idle in-Windows power draw, and a mixed AVX power draw given by OCCT (a tool used for stability testing). The difference between the two, with a good power supply that is nice and efficient in the intended range (85%+ from 50W and up), we get a good qualitative comparison between processors. I say qualitative as these numbers aren’t absolute, as these are at-wall VA numbers based on power you are charged for, rather than consumption. I am working with our PSU reviewer, E.Fylladikatis, in order to find the best way to do the latter, especially when working at scale.

Nonetheless, here are our recent results for Kaby Lake at stock frequencies:

Power Delta (Long Idle to OCCT)

What amazes me, if anything, is how close the Core i7 and Core i3 parts are to their TDP in our measurements. Previously, such as with the Core i7-6700K and Core i7-4790K, we saw +20W on our system compared to TDP, but the Core i7-7700K is pretty much bang on at 90W (for a 91W rated part). Similarly, the Core i3-7350K is rated at 60W and we measured it at 55W. The Core i5-7600K is a bit different due to no hyperthreading meaning the AVX units aren’t loaded as much, but more on that in that review.

To clarify, our tests were performed on retail units. No engineering sample trickery here.

With power on the money, this perhaps mean that Intel is getting the voltages of each CPU to where they should be based on the quality of the silicon. In previous generations, Intel would over estimate the voltage needed in order to capture more CPUs within a given yield – however AMD has been demonstrating of late that it is possible to tailor the silicon more based on internal metrics. Either our samples are flukes, or Intel is doing something similar here.

With power consumption in mind, let’s move on to Overclocking, and watch some sand burn a hole in a PCB (hopefully not).

Overclocking

At this point I’ll assume that as an AnandTech reader, you are au fait with the core concepts of overclocking, the reason why people do it, and potentially how to do it yourself. The core enthusiast community always loves something for nothing, so Intel has put its high-end SKUs up as unlocked for people to play with. As a result, we still see a lot of users running a Sandy Bridge i7-2600K heavily overclocked for a daily system, as the performance they get from it is still highly competitive.

Despite that, the i7-7700K has somewhat of an uphill battle. As a part with a 4.5 GHz turbo frequency, if users are expecting a 20-30% increase for a daily system then we will be pushing 5.4-5.8 GHz, which for daily use with recent processors has not happened.

There’s also a new feature worth mentioning before we get into the meat: AVX Offset. We go into this more in our bigger overclocking piece, but the crux is that AVX instructions are power hungry and hurt stability when overclocked. The new Kaby Lake processors come with BIOS options to implement an offset for these instructions in the form of a negative multiplier. As a result, a user can stick on a high main overclock with a reduced AVX frequency for when the odd instruction comes along that would have previously caused the system to crash.

Because of this, we took our overclocking methods in two ways. First, we left the AVX Offset alone, meaning our OCCT mixed-AVX stability test got the full brunt of AVX power and the increased temperature/power reading there in.  We then applied a second set of overclocks with a -10 offset, meaning that at 4.5 GHz the AVX instructions were at 3.5 GHz. This did screw up some of our usual numbers that rely on the AVX part to measure power, but here are our results:

At stock, our Core i7-7700K ran at 1.248 volts at load, drawing 90 watts (the column marked ‘delta’), and saw a temperature of 79C using our 2kg copper cooling.

After this, we put the CPU on a 20x multiplier, set it to 1.000 volt (which didn’t work, so 1.100 volts instead), gave the load-line calibration setting to Level 1 (constant voltage on ASRock boards), and slowly went up in 100 MHz jumps. Every time the POV-Ray/OCCT stability tests failed, the voltage was raised 0.025V.

This gives a few interesting metrics. For a long time, we did not need a voltage increase: 1.100 volts worked as a setting all the way up to 4.2 GHz, which is what we’ve been expecting for a 14nm processor at such a frequency. From there the voltage starts increasing, but at 4.5 GHz we needed more voltage in a manual overclock to achieve stability than the CPU gave itself at stock frequency. So much for overclocking! It wasn’t until that 4.3-4.5 GHz set of results that the CPU started to get warm, as shown by the OCCT temperature values.

At 4.8 GHz, the Core i7-7700K passed POV-Ray with ease, however the 1.400 volts needed at that point were pushing the processor up to 95C during OCCT and its mixed AVX workload. At that point I decided to call an end to it, where the CPU was now drawing 122W from idle to load. The fact that it is only 122W is surprisingly low – I would have thought we would be nearing 160W at this point, other i7 overclockable processors at this level in the past.

The second set of results is with the AVX offset. This afforded stability at 4.8 GHz and 4.9 GHz, however at 5.0 GHz and 1.425 volts the CPU was clearly going into thermal recovery modes, as given by the lower scores in POV-Ray.

Based on what we’ve heard out in the ether, our CPU sample is somewhat average to poor in terms of overclocking performance. Some colleagues at the motherboard manufacturers are seeing 5.0 GHz at 1.3 volts (with AVX offset) although I’m sure they’re not talking in terms of a serious reasonable stability. 

Gaming: Shadow of Mordor Conclusions: The New Champion
Comments Locked

125 Comments

View All Comments

  • Toss3 - Tuesday, January 3, 2017 - link

    TBH they shouldn't just post mins, but decent FCAT analyses like the ones over on Guru3d.com
  • User.Name - Tuesday, January 3, 2017 - link

    Well that involves a lot more hardware and time to record/analyze the results, which is why I suggested looking at minimum framerates. But you're right that would be a good improvement too.
  • edzieba - Tuesday, January 3, 2017 - link

    "For one thing, average framerates are meaningless when doing CPU tests. You need to be looking at minimum framerates."

    Framerate needs to be dropped entirely. Instead, frame render times (specifically range and variance) give a better picture of perceived 'responsiveness', as well as render times being convertible to an FPS value (though not vice versa).
  • Alexvrb - Tuesday, January 3, 2017 - link

    Agreed, when reviewing CPUs it would stand to reason that you'd want to use games that tax the CPU.
  • Notmyusualid - Friday, January 6, 2017 - link

    Your second link there was especially interesting - thats why I went for more cores than four.

    My 14C/28T Xeon has to feed 2x 1070 FTWs. I don't think quad core & multi-gpu are that great together, in my experience.

    For all the talk about games don't use more than 'x' cores, I see my cores / threads nicely loaded up for many games. Even MW3 shows activity over 12 Threads, however small, and thats old now.

    I just got a 6950X for a song, and the scouser seller backed out on me AFTER I paid. So I get to wait 5 to 7 working days for my money back (thanks PayPal), and I won't get to see how much frequency would have affected my everyday computing. I won't be paying north of 1400 GBP for one, that I can tell you.
  • Mondozai - Tuesday, January 3, 2017 - link

    Who the fuck is testing with GTX 980@1080p? It should be a GTX 1080@1080p because as games' visual demands go up progressively, it will show how the processor ages. This review is useless from that regard.

    Go to Sweclockers or any other website for a real review. AT has fallen so fucking much it's hilarious.
  • Gasaraki88 - Tuesday, January 3, 2017 - link

    Yeah... surprisingly Tom's Hardware has really indepth reviews now a days just like the olden times. Considering that Microsoft has said that OSes lower than Windows 10 will not be supported on Kaby Lake, i'm surprised they are still using Windows 7 to to their tests.
  • Shadowmaster625 - Tuesday, January 3, 2017 - link

    There are lots of people that use a 980 with something like a 2500k or 2600k and might be wondering what a new cpu would do for them.
  • dakishimesan - Tuesday, January 3, 2017 - link

    Also using the same testing setup allows the results to be directly comparable to previous chips. They already mentioned in one of the articles regarding Kaby Lake (I think it was the i5 review) that they will be rolling out a new testbed and testing suite in February.
  • BrokenCrayons - Tuesday, January 3, 2017 - link

    From the Test Bed and Setup page:

    "This is also typically run at JEDEC subtimings where possible."

    -and-

    "Our testing methodology is ‘out-of-the-box’, with the latest public BIOS installed and XMP enabled, and thus subject to the whims of this feature."

    After reading those two lines, I really don't know what Anandtech's memory settings were like for this article.

Log in

Don't have an account? Sign up now