Power Consumption

As with all the major processor launches in the past few years, performance is nothing without a good efficiency to go with it. Doing more work for less power is a design mantra across all semiconductor firms, and teaching silicon designers to build for power has been a tough job (they all want performance first, naturally). Of course there might be other tradeoffs, such as design complexity or die area, but no-one ever said designing a CPU through to silicon was easy. Most semiconductor companies that ship processors do so with a Thermal Design Power, which has caused some arguments recently based presentations broadcast about upcoming hardware.

Yes, technically the TDP rating is not the power draw. It’s a number given by the manufacturer to the OEM/system designer to ensure that the appropriate thermal cooling mechanism is employed: if you have a 65W TDP piece of silicon, the thermal solution must support at least 65W without going into heat soak.  Both Intel and AMD also have different ways of rating TDP, either as a function of peak output running all the instructions at once, or as an indication of a ‘real-world peak’ rather than a power virus. This is a contentious issue, especially when I’m going to say that while TDP isn’t power, it’s still a pretty good metric of what you should expect to see in terms of power draw in prosumer style scenarios.

So for our power analysis, we do the following: in a system using one reasonable sized memory stick per channel at JEDEC specifications, a good cooler with a single fan, and a GTX 770 installed, we look at the long idle in-Windows power draw, and a mixed AVX power draw given by OCCT (a tool used for stability testing). The difference between the two, with a good power supply that is nice and efficient in the intended range (85%+ from 50W and up), we get a good qualitative comparison between processors. I say qualitative as these numbers aren’t absolute, as these are at-wall VA numbers based on power you are charged for, rather than consumption. I am working with our PSU reviewer, E.Fylladikatis, in order to find the best way to do the latter, especially when working at scale.

Nonetheless, here are our recent results for Kaby Lake at stock frequencies:

Power Delta (Long Idle to OCCT)

The Core i3-7350K, by virtue of its higher frequency, seems to require a good voltage to get up to speed. This is more than enough to go above and beyond the Core i5, which despite having more cores, is in the nicer part (efficiency wise) in the voltage/frequency curve. As is perhaps to be expected, the Core i7-2600K uses more power, having four cores with hyperthreading and a much higher TDP.

Overclocking

At this point I’ll assume that as an AnandTech reader, you are au fait with the core concepts of overclocking, the reason why people do it, and potentially how to do it yourself. The core enthusiast community always loves something for nothing, so Intel has put its high-end SKUs up as unlocked for people to play with. As a result, we still see a lot of users running a Sandy Bridge i7-2600K heavily overclocked for a daily system, as the performance they get from it is still highly competitive.

There’s also a new feature worth mentioning before we get into the meat: AVX Offset. We go into this more in our bigger overclocking piece, but the crux is that AVX instructions are power hungry and hurt stability when overclocked. The new Kaby Lake processors come with BIOS options to implement an offset for these instructions in the form of a negative multiplier. As a result, a user can stick on a high main overclock with a reduced AVX frequency for when the odd instruction comes along that would have previously caused the system to crash.

For our testing, we overclocking all cores under all conditions:

The overclocking experience with the Core i3-7350K matched that from our other overclockable processors - around 4.8-5.0 GHz. The stock voltage was particularly high, given that we saw 1.100 volts being fine at 4.2 GHz. But at the higher frequencies, depending on the quality of the CPU, it becomes a lot tougher maintain a stable system. With the Core i3, temperature wasn't really a feature here with our cooler, and even hitting 4.8 GHz was not much of a strain on the power consumption either - only +12W over stock. The critical thing here is voltage and stability, and it would seem that these chips would rather hit the voltage limit first (and our 1.4 V limit is really a bit much for a 24/7 daily system anyway). 

A quick browse online shows a wide array of Core i3-7350K results, from 4.7 GHz to 5.1 GHz. Kaby Lake, much like previous generations, is all about the luck of the draw - if you want to push it to the absolute limit.

Gaming: Shadow of Mordor Core i3-7350K vs Core i7-2600K: More MHz Cap'n!
Comments Locked

186 Comments

View All Comments

  • TheinsanegamerN - Friday, February 3, 2017 - link

    This chip is 6 years late. Back when sandy bridge was the newest chip, a dual core i3 was a super relevant choice for gaming, a quad core was overkill.

    Today, for gaming builds, a i5 chip is almost always a better choice, unless you only play games that are single threaded. And the i3 is more power hungry then locked quad cores.

    At $130, this would be a great choice, but ATM, the i3k is overpriced for what it offers for a modern system.
  • nathanddrews - Friday, February 3, 2017 - link

    I'd argue that with the introduction of this i3 K-variant and the new hyperthreaded Pentium, Intel just gave a lot of people a reason to not by an i5. The message from Intel seems to be this:

    "If you need great single-threaded performance with some mild multi-threaded, get the Pentium or i3. If you need great multi-threaded performance with great single-threaded, get an i7."

    I'd say they are preemptively stacking the product deck prior to the release of AMD Ryzen - offering entry-level gamers more options without diluting their HEDT status.
  • BedfordTim - Friday, February 3, 2017 - link

    In many of the games an i3-6100 offers effectively the same performance and is $50 cheaper. It isn't a case of the i-3750k offering great performance, so much as the games are not CPU limited. This points towards an even more expensive graphics card and the even cheaper CPU.
  • jayfang - Friday, February 3, 2017 - link

    Agree. Whatever about actual performance, it seems quite clear the cool factor of "unlocked" Ryzen's and joining the "overclocking community" is getting a pre-emptive strike from Intel.
  • Michael Bay - Saturday, February 4, 2017 - link

    OC never had a "cool factor".
  • eldakka - Sunday, February 5, 2017 - link

    In the late 90's into the early 00's, when people would travel for hours carting their PC to a LAN gaming event with 100's (or even thousands) of other people, having an OC'ed machine was indeed cool amongst that Geek crowd.

    /em remembers his dual celeron 300A's OC'ed to 450MHz (yes, that's Mega - not Giga - hertz).
  • drgoodie - Tuesday, February 7, 2017 - link

    I had a dual 366Mhz Celeron box OC'd to 550Mhz. It was cool back then.
  • dsraa - Tuesday, February 14, 2017 - link

    It was indeed and still is very very cool......I had been OC'ing my systems way back to the original Pentium 100, and then got a Celeron 300, OMG those were the days....If you don't think its cool, what the hell are you doing on Anandtech??!??!!!?
  • DLimmer - Wednesday, February 15, 2017 - link

    You may have missed the joke. This was a play on words; Overclocking produces more heat so it's "not cool."
  • DLimmer - Wednesday, February 15, 2017 - link

    OC is "hot, hot, hot"!

Log in

Don't have an account? Sign up now