Power Consumption

Intel isn't really exploiting 22nm for significantly higher default or max turbo frequencies. While it does seem like you'll hit turbo frequencies more often with Ivy, most of what 22nm offers will be realized as power savings.

The data in the charts below is from our original 3770K preview, however I've also provided a table comparing the 3770K to the 2700K using Intel's own Z77 motherboard which is a bit more power hungry than our typical testbed:

Power Consumption Comparison
Intel DZ77GA-70K Idle Load (x264 2nd pass)
Intel Core i7 3770K 80.1W 146.4W
Intel Core i7 2700K 79.4W 177.6W

As you can see, there are no savings at idle and a reasonably significant improvement under load.

The same is echoed on our earlier chip in a more power efficient platform:

Power Consumption—Idle

Power Consumption—Load (x264 HD 3.03 2nd Pass)

I was also curious to see what power consumption would look like compared to other low-end GPUs. For these next results I used the 3770K alone, without a discrete card and measured power consumption. I then added in discrete GPUs from our HD 4000 comparisons and looked at both idle and load power while playing Metro 2033:

GPU Power Consumption—Idle

Obviously at idle it's impossible to beat the HD 4000, the GPU is largely stopped/gated when idle keeping power consumption to a minimum. Under load is where things get interesting:

GPU Power Consumption—Load (Metro 2033)

Ivy's GPU is much more power efficient than SNB's, however Intel still has a way to go before it starts to equal the power efficiency of modern discrete GPU architectures. Remember the HD 4000 is on Intel's 22nm process here while the GT 440 is built on TSMC's 40nm process.

Intel HD 4000 Performance: Compute & Synthetics Quick Sync Image Quality & Performance


View All Comments

  • cjb110 - Tuesday, April 24, 2012 - link

    Considering that on most games both the Xbox and PS 3tend to be sub 720p, the iGPU in Ivy Bridge is impressive. Has anybody compared the 3? Reply
  • tipoo - Tuesday, April 24, 2012 - link

    You have a unified shader GPU with similar performance to the x1900 series but more flexible, and something like a Geforce 7800 with some parts of the 7600 like lower ROPs and memory bandwidth...7 years later, if even an IGP didn't beat those, it would be pretty sad. Those were ~200gflop cards, todays top end is over 3000, a lower-mid range chip like this I would expect to be in the upper hundreds. Reply
  • gammaray - Tuesday, April 24, 2012 - link

    Why do Intel and AMD even started building IGPs in the first place?

    Why can't they just put a video card in every desktop and laptop?

    And if they continue making IGPs, whats their goal?

    Do they eventually wanna get rid of video card makers?
  • versesuvius - Tuesday, April 24, 2012 - link

    Better yet, why not just put the graphics on a chip like the CPU? That way the "upgrade path" is a lot clearer, not to mention "possible". It will also offer the possibility of having those chips in different flavors, for example good video transcoder or good gamer. There is room for that on the motherboard now that the north bridge is gone. Or they can review the IBM boards from 286 days and learn from their clean, and very efficient design.

    Unfortunately the financial model of the IT industry from a collective viewpoint entails throwing a lot of good hardware away just for a small advantage. Just the way many will throw away their HD3000 IGPs without having ever used it. The comparison is cruel but that should not be what distinguishes them from the toilet paper industry.

    As of late, Anand has taken to reminding us that technology has taken leaps ahead of our wishes and that we need time to absorb it. That is not the case. No wish is ever materialized. We only have to take whatever is offered and marvel at the only parameter that can be measured: speed. Less energy consumption is fine, but I suppose that comes with the territory (i.e. can Intel or AMD produce 22 nm chips that consume the same watts as 65 nm chips with the same number of transistors?).
  • tipoo - Wednesday, April 25, 2012 - link

    Cost, size, power draw. All reduced by putting everything on one chip. I'm not sure if AMD wants to get rid of discreet graphics cards considering that's their one profitable division, but Intel sure does :) Reply
  • klmccaughey - Tuesday, April 24, 2012 - link

    I've tried every setup possible and have never got quicksync to work at all. It won't even work with discrete graphics enabled and my monitor hooked up to the intel chip on my Z68 board.

    I have tried mediacoder (error 14), media converter 7, MEdia Espresso. I have downloaded the media SDK, I have tried the new FFmthingy from the intel engineer. Nothing, nada. It has never ever worked. AMD media converter will convert a few limited formats that went out of fashion 5 years ago (of all my 1.5TB of video the only thing it would touch was old episodes of Becker).

    All in all I have got nothing whatsoever from any video accelerated encoding and I have always had to go back to my tried and trusted handbrake.

    I don't think it actually works - I've never heard of anyone getting it working and the forums on mediacoder are full of people who have given up.
  • klmccaughey - Tuesday, April 24, 2012 - link

    *disabled (for discrete graphics) Reply
  • JarredWalton - Tuesday, April 24, 2012 - link

    What input format are you using? I've only tried it on laptops, and I've done MOV input from my Nikon D3100 camera with no issues whatsoever. I've also done a WMV input file (the sample Wildlife.WMV file from Windows 7) and it didn't have any trouble. If you're trying to do a larger video, that might be an issue, or it might just be a problem with the codec used on the original video. Reply
  • dealcorn - Tuesday, April 24, 2012 - link

    Intel has big eyes for the workstation graphics market and has had some success at the bottom of this market. Will IB's IGD advances enhance Intel's access to this market? Reply
  • chizow - Tuesday, April 24, 2012 - link

    Hope you're not counting yourself, you've proven long ago your opinion isn't worth paying attention to. Reply

Log in

Don't have an account? Sign up now