Homework: How Turbo Mode Works

AMD and Intel both figured out the practical maximum power consumption of a desktop CPU. Intel actually discovered it first, through trial and error, in the Prescott days. At the high end that's around 130W, for the upper mainstream market that's 95W. That's why all high end CPUs ship with 120 - 140W TDPs.

Regardless of whether you have one, two, four, six or eight cores - the entire chip has to fit within that power envelope. A single core 95W chip gets to have a one core eating up all of that power budget. This is where we get very high clock speed single core CPUs from. A 95W dual core processor means that individually the cores have to use less than the single 95W processor, so tradeoffs are made: each core runs at a lower clock speed. A 95W quad core processor requires that each core uses less power than both a single or dual core 95W processor, resulting in more tradeoffs. Each core runs at a lower clock speed than the 95W dual core processor.

The diagram below helps illustrate this:

  Single Core Dual Core Quad Core Hex Core
TDP
Tradeoff

 

The TDP is constant, you can't ramp power indefinitely - you eventually run into cooling and thermal density issues. The variables are core count and clock speed (at least today), if you increase one, you have to decrease the other.

Here's the problem: what happens if you're not using all four cores of the 95W quad core processor? You're only consuming a fraction of the 95W TDP because parts of the chip are idle, but your chip ends up being slower than a 95W dual core processor since its clocked lower. The consumer has to thus choose if they should buy a faster dual core or a slower quad core processor.

A smart processor would realize that its cores aren't frequency limited, just TDP limited. Furthermore, if half the chip is idle then the active cores could theoretically run faster.

That smart processor is Lynnfield.

Intel made a very important announcement when Nehalem launched last year. Everyone focused on cache sizes, performance or memory latency, but the most important part of Nehalem was far more subtle: the Power Gate Transistor.

Transistors are supposed to act as light switches - allowing current to flow when they're on, and stopping the flow when they're off. One side effect of constantly reducing transistor feature size and increasing performance is that current continues to flow even when the transistor is switched off. It's called leakage current, and when you've got a few hundred million transistors that are supposed to be off but are still using current, power efficiency suffers. You can reduce leakage current, but you also impact performance when doing so; the processes with the lowest leakage, can't scale as high in clock speed.

Using some clever materials engineering Intel developed a very low resistance, low leakage, transistor that can effectively drop any circuits behind it to near-zero power consumption; a true off switch. This is the Power Gate Transistor.

On a quad-core Phenom II, if two cores are idle, blocks of transistors are placed in the off-state but they still consume power thanks to leakage current. On any Nehalem processor, if two cores are idle, the Power Gate transistors that feed the cores their supply current are turned off and thus the two cores are almost completely turned off - with extremely low leakage current. This is why nothing can touch Nehalem's idle power:

Since Nehalem can effectively turn off idle cores, it can free up some of that precious TDP we were talking about above. The next step then makes perfect sense. After turning off idle cores, let's boost the speed of active cores until we hit our TDP limit.

On every single Nehalem (Lynnfield included) lies around 1 million transistors (about the complexity of a 486) whose sole task is managing power. It turns cores off, underclocks them and is generally charged with the task of making sure that power usage is kept to a minimum. Lynnfield's PCU (Power Control Unit) is largely the same as what was in Bloomfield. The architecture remains the same, although it has a higher sampling rate for monitoring the state of all of the cores and demands on them.

The PCU is responsible for turbo mode.

New Heatsinks and Motherboards Lynnfield's Turbo Mode: Up to 17% More Performance
Comments Locked

343 Comments

View All Comments

  • Jamahl - Tuesday, September 8, 2009 - link

    Digusting! How much money did intel bung you for this disgrace?
  • strikeback03 - Tuesday, September 8, 2009 - link

    How is that disgusting? It is the stock configuration of the processor. They are not doing all this testing as an e-pissing contest of who has better performance per clock, it is a comparison of retail products in real-world applications. If (and according to the review, when) AMD has something similar, I'd imagine they will test with that turned on too.
  • Jamahl - Tuesday, September 8, 2009 - link

    Why not benchmark the Phenom 2 with fusion for gaming anand???
  • Gary Key - Tuesday, September 8, 2009 - link

    "Why not benchmark the Phenom 2 with fusion for gaming anand??? "

    Have you actually tried using Fusion with Windows 7 x64? It is a total mess. I will be happy to show some results with it, most will say DNF, but that might not make you happy. ;) That said, AMD is working on it, especially trying to get it to play nice with AOD.

    In the meantime, here is the current list of items to watch out for - http://game.amd.com/us-en/drivers_fusion.aspx?p=3&...">http://game.amd.com/us-en/drivers_fusion.aspx?p=3&... .
  • Jamahl - Tuesday, September 8, 2009 - link

    What? It's AMD's fault that an unreleased 64-bit os is causing issues with their software?

    How many people are using Window 7 64 bit who visit this website?
  • Gary Key - Tuesday, September 8, 2009 - link

    It is just as bad in 64-bit Vista. I imagine a fair amount of people that visit the site are using the RTM version of Win7.
  • Jamahl - Tuesday, September 8, 2009 - link

    were you benchmarking a processor at 2.66 gigahertz or a processor at 3.2 gigahertz?
  • Gary Key - Tuesday, September 8, 2009 - link

    Do you and SnakeOil live in the same house or is the IP address just that similar? LOL.. The processors were benched as they come out of the box. For Lynnfield that means turbo was on as we stated in the article and here in the comments. For Bloomfield, that also means turbo was on, just as it comes out of the retail box.

    If you check all the other reviews on the web at the main sites, everyone tested with turbo on in the primary benchmarks. So I guess you can say there is a huge conspiracy between us to actually utilize the processors as Intel intended for the users.

    Apparently, we all failed at covering it up, so congratulations on discovering the Freeturbomasons. A now not so secret fraternal organization bent on world domination through the use of turbo frequencies inside processors carrying the blue "i" logo.

    For the AMD Phenom II x4 series, they were benched with all cores enabled just as they come out of the box, even though you can disable each core in the BIOS just like you can disable turbo on the i7/i5. I guess to make things fair, we should disable the cores on the 965 BE as having that "feature" turned on is cheating.

    Anyway, thanks for making my day, I needed some much deserved laughter. :)
  • Chlorus - Tuesday, September 8, 2009 - link

    Seeing as how Fusion isn't even out yet, that would be hard to do...go troll somewhere else. What is with the idiots coming out of the woodwork on this post? You've got the standard fanboys, as well as insecure LGA-1366 owners who feel the need to defend their purchase, and insecure purchasers of 1156 products who are afraid their choice might be bested in a benchmark somewhere.
  • Roland00 - Tuesday, September 8, 2009 - link

    Would you please rerun page 9 with an overclocked 975 and the 870. I am wondering how much the difference will grow when the gpus are fed more information due to the faster cpus. Something like 3.8 to 4.0 ghz on both cpus with turbo off (a good overclock yet not in the unreasonable area)

    If you are investing 400+ dollars in gpus, and you are building it yourself you are probably going to overclock.

Log in

Don't have an account? Sign up now