Power Efficiency: Putting A Lid On Fiji

Last, but certainly not least, before ending our tour of the Fiji GPU we need to talk about power.

Power is, without question, AMD’s biggest deficit going into the launch of R9 Fury X. With Maxwell 2 NVIDIA took what they learned from Tegra and stepped up their power efficiency in a major way, which allowed them to not only outperform AMD’s Hawaii GPUs, but to do so while consuming significantly less power. In this 4th year of 28nm the typical power efficiency gains that come from a smaller process are another year off, so both AMD and NVIDIA have needed to invest in power efficiency at an architectural level for 28nm.

The power situation on Fiji in turn is a bit of a mixed bag, but largely positive for AMD. The good news here is that AMD has indeed taken power efficiency very seriously for Fiji, and in turn has made a number of changes to boost power efficiency and bring it more in line with what NVIDIA has achieved, leading to R9 Fury X being rated for the same 275W Typical Board Power (TBP) as the R9 390X, and just 25W more than R9 290X. The bad news, as we’ll see in our benchmarks, is that AMD won’t quite meet NVIDIA’s power efficiency numbers; but they had a significant gap to close and they have done a very admirable job in coming this far.


A basic implementation of clock gating. Image Source: Mahesh Dananjaya - Clock Gating

So what has AMD done to better control power consumption? Perhaps the biggest improvement here is that AMD has improved their clock gating technology by implementing multi-level clock gating throughout the chip, in order to better cut off parts of the GPU that are not in use and thereby reduce their power consumption. With clock gating the clock signal is turned off to a functional unit, leaving said unit turned on but not doing any work or switching transistors, which allows for significant power savings even without turning said unit off via power gating (and without the time-cost of bringing it back up). Even turning off a functional unit for a couple of dozen cycles, say while the geometry engines wait on the shaders to complete their work, brings down power consumption in load states as well as the more obvious idle states.

Meanwhile AMD has taken some lessons from their recently-launched Carrizo APU – which is also based on GCN 1.2 and designed around improving power efficiency – in order to boost power efficiency for Fiji. What AMD has disclosed to us is that the power flow for Fiji is based on what they’ve learned from the APUs, which in turn has allowed AMD to better control/map several aspects of Fiji’s voltage needs for better operation. Voltage adaptive operation, for example, allows AMD to use a lower voltage that’s closer to Fiji’s real voltage needs, reducing the amount of power wasted by operating Fiji at a voltage higher than it needs to operate. VAO essentially uses thinner voltage safeguards to accomplish this, pulling back the clockspeed momentarily if the supply voltage drops below Fiji’s operational requirements.

Similarly, AMD has also put a greater focus on the binning process to better profile chips before they leave the factory. This includes a tighter voltage/frequency curve (enabled by VSO) to cut down on wasted voltage, but it also includes new processes to better identify and compensate for leakage on a per-chip basis. Leakage is the eternal scourge for chip designers, and with 28nm it has only gotten worse. Even with the now highly-mature process, leakage can still consume (or rather allows to escape) quite a bit of power if not controlled for. This is also one of the reasons that FinFETs will be so important in TSMC’s next-generation 16nm manufacturing process, as FinFETs cut down on leakage.

AMD’s third power optimization comes from the use of HBM, which along with its greater bandwidth also offers lower power consumption relative to even the 512-bit wide 5Gbps GDDR5 memory bus AMD used on R9 290X. On R9 290X AMD estimates that memory power consumption was 15-20% (37-50W) of their 250W TDP, largely due to the extensive PHYs required to handle the complicated bus signaling of GDDR5.

By AMD’s own metrics, HBM delivers better than 3x the bandwidth per watt of GDDR5 thanks to the simpler bus and lower operating voltage of 1.3v. Given that AMD opted to spend some of their gains on increasing memory bandwidth as opposed to just power savings, the final power savings aren’t 3X, but by AMD’s estimates the amount of power they’re spending on HBM is around 15-20W, which has saved R9 Fury X around 20-30W of power relative to R9 290X. These are savings that AMD can simply keep, or as in the case of R9 Fury X, spend some of them on giving the card more power headroom for higher performance.

The final element in AMD’s plan to improve energy efficiency on Fiji is a bit more brute-force but none the less important, and that’s temperature controls. As our long-time readers may recall from the R9 290 (Hawaii) launch in 2013, with the reference R9 290X AMD picked a higher temperature gradient over lower operating temperatures in order to maximize the cooling efficiency of their reference cooler. The tradeoff was that they had to accept higher leakage as a result of the higher temperatures, though as AMD’s second-generation 28nm product they felt they had leakage under control.


An example of the temperature versus power consumption principle on an Intel Core i7-2600K. Image Credit: AT Forums User "Idontcare"

But with R9 Fury X in particular and its large, overpowered closed loop liquid cooler, AMD has gone in the opposite direction. AMD no longer needs to rely on temperature gradients to boost cooler performance, and as a result they’ve significantly dialed down the average operating temperature of the Fiji GPU in R9 Fury X in order to further mitigate leakage and reduce overall power consumption. Whereas R9 290X would go to 95C, R9 Fury X essentially tops out at 65C, as that’s the point after which it will start ramping up the fan speed rather than allow the GPU to get any warmer. This 30C reduction in GPU temperature undoubtedly saves AMD some power on leakage, and while the precise amount isn’t disclosed, as leakage is a non-linear relationship the results could be rather significant for Fiji.

To put this to the test, we did a bit of experimenting with Crysis 3 to look at power consumption over time. While the R9 Fury X doesn’t allow us to let it run any warmer, we are able to monitor power consumption at the start of the benchmark run when the card has just left idle at around 40C, and compare it to when the run is terminated at 65C.

Crysis 3 Power Consumption
  GPU Temperature Power Consumption @ Wall
Start Of Run 40C 388W
15 Minutes, Equilibrium 65C 408W

What we find is that Fury’s power consumption increases by 20W at the wall between the start and the end, and this despite the fact that the scene is unchanged, the framerate is unchanged, and the CPU usage is unchanged. The roughly 18W difference after the PSU comes from the video card, its power consumption increasing with the GPU temperature and a slighter bump from the approximately 100RPM increase in fan speeds. Had AMD allowed Fury X to go to 83C (the same temperature as the GTX 980 Ti), it likely would have been closer to a 300W TBP card, and 95C would be higher yet, indicating just how important temperature controls are for AMD in order to get the best energy efficiency as is possible out of Fiji.

Last, but not least on the subject of power consumption, we need to quickly discuss the driver situation. AMD tells us that for R9 Fury X they were somewhat conservative on how they adjusted clockspeeds, favoring performance over power savings. As a result R9 Fury X doesn’t downclock as often as it could, staying at 1050MHz more often, practically running at maximum clockspeeds whenever a real load is put on it so that it offers the best performance possible should it be needed.

What AMD is telling us right now is that future drivers for Fiji products will be better tuned than what we’re seeing on Fury X, such that those parts won’t run at their full load clocks quite so aggressively. The nature of this claim invites a wait-and-see approach, but based on what we’re seeing with R9 Fury X so far, it’s not an unrealistic goal for AMD. More aggressive power control and throttling not only improves power consumption under light loads, but it also stands to improve power consumption under full load. GCN can switch voltages as quickly as 10 microseconds, or hundreds of times in the span of time it takes for a GPU to render a single frame, so there are opportunities there for the GPU to take short breaks whenever a bottleneck is occurring in the rendering process and the card’s full 1050MHz isn’t required for a thousand cycles or so.

On that note, AMD has also told us to keep our eyes peeled for what they deliver with the R9 Fury (vanilla). Without its closed loop liquid cooler, the R9 Fury will not have the same overbuilt cooling apparatus available, and as a result it sounds like AMD will take a more aggressive approach in-line with the above to better control power consumption.

Fiji’s Layout High Bandwidth Memory: Wide & Slow Makes It Fast
Comments Locked

458 Comments

View All Comments

  • chizow - Sunday, July 5, 2015 - link

    @piiman - I guess we'll see soon enough, I'm confident it won't make any difference given GPU prices have gone up and up anyways. If anything we may see price stabilization as we've seen in the CPU industry.
  • medi03 - Sunday, July 5, 2015 - link

    Another portion of bulshit from nVidia troll.

    AMD never ever had more than 25% of CPU share. Doom to Intel, my ass.
    Even in Prescott times Intell was selling more CPUs and for higher price.
  • chizow - Monday, July 6, 2015 - link

    @medi03 AMD was up to 30% a few times and they did certainly have performance leadership at the time of K8 but of course they wanted to charge anyone for the privilege. Higher price? No, $450 for entry level Athlon 64, much more than what they charged in the past and certainly much more than Intel was charging at the time going up to $1500 on the high end with their FX chips.
  • Samus - Monday, July 6, 2015 - link

    Best interest? Broken up for scraps? You do realize how important AMD is to people who are Intel\NVidia fans right?

    Without AMD, Intel and NVidia are unchallenged, and we'll be back to paying $250 for a low-end video card and $300 for a mid-range CPU. There would be no GTX 750's or Pentium G3258's in the <$100 tier.
  • chizow - Monday, July 6, 2015 - link

    @Samus, they're irrelevant in the CPU market and have been for years, and yet amazingly, prices are as low as ever since Intel began dominating AMD in performance when they launched Core 2. Since then I've upgraded 5x and have not paid more than $300 for a high-end Intel CPU. How does this happen without competition from AMD as you claim? Oh right, because Intel is still competing with itself and needs to provide enough improvement in order to entice me to buy another one of their products and "upgrade".

    The exact same thing will happen in the GPU sector, with or without AMD. Not worried at all, in fact I'm looking forward to the day a company with deep pockets buys out AMD and reinvigorates their products, I may actually have a reason to buy AMD (or whatever it is called after being bought out) again!
  • Iketh - Monday, July 6, 2015 - link

    you overestimate the human drive... if another isn't pushing us, we will get lazy and that's not an argument... what we'll do instead to make people upgrade is release products in steps planned out much further into the future that are even smaller steps than how intel is releasing now
  • silverblue - Friday, July 3, 2015 - link

    I think this chart shows a better view of who was the underdog and when:

    http://i59.tinypic.com/5uk3e9.jpg

    ATi were ahead for the 9xxx series, and that's it. Moreover, NVIDIA's chipset struggles with Intel were in 2009 and settled in early 2011, something that would've benefitted NVIDIA far more than Intel's settlement with AMD as it would've done far less damage to NVIDIA's financials over a much shorter period of time.

    The lack of higher end APUs hasn't helped, nor has the issue with actually trying to get a GPU onto a CPU die in the first place. Remember that when Intel tried it with Clarkdale/Arrandale, the graphics and IMC were 45nm, sitting alongside everything else which was 32nm.
  • chizow - Friday, July 3, 2015 - link

    I think you have to look at a bigger sample than that, riding on the 9000 series momentum, AMD was competitive for years with a near 50/50 share through the X800/X1900 series. And then G80/R600 happened and they never really recovered. There was a minor blip with Cypress vs. Fermi where AMD got close again but Nvidia quickly righted things with GF106 and GF110 (GTX 570/580).
  • Scali - Tuesday, July 7, 2015 - link

    nVidia wasn't the underdog in terms of technology. nVidia was the choice of gamers. ATi was big because they had been around since the early days of CGA and Hercules, and had lots of OEM contracts.
    In terms of technology and performance, ATi was always struggling to keep up with nVidia, and they didn't reach parity until the Radeon 8500/9700-era, even though nVidia was the newcomer and ATi had been active in the PC market since the mid-80s.
  • Frenetic Pony - Thursday, July 2, 2015 - link

    Well done analysis, though the kick in the head was Bulldozer and it's utter failure. Core 2 wasn't really AMD's downfall so much as Core/Sandy Bridge, which came at the exact wrong time for the utter failure of Bulldozer. This combined with AMD's dismal failure to market its graphics card has cost them billions. Even this article calls the 290x problematic, a card that offered the same performance as the original Titan at a fraction of the price. Based on empirical data the 290/x should have been almost continuously sold until the introduction of Nvidia's Maxwell architecture.

    Instead people continued to buy the much less performant per dollar Nvidia cards and/or waited for "the good GPU company" to put out their new architecture. AMD's performance in marketing has been utterly appalling at the same time Nvidia's has been extremely tight. Whether that will, or even can, change next year remains to be seen.

Log in

Don't have an account? Sign up now