Power Consumption & Noise

The Mac Pro’s thermal core makes a lot of sense from an area efficiency standpoint as the chances that you have all three processors in the system (Xeon CPU + dual AMD FirePro GPUs) running at max speed at the same time is highly unlikely. By having all three players share one large heatsink Apple can optimize for the most likely usage scenarios where at most one processor is running at close to max TDP. From the moment the system was introduced however I was curious to see if its thermal design could deal with all three processors going full tilt. I decided to test both scenarios out.

The vast majority of the time the Mac Pro’s single fan spins at a paltry 769 RPM. At this speed the Mac Pro is virtually silent. I have a Promise Pegasus R6 Thunderbolt chassis with four 3.5” 7200 RPM hard drives inside. The Pegasus sits four feet away from my Mac Pro and at idle the hum of its drives drowns out the idle fan noise of the Mac.

Even running heavily threaded CPU benchmarks does nothing to impact fan speed. It's not until you run a prolonged workload that you can hear an increase in noise from the Mac Pro. I ran FurMark under OS X in parallel with a relatively heavy 4K render in Final Cut Pro. I was rendering a 20 minute 4K project with two effects applied across the entire timeline. The basic render used up 4 - 8 threads, while the effects ensured the compute GPU had some work to do. FurMark obviously kept the display GPU busy. After around 25% of the rendering task was complete the Mac Pro’s fan smoothly scaled up to 1400 RPM, then 1600 RPM and finally stopping at what I believe is full speed: 1900 RPM.

I crudely measured noise sitting at my desk, approximately 18” away from the Mac Pro at all of these points:

Mac Pro (Late 2013) Fan Noise vs. RPM
Fan Speed (RPM) Measured Sound Pressure Level @ 18"
Off 32.2 dBA
770 RPM 32.4 dBA
890 RPM 32.8 dBA
1097 RPM 34.1 dBA
1193 RPM 35.2 dBA
1440 RPM 38.6 dBA
1500 RPM 39.4 dBA
1900 RPM 44.7 dBA

System noise isn’t an issue the vast majority of the time. Even if you’re running for short bursts with everything cranking away on a problem. It’s only under prolonged use of the CPU + both GPUs that the fan has to spin up this high. I quit FurMark and kept the FCP render going, which brought system power down from ~380W to ~326W. It also slowly dropped fan speed from 1900 RPM down to 1500 RPM before smoothly alternating between 1200 RPM and 1450 RPM. The way the Mac Pro’s fan ramps is actually just as important as its ability to remove heat. There are no dramatic increases in speed, just an incredibly smooth curve that the Pro’s fan moves up and down.

The system never gets annoyingly loud, it just varies between silent and calmly reminding you that there’s a large fan inside. Even at full bore the Mac Pro’s fan noise is more pleasant than harsh. It’s not magic, just good system design coupled with the fact that even with a pair of D700s Apple went a bit conservative on GPU speed. Had this been a system capable of pulling 800W at the wall, we’d be having a different discussion about noise.

My next task was to see what actually happens in this worst case scenario. If you’re running all of the parts at full tilt, are any of them going to throttle? I have to work pretty hard to get the fan to spin up under OS X, but in Windows it’s a lot easier since I can just toss a single multi-GPU workload at the problem.

I started out by running LuxMark, an OpenCL workload, on both GPUs as well as a multithreaded 7-Zip benchmark on all of the CPU cores. I monitored both CPU and GPU frequencies. The result was no throttling across the board:

Getting an accurate reading on GPU frequencies from Tahiti based GPUs ends up being harder than I expected, but I saw what Ryan reminded me is typical behavior where the GPUs alternate between their 650MHz base clock and 850MHz max turbo. We don’t have good tools to actually measure their behavior in between unfortunately.

The same was true for the CPU. Even with all 12 cores taxed heavily, I never saw any drops below the CPU’s 2.7GHz base clock.

Next I tried a heavier workload on the CPU: a H.264 video encode. Here I just ran the x264 5.01 benchmark in parallel with the LuxMark workload. Once again, I saw no drop in CPU or GPU clocks although I believe I was approaching the limits of where that would hold true. The system was pulling an average of 410W at that point, with peak power draw at 429W.

If you’re wondering, there was little to no impact on the x264 benchmark from having LuxMark run in the background. The first rendering pass took about a 3% hit, likely due to the CPU not being able to turbo as high/at all, but the second heavily threaded pass was on par with my standalone run without LuxMark in the background. LuxMark on the other hand saw around a 14% reduction in performance, from 2040K samples per second down to 1750K when run in parallel with the x264 test. We’re still talking about two extremely compute intensive tasks, the fact that I can run both with little performance reduction is an example of the sort of performance scaling that’s possible if you leverage all of the compute in the Mac Pro.

So far I wasn’t surprised by the platform’s behavior. The Mac Pro’s thermal core and fan was enough to handle a real world workload without throttling. But what about a power virus? For this I ran a combination of FurMark at 2560 x 1440 and Prime95 (in-place large FFTs) across all GPUs and CPU cores.

The GPUs peaked at 97C, which is pretty much as high as a 28nm Tahiti should ever go. The CPU hit a similar temperature, with most cores hovering around 95C. GPU clocks seemed ok, with both GPUs running between 650 - 850MHz (base - boost). The CPU on the other hand dropped down to 2.1GHz (I even saw a short excursion down to 2GHz). Average power while running this workload was 437W, peaking at 463W before CPU throttling kicked in. If you plot out a graph of power vs. time you can see the CPU throttling kick in during the workload.

I have to stress that I haven’t been able to get this to happen in any normal workload, only what’s effectively a power virus for the GPUs and something quite unrealistic for the CPUs. Either way it shows us the upper limit of what the thermal core can do.

Overall power consumption is down considerably compared to previous generation Mac Pros:

Mac Pro Power Consumption
  Idle Multithreaded CPU Load
Apple Mac Pro (Late 2013/2.7GHz-12C) 85.5W 190.7W (Cinebench 11.5)
Apple Mac Pro (Early 2009/2.93GHz-8C) 125.3W 254W (Cinebench R10)
Apple PowerMac G5 Dual 2.5GHz 120.8W 252.2W (Cinebench R10)

Idle power consumption is down and power efficiency under load is way better than the early 2009 Mac Pro I have here. As far as I can tell, max power consumption for either compute system (CPU alone or dual GPUs alone) is around 300 - 320W. With both subsystems combined you'll get the 437W figure I talked about earlier.

You don't even need a power virus to get there. Running two realistic workloads that are heavy enough won't get you up to the 463W point where I saw throttling, but it'll get you close to the max average power I recorded on the Mac Pro.

Thunderbolt 2 Final Words
Comments Locked

267 Comments

View All Comments

  • zepi - Wednesday, January 1, 2014 - link

    How about virtualization and for example VT-d support with multiple gpu's and thunderbolts etc?

    Ie. Running windows in a virtual machine with half a dozen cores + another GPU while using rest for the OSX simultaneously?

    I'd assume some people would benefit of having both OSX and Windows content creation applications and development environments available to them at the same time. Not to mention gaming in a virtual machine with dedicated GPU instead of virtual machine overhead / incompatibility etc.
  • japtor - Wednesday, January 1, 2014 - link

    This is something I've wondered about too, for a while now really. I'm kinda iffy on this stuff, but last I checked (admittedly quite a while back) OS X wouldn't work as the hypervisor and/or didn't have whatever necessary VT-d support. I've heard of people using some other OS as the hypervisor with OS X and Windows VMs, but then I think you'd be stuck with hard resource allocation in that case (without restarting at least). Fine if you're using both all the time but a waste of resources if you predominantly use one vs the other.
  • horuss - Thursday, January 2, 2014 - link

    Anyway, I still would like to see some virtualization benchs. In my case, I can pretty much make it as an ideal home server with external storage while taking advantage of the incredible horse power to run multiple vms for my tests, for development, gaming and everything else!
  • iwod - Wednesday, January 1, 2014 - link

    I have been how likely we get a Mac ( Non Pro ) Spec.
    Nvidia has realize those extra die space wasted for GPGPU wasn't worth it. Afterall their main target are gamers and gaming benchmarks. So they decided for Kepler they have two line, one for GPGPU and one on the mainstream. Unless they change course again I think Maxwell will very likely follow the same route. AMD are little difference since they are betting on their OpenCL Fusion with their APU, therefore GPGPU are critical for them.
    That could means Apple diverge their product line with Nvidia on the non Professional Mac like iMac and Macbook Pro ( Urg.. ) while continue using AMD FirePro on the Mac Pro Line.

    Last time it was rumoured Intel wasn't so interested in getting a Broadwell out for Desktop, the 14nm die shrink of Haswell. Mostly because Mobile / Notebook CPU has over taken Desktop and will continue to do so. It is much more important to cater for the biggest market. Not to mention die shrink nowadays are much more about Power savings then Performance Improvements. So Intel could milk the Desktop and Server Market while continue to lead in Mobile and try to catch up with 14nm Atom SoC.

    If that is true, the rumor of Haswell-Refresh on Desktop could mean Intel is no longer delaying Server Product by a single cycle. They will be doing the same for Desktop as well.

    That means there could be a Mac Pro with Haswell-EP along with Mac with a Haswell-Refresh.
    And by using Nvidia Gfx instead of AMD Apple dont need to worry about Mac eating into Mac Pro Market. And there could be less cost involve with not using a Pro Gfx card, only have 3 TB display, etc.
  • words of peace - Wednesday, January 1, 2014 - link

    I keep thinking that if the MP is a good seller, maybe Apple could enlarge the unit so it contains a four sided heatsink, this could allow for dual CPU.
  • Olivier_G - Wednesday, January 1, 2014 - link

    Hi,

    I don't understand the comment about the lack of HiDPI mode here?

    I would think it's simply the last one down the list, listed as 1920x1080 HiDPI, it does make the screen be perceived as such for apps, yet photos and text render at 4x resolution, which is what we're looking for i believe?

    i tried such mode on my iMac out of curiosity and while 1280x720 is a bit ridiculously small it allowed me to confirm it does work since OSX mavericks. So I do expect the same behaviour to use my 4K monitor correctly with mac pro?

    Am I wrong?
  • Gigaplex - Wednesday, January 1, 2014 - link

    The article clearly states that it worked at 1920 HiDPI but the lack of higher resolutions in HiDPI mode is the problem.
  • Olivier_G - Wednesday, January 1, 2014 - link

    Well no it does not state that at all I read again and he did not mention trying the last option in the selector.
  • LumaForge - Wednesday, January 1, 2014 - link

    Anand,

    Firstly, thank you very much for such a well researched and well thought out piece of analysis - extremely insightful. I've been testing a 6 core and 12 core nMP all week using real-life post-production workflows and your scientific analysis helps explain why I've gotten good and OK results in some situations and not always seen the kinds of real-life improvements I was expecting in others.

    Three follow up questions if I may:

    1) DaVinci Resolve 10.1 ... have you done any benchmarking on Resolve with 4K files? ... like FCP X 10.1, BMD have optimized Resolve 10.1 to take full advantage of split CPU and GPU architecture but I'm not seeing the same performance gains as with FCP x 10.1 .... wondering if you have any ideas on system optimization or the sweet spot? I'm still waiting for my 8 core to arrive and that may be the machine that really takes advantage of the processor speed versus cores trade-off you identify.

    2) Thunderbolt 2 storage options? ... external storage I/O also plays a significant role in overall sustained processing performance especially with 4K workflows ... I posted a short article on Creative Cow SAN section detailing some of my findings (no where as detailed or scientific as your approach I'm afraid) ... be interested to know your recommendations on Tbolt2 storage.

    http://forums.creativecow.net/readpost/197/859961

    3) IP over Tbolt2 as peer-to-peer networking topology? ... as well as running the nMPs in DAS, NAS and SAN modes I've also been testing IP over Tbolt2 .... only been getting around 500 MB/s sustained throughput between two nMPs ... if you look at the AJA diskwhack tests I posted on Creative Cow you'll see that the READ speeds are very choppy ... looks like a read-ahead caching issue somewhere in the pipeline or lack of 'Jumbo Frames' across the network ... have you played with TCP/IP over Thunderbolt2 yet and come to any conclusions on how to optimize throughput?

    Keep up the good work and all the best for 2014.

    Cheers,
    Neil
  • modeleste - Wednesday, January 1, 2014 - link

    I noticed that the Toshiba 65" 4k TV is about the same price as the Sharp 32" The reviews seem nice.

    Does anyone have any ide what the issues would be with using this display?

Log in

Don't have an account? Sign up now