The Mac Pro Review (Late 2013)
by Anand Lal Shimpi on December 31, 2013 3:18 PM ESTGaming Performance
As I mentioned earlier, under OS X games have to specifically be written to use both GPUs in the new Mac Pro. Under Windows however it's just a matter of enabling CrossFire X. I ran the new Mac Pro with dual FirePro D700s through a few of Ryan's 2014 GPU test suite games. The key comparison here is AMD's Radeon R9 280X CF. I've put all of the relevent information about the differences between the GPUs in the table below:
Mac Pro (Late 2013) GPU Comparison | ||||||
AMD Radeon R9 280X | AMD FirePro D700 | |||||
SPs | 2048 | 2048 | ||||
GPU Clock (base) | 850MHz | 650MHz | ||||
GPU Clock (boost) | 1000MHz | 850MHz | ||||
Single Precision GFLOPS | 4096 GFLOPS | 3481 GFLOPS | ||||
Texture Units | 128 | 128 | ||||
ROPs | 32 | 32 | ||||
Transistor Count | 4.3 Billion | 4.3 Billion | ||||
Memory Interface | 384-bit GDDR5 | 384-bit GDDR5 | ||||
Memory Datarate | 6000MHz | 5480MHz | ||||
Peak GPU Memory Bandwidth | 288 GB/s | 264 GB/s | ||||
GPU Memory | 3GB | 6GB |
Depending on thermal conditions the 280X can be as little as 17% faster than the D700 or as much as 30% faster, assuming it's not memory bandwidth limited. In the case of a memory bandwidth limited scenario the gap can shrink to 9%.
All of the results below are using the latest Radeon WHQL drivers at the time of publication (13-12_win7_win8_64_dd_ccc_whql.exe) running 64-bit Windows 8.1. Keep in mind that the comparison cards are all run on our 2014 GPU testbed, which is a 6-core Ivy Bridge E (i7-4960X) running at 4.2GHz. In other words, the other cards will have a definite CPU performance advantage (20 - 30% depending on the number of active cores).
You'll notice that I didn't run anything at 4K for these tests. Remember CrossFire at 4K is still broken on everything but the latest GCN 1.1 hardware from AMD.
Battlefield 3 starts out telling the story I expected to see. A pair of 280Xes ends up being 16% faster than the dual FirePro D700 setup in the Mac Pro. You really start to get an idea of where the Mac Pro's high-end GPU configuration really lands.
Bioshock ends up at the extreme end of what we'd expect to see between the 280X and D700. I tossed in a score from Bioshock under OS X, which obviously doesn't have CF working and ends up at less than half of the performance of the D700. If you're going to do any heavy 3D gaming, you'll want to do it under Windows still.
Not all games will scale well across multiple GPUs: Company of Heroes 2 is one of them. There's no performance uplift from having two 280Xes and thus the D700 performs like a slower single GPU R9 280X.
Metro is the one outlier in our test suite. Although CrossFire is clearly working under Windows, under Metro the D700 behaves as if it wasn't. I'm not sure what's going on here, but this does serve as a reminder that relying on multi-GPU setups to increase performance does come with a handful of these weird cases - particularly if you're using non-standard GPU configurations.
267 Comments
View All Comments
zepi - Wednesday, January 1, 2014 - link
How about virtualization and for example VT-d support with multiple gpu's and thunderbolts etc?Ie. Running windows in a virtual machine with half a dozen cores + another GPU while using rest for the OSX simultaneously?
I'd assume some people would benefit of having both OSX and Windows content creation applications and development environments available to them at the same time. Not to mention gaming in a virtual machine with dedicated GPU instead of virtual machine overhead / incompatibility etc.
japtor - Wednesday, January 1, 2014 - link
This is something I've wondered about too, for a while now really. I'm kinda iffy on this stuff, but last I checked (admittedly quite a while back) OS X wouldn't work as the hypervisor and/or didn't have whatever necessary VT-d support. I've heard of people using some other OS as the hypervisor with OS X and Windows VMs, but then I think you'd be stuck with hard resource allocation in that case (without restarting at least). Fine if you're using both all the time but a waste of resources if you predominantly use one vs the other.horuss - Thursday, January 2, 2014 - link
Anyway, I still would like to see some virtualization benchs. In my case, I can pretty much make it as an ideal home server with external storage while taking advantage of the incredible horse power to run multiple vms for my tests, for development, gaming and everything else!iwod - Wednesday, January 1, 2014 - link
I have been how likely we get a Mac ( Non Pro ) Spec.Nvidia has realize those extra die space wasted for GPGPU wasn't worth it. Afterall their main target are gamers and gaming benchmarks. So they decided for Kepler they have two line, one for GPGPU and one on the mainstream. Unless they change course again I think Maxwell will very likely follow the same route. AMD are little difference since they are betting on their OpenCL Fusion with their APU, therefore GPGPU are critical for them.
That could means Apple diverge their product line with Nvidia on the non Professional Mac like iMac and Macbook Pro ( Urg.. ) while continue using AMD FirePro on the Mac Pro Line.
Last time it was rumoured Intel wasn't so interested in getting a Broadwell out for Desktop, the 14nm die shrink of Haswell. Mostly because Mobile / Notebook CPU has over taken Desktop and will continue to do so. It is much more important to cater for the biggest market. Not to mention die shrink nowadays are much more about Power savings then Performance Improvements. So Intel could milk the Desktop and Server Market while continue to lead in Mobile and try to catch up with 14nm Atom SoC.
If that is true, the rumor of Haswell-Refresh on Desktop could mean Intel is no longer delaying Server Product by a single cycle. They will be doing the same for Desktop as well.
That means there could be a Mac Pro with Haswell-EP along with Mac with a Haswell-Refresh.
And by using Nvidia Gfx instead of AMD Apple dont need to worry about Mac eating into Mac Pro Market. And there could be less cost involve with not using a Pro Gfx card, only have 3 TB display, etc.
words of peace - Wednesday, January 1, 2014 - link
I keep thinking that if the MP is a good seller, maybe Apple could enlarge the unit so it contains a four sided heatsink, this could allow for dual CPU.Olivier_G - Wednesday, January 1, 2014 - link
Hi,I don't understand the comment about the lack of HiDPI mode here?
I would think it's simply the last one down the list, listed as 1920x1080 HiDPI, it does make the screen be perceived as such for apps, yet photos and text render at 4x resolution, which is what we're looking for i believe?
i tried such mode on my iMac out of curiosity and while 1280x720 is a bit ridiculously small it allowed me to confirm it does work since OSX mavericks. So I do expect the same behaviour to use my 4K monitor correctly with mac pro?
Am I wrong?
Gigaplex - Wednesday, January 1, 2014 - link
The article clearly states that it worked at 1920 HiDPI but the lack of higher resolutions in HiDPI mode is the problem.Olivier_G - Wednesday, January 1, 2014 - link
Well no it does not state that at all I read again and he did not mention trying the last option in the selector.LumaForge - Wednesday, January 1, 2014 - link
Anand,Firstly, thank you very much for such a well researched and well thought out piece of analysis - extremely insightful. I've been testing a 6 core and 12 core nMP all week using real-life post-production workflows and your scientific analysis helps explain why I've gotten good and OK results in some situations and not always seen the kinds of real-life improvements I was expecting in others.
Three follow up questions if I may:
1) DaVinci Resolve 10.1 ... have you done any benchmarking on Resolve with 4K files? ... like FCP X 10.1, BMD have optimized Resolve 10.1 to take full advantage of split CPU and GPU architecture but I'm not seeing the same performance gains as with FCP x 10.1 .... wondering if you have any ideas on system optimization or the sweet spot? I'm still waiting for my 8 core to arrive and that may be the machine that really takes advantage of the processor speed versus cores trade-off you identify.
2) Thunderbolt 2 storage options? ... external storage I/O also plays a significant role in overall sustained processing performance especially with 4K workflows ... I posted a short article on Creative Cow SAN section detailing some of my findings (no where as detailed or scientific as your approach I'm afraid) ... be interested to know your recommendations on Tbolt2 storage.
http://forums.creativecow.net/readpost/197/859961
3) IP over Tbolt2 as peer-to-peer networking topology? ... as well as running the nMPs in DAS, NAS and SAN modes I've also been testing IP over Tbolt2 .... only been getting around 500 MB/s sustained throughput between two nMPs ... if you look at the AJA diskwhack tests I posted on Creative Cow you'll see that the READ speeds are very choppy ... looks like a read-ahead caching issue somewhere in the pipeline or lack of 'Jumbo Frames' across the network ... have you played with TCP/IP over Thunderbolt2 yet and come to any conclusions on how to optimize throughput?
Keep up the good work and all the best for 2014.
Cheers,
Neil
modeleste - Wednesday, January 1, 2014 - link
I noticed that the Toshiba 65" 4k TV is about the same price as the Sharp 32" The reviews seem nice.Does anyone have any ide what the issues would be with using this display?