Gaming Performance

As I mentioned earlier, under OS X games have to specifically be written to use both GPUs in the new Mac Pro. Under Windows however it's just a matter of enabling CrossFire X. I ran the new Mac Pro with dual FirePro D700s through a few of Ryan's 2014 GPU test suite games. The key comparison here is AMD's Radeon R9 280X CF. I've put all of the relevent information about the differences between the GPUs in the table below:

Mac Pro (Late 2013) GPU Comparison
  AMD Radeon R9 280X AMD FirePro D700
SPs 2048 2048
GPU Clock (base) 850MHz 650MHz
GPU Clock (boost) 1000MHz 850MHz
Single Precision GFLOPS 4096 GFLOPS 3481 GFLOPS
Texture Units 128 128
ROPs 32 32
Transistor Count 4.3 Billion 4.3 Billion
Memory Interface 384-bit GDDR5 384-bit GDDR5
Memory Datarate 6000MHz 5480MHz
Peak GPU Memory Bandwidth 288 GB/s 264 GB/s
GPU Memory 3GB 6GB

Depending on thermal conditions the 280X can be as little as 17% faster than the D700 or as much as 30% faster, assuming it's not memory bandwidth limited. In the case of a memory bandwidth limited scenario the gap can shrink to 9%.

All of the results below are using the latest Radeon WHQL drivers at the time of publication (13-12_win7_win8_64_dd_ccc_whql.exe) running 64-bit Windows 8.1. Keep in mind that the comparison cards are all run on our 2014 GPU testbed, which is a 6-core Ivy Bridge E (i7-4960X) running at 4.2GHz. In other words, the other cards will have a definite CPU performance advantage (20 - 30% depending on the number of active cores).

You'll notice that I didn't run anything at 4K for these tests. Remember CrossFire at 4K is still broken on everything but the latest GCN 1.1 hardware from AMD.

Battlefield 3 - 2560x1440 - Ultra Quality + 4x MSAA

Battlefield 3 starts out telling the story I expected to see. A pair of 280Xes ends up being 16% faster than the dual FirePro D700 setup in the Mac Pro. You really start to get an idea of where the Mac Pro's high-end GPU configuration really lands.

Bioshock Infinite - 2560x1440 - Ultra Quality + DDoF

Bioshock ends up at the extreme end of what we'd expect to see between the 280X and D700. I tossed in a score from Bioshock under OS X, which obviously doesn't have CF working and ends up at less than half of the performance of the D700. If you're going to do any heavy 3D gaming, you'll want to do it under Windows still.

Company of Heroes 2 - 2560x1440 - Maximum Quality + Med. AA

Not all games will scale well across multiple GPUs: Company of Heroes 2 is one of them. There's no performance uplift from having two 280Xes and thus the D700 performs like a slower single GPU R9 280X.

Company of Heroes 2 - Min. Frame Rate - 2560x1440 - Maximum Quality + Med. AA

Metro: Last Light - 2560x1440 - High Quality

Metro is the one outlier in our test suite. Although CrossFire is clearly working under Windows, under Metro the D700 behaves as if it wasn't. I'm not sure what's going on here, but this does serve as a reminder that relying on multi-GPU setups to increase performance does come with a handful of these weird cases - particularly if you're using non-standard GPU configurations.

GPU Choices 4K Support & The 4K Experience
Comments Locked

267 Comments

View All Comments

  • zepi - Wednesday, January 1, 2014 - link

    How about virtualization and for example VT-d support with multiple gpu's and thunderbolts etc?

    Ie. Running windows in a virtual machine with half a dozen cores + another GPU while using rest for the OSX simultaneously?

    I'd assume some people would benefit of having both OSX and Windows content creation applications and development environments available to them at the same time. Not to mention gaming in a virtual machine with dedicated GPU instead of virtual machine overhead / incompatibility etc.
  • japtor - Wednesday, January 1, 2014 - link

    This is something I've wondered about too, for a while now really. I'm kinda iffy on this stuff, but last I checked (admittedly quite a while back) OS X wouldn't work as the hypervisor and/or didn't have whatever necessary VT-d support. I've heard of people using some other OS as the hypervisor with OS X and Windows VMs, but then I think you'd be stuck with hard resource allocation in that case (without restarting at least). Fine if you're using both all the time but a waste of resources if you predominantly use one vs the other.
  • horuss - Thursday, January 2, 2014 - link

    Anyway, I still would like to see some virtualization benchs. In my case, I can pretty much make it as an ideal home server with external storage while taking advantage of the incredible horse power to run multiple vms for my tests, for development, gaming and everything else!
  • iwod - Wednesday, January 1, 2014 - link

    I have been how likely we get a Mac ( Non Pro ) Spec.
    Nvidia has realize those extra die space wasted for GPGPU wasn't worth it. Afterall their main target are gamers and gaming benchmarks. So they decided for Kepler they have two line, one for GPGPU and one on the mainstream. Unless they change course again I think Maxwell will very likely follow the same route. AMD are little difference since they are betting on their OpenCL Fusion with their APU, therefore GPGPU are critical for them.
    That could means Apple diverge their product line with Nvidia on the non Professional Mac like iMac and Macbook Pro ( Urg.. ) while continue using AMD FirePro on the Mac Pro Line.

    Last time it was rumoured Intel wasn't so interested in getting a Broadwell out for Desktop, the 14nm die shrink of Haswell. Mostly because Mobile / Notebook CPU has over taken Desktop and will continue to do so. It is much more important to cater for the biggest market. Not to mention die shrink nowadays are much more about Power savings then Performance Improvements. So Intel could milk the Desktop and Server Market while continue to lead in Mobile and try to catch up with 14nm Atom SoC.

    If that is true, the rumor of Haswell-Refresh on Desktop could mean Intel is no longer delaying Server Product by a single cycle. They will be doing the same for Desktop as well.

    That means there could be a Mac Pro with Haswell-EP along with Mac with a Haswell-Refresh.
    And by using Nvidia Gfx instead of AMD Apple dont need to worry about Mac eating into Mac Pro Market. And there could be less cost involve with not using a Pro Gfx card, only have 3 TB display, etc.
  • words of peace - Wednesday, January 1, 2014 - link

    I keep thinking that if the MP is a good seller, maybe Apple could enlarge the unit so it contains a four sided heatsink, this could allow for dual CPU.
  • Olivier_G - Wednesday, January 1, 2014 - link

    Hi,

    I don't understand the comment about the lack of HiDPI mode here?

    I would think it's simply the last one down the list, listed as 1920x1080 HiDPI, it does make the screen be perceived as such for apps, yet photos and text render at 4x resolution, which is what we're looking for i believe?

    i tried such mode on my iMac out of curiosity and while 1280x720 is a bit ridiculously small it allowed me to confirm it does work since OSX mavericks. So I do expect the same behaviour to use my 4K monitor correctly with mac pro?

    Am I wrong?
  • Gigaplex - Wednesday, January 1, 2014 - link

    The article clearly states that it worked at 1920 HiDPI but the lack of higher resolutions in HiDPI mode is the problem.
  • Olivier_G - Wednesday, January 1, 2014 - link

    Well no it does not state that at all I read again and he did not mention trying the last option in the selector.
  • LumaForge - Wednesday, January 1, 2014 - link

    Anand,

    Firstly, thank you very much for such a well researched and well thought out piece of analysis - extremely insightful. I've been testing a 6 core and 12 core nMP all week using real-life post-production workflows and your scientific analysis helps explain why I've gotten good and OK results in some situations and not always seen the kinds of real-life improvements I was expecting in others.

    Three follow up questions if I may:

    1) DaVinci Resolve 10.1 ... have you done any benchmarking on Resolve with 4K files? ... like FCP X 10.1, BMD have optimized Resolve 10.1 to take full advantage of split CPU and GPU architecture but I'm not seeing the same performance gains as with FCP x 10.1 .... wondering if you have any ideas on system optimization or the sweet spot? I'm still waiting for my 8 core to arrive and that may be the machine that really takes advantage of the processor speed versus cores trade-off you identify.

    2) Thunderbolt 2 storage options? ... external storage I/O also plays a significant role in overall sustained processing performance especially with 4K workflows ... I posted a short article on Creative Cow SAN section detailing some of my findings (no where as detailed or scientific as your approach I'm afraid) ... be interested to know your recommendations on Tbolt2 storage.

    http://forums.creativecow.net/readpost/197/859961

    3) IP over Tbolt2 as peer-to-peer networking topology? ... as well as running the nMPs in DAS, NAS and SAN modes I've also been testing IP over Tbolt2 .... only been getting around 500 MB/s sustained throughput between two nMPs ... if you look at the AJA diskwhack tests I posted on Creative Cow you'll see that the READ speeds are very choppy ... looks like a read-ahead caching issue somewhere in the pipeline or lack of 'Jumbo Frames' across the network ... have you played with TCP/IP over Thunderbolt2 yet and come to any conclusions on how to optimize throughput?

    Keep up the good work and all the best for 2014.

    Cheers,
    Neil
  • modeleste - Wednesday, January 1, 2014 - link

    I noticed that the Toshiba 65" 4k TV is about the same price as the Sharp 32" The reviews seem nice.

    Does anyone have any ide what the issues would be with using this display?

Log in

Don't have an account? Sign up now