The Mac Pro (Late 2013)

When Anand reviewed the Mac Pro late last year, he received the full fat 12 core edition, using the E5-2697 v2 CPU with a 2.7 GHz rating. The CPU choices for the Mac Pro include 4, 6 and 8 core models, all with HyperThreading. Interestingly enough, the 4/6/8 core models all come from the E5-16xx line, meaning the CPUs are designed with single processor systems in mind. But to get to the 12 core/24 thread model at the high end, Apple used the E5-2697 v2, a processor optimized for dual CPU situations. Based on the die shots on the previous page, this has repercussions, but as Anand pointed out, it all comes down to power usage and turbo performance.

Mac Pro (Late 2013) CPU Options
Intel CPU E5-1620 v2 E5-1650 v2 E5-1680 v2 E5-2697 v2
Cores / Threads 4 / 8 6 / 12 8 / 16 12 / 24
CPU Base Clock 3.7GHz 3.5GHz 3.0GHz 2.7GHz
Max Turbo (1C) 3.9GHz 3.9GHz 3.9GHz 3.5GHz
L3 Cache 10MB 12MB 25MB 30MB
TDP 130W 130W 130W 130W
Intel SRP $294 $583 ? $2614

The Mac Pro is designed within a peak 450W envelope, and Intel has options with its CPUs. For the same TDP limit, Intel can create many cores as low frequency, or fewer cores at higher frequency. This is seen in the options on the Mac Pro – all the CPU choices have the same 130W TDP, but the CPU base clocks change as we rise up the core count. Moving from 4 cores to 8 cores keeps the maximum turbo (single core performance) at 3.9 GHz, but the base clock decreases the more cores are available. Finally at the 12-core model, the base frequency is at its lowest of the set, as well as the maximum turbo.

This has repercussions on workloads, especially for workstations. For the most part, the types of applications used on workstations are highly professional, and have big budgets with plenty of engineers designed to extract performance. That should bode well for the systems with more cores, despite the frequency per core being lower. However, it is not always that simple – the mathematics for the problem has to be able to take advantage of parallel computing. Simple programs run solely on one core because that is the easiest to develop, but if the mathematics wholly linear, then even enterprise software is restricted. This would lend a positive note to the higher turbo frequency CPUs. Intel attempts to keep the turbo frequency similar as long as it can while retaining the maximum TDP to avoid this issue; however at the 12-core model this is not possible. Quantifying your workload before making a purchase is a key area that users have to consider.

Benchmark Configuration

I talk about the Mac Pro a little because the processors we have for a ‘regular’ test today are 8-core and 12-core models. The 12-core is the same model that Anand tested in the Mac Pro – the Xeon E5-2697v2. The 8-core model we are testing today is different to the one offered in the Mac Pro, in terms of frequency and TDP:

Intel SKU Comparison
  Core i7-4960X Xeon E5-2687W v2 Xeon E5-2697 v2
Release Date September 10, 2013 September 10, 2013 September 10, 2013
Cores 6 8 12
Threads 12 16 24
Base Frequency 3600 3400 2700
Turbo Frequency 4000 4000 3500
L3 Cache 15 MB 25 MB 30 MB
Max TDP 130 W 150 W 130 W
Max Memory Size 64 GB 256 GB 768 GB
Memory Channels 4 4 4
Memory Frequency DDR3-1866 DDR3-1866 DDR3-1866
PCIe Revision 3.0 3.0 3.0
PCIe Lanes 40 40 40
Multi-Processor 1P 2P 2P
VT-x Yes Yes Yes
VT-d Yes Yes Yes
vPro No Yes Yes
Memory Bandwidth 59.7 GB/s 59.7 GB/s 59.7 GB/s
Price $1059 $2112 $2618

The reason for this review is to put these enterprise class processors through the normal (rather than server) benchmarks I run at AnandTech for processors. Before I started writing about technology, as an enthusiast, it was always interesting to hear of the faster Xeons and how much that actually made a difference to my normal computing. I luckily have that opportunity and would like to share it with our readers.

The system set up is as follows:

Test Setup
Motherboards GIGABYTE GA-6PXSV3
MSI X79A-GD45 Plus for 3x GPU Configurations
Memory 8x4 GB Kingston DDR3-1600 11-11-11 ECC
Storage OCZ Vertex 3 256 GB
Power Supply OCZ 1250 ZX Series
CPU Cooler Corsair H80i
NVIDIA GPU MSI GTX 770 Lightning 2GB
AMD GPU ASUS HD 7970 3GB

Many thanks to...

We must thank the following companies for kindly providing hardware for our test bed:

Thank you to GIGABYTE Server for providing us with the Motherboard and CPUs
Thank you to OCZ for providing us with 1250W Gold Power Supplies and SSDs.
Thank you to Kingston for the ECC Memory kit
Thank you to ASUS for providing us with the AMD HD7970 GPUs and some IO Testing kit.
Thank you to MSI for providing us with the NVIDIA GTX 770 Lightning GPUs.

Power Consumption

Power consumption was tested on the system as a whole with a wall meter connected to the OCZ 1250W power supply, while in a single MSI GTX 770 Lightning GPU configuration. This power supply is Gold rated, and as I am in the UK on a 230-240 V supply, leads to ~75% efficiency > 50W, and 90%+ efficiency at 250W, which is suitable for both idle and multi-GPU loading. This method of power reading allows us to compare the power management of the UEFI and the board to supply components with power under load, and includes typical PSU losses due to efficiency. These are the real world values that consumers may expect from a typical system (minus the monitor) using this motherboard.

While this method for power measurement may not be ideal, and you feel these numbers are not representative due to the high wattage power supply being used (we use the same PSU to remain consistent over a series of reviews, and the fact that some boards on our test bed get tested with three or four high powered GPUs), the important point to take away is the relationship between the numbers. These boards are all under the same conditions, and thus the differences between them should be easy to spot.

Power Consumption - Idle

Power Consumption - Long Idle

Power Consumption - OCCT

At idle, the Xeons are on par with the Core i7-4960X for power consumption in the GIGABYTE motherboard. At load the extra TDP of the E5-2687W v2 can be seen.

DPC Latency

Deferred Procedure Call latency is a way in which Windows handles interrupt servicing. In order to wait for a processor to acknowledge the request, the system will queue all interrupt requests by priority. Critical interrupts will be handled as soon as possible, whereas lesser priority requests, such as audio, will be further down the line. So if the audio device requires data, it will have to wait until the request is processed before the buffer is filled. If the device drivers of higher priority components in a system are poorly implemented, this can cause delays in request scheduling and process time, resulting in an empty audio buffer – this leads to characteristic audible pauses, pops and clicks. Having a bigger buffer and correctly implemented system drivers obviously helps in this regard. The DPC latency checker measures how much time is processing DPCs from driver invocation – the lower the value will result in better audio transfer at smaller buffer sizes. Results are measured in microseconds and taken as the peak latency while cycling through a series of short HD videos - less than 500 microseconds usually gets the green light, but the lower the better.

DPC Latency Maximum

The DPC latency of the Xeons is closer to the 100 mark, which we saw during Sandy Bridge.  Newer systems seem to be increasing the DPC latency - so far all Haswell consumer CPUs are at the 140+ line.

Intel Xeon E5-2697 v2 and Xeon E5-2687W v2 Review: 12 and 8 Cores Real World CPU Benchmarks: Rendering, Compression, Video Conversion
Comments Locked

71 Comments

View All Comments

  • Ian Cutress - Tuesday, March 18, 2014 - link

    I need to spend some time to organise this with my new 2014 benchmark setup. That and I've never used bench to add data before. But I will be putting some data in there for everyone :)
  • Maxal - Tuesday, March 18, 2014 - link

    There is one sad thing - disappearance of 2C/4T high clock speed CPUs, as Oracle Enterprise Edition charges by cores.....and sometimes you need just small installation but with EE features...
  • Rick83 - Tuesday, March 18, 2014 - link

    Wouldn't L3/thread be a more useful metric than L3/core in the big table?
    HT will only really work after all, if both threads are in cache, and if you can get a CPU with HT and one without, as is the case with the Xeons, you'd get the one without because you are running more concurrent threads. That means that under optimum conditions, you have 2 threads per core that are active, and thus 2x#cores threads that need to be in the data caches.
  • HalloweenJack - Tuesday, March 18, 2014 - link

    holy shit anandtech you really have gone to the dogs - comparing a £2000 cpu against a £100 apu and saying its better..... and really? wheres the AMD AM3+ cpu`s? 8350 or 9590? seriously
  • Ian Cutress - Tuesday, March 18, 2014 - link

    Let's see. I'm not comparing it against a £100 APU, I'm comparing it against the $1000 Core i7-4960X to see the difference. We're using a new set of benchmarks for 2014, which I have already run on the APU so I include them here as a point of reference for AMD's new highest performance line. It is interesting to see where the APU and Xeon line up in the benchmarks to show the difference (if any). AMD's old high end line has stagnated - I have not tested those CPUs in our new 2014 set of benchmarks. There have been no new AM3+ platforms or CPUs this year, or almost all of last year. Testing these two CPUs properly took the best part of three weeks, including all the other work such as news, motherboard reviews, Mobile World Congress coverage, meetings, extra testing, bug fixing, conversing with engineers on how to solve issues. Sure, let's just stop all that and pull out an old system to test. If I had the time I really would, but I was able to get these processors from GIGABYTE, not Intel, for a limited time. I have many other projects (memory scaling, Gaming CPU) that would take priority if I had time.

    AKA I think you missed the point of the article. If you have a magical portal to Narnia, I'd happily test until I was blue in the face and go as far back to old Athlon s939 CPUs. But the world moves faster than that.
  • deadrats - Tuesday, March 18, 2014 - link

    any chance of updating this article with some x265 and/or Divx265 benchmarks? hevc is much more processor intensive and threading friendly, so these encoders may be perfect for showing a greater separation between the various core configurations.
  • Ian Cutress - Tuesday, March 18, 2014 - link

    If you have an encoder in mind drop me an email. Click my name at the top of the article.
  • bobbozzo - Tuesday, March 18, 2014 - link

    Hi,

    1. please change the charts' headings on the first page to say 'Cores/Threads' instead of 'Cores'.

    2. it wasn't clear on the first page that this is talking about workstation CPUs.

    3. "Intel can push core counts, frequency and thus price much higher than in the consumer space"
    I would have said core counts and cache...
    Don't the consumer parts have the highest clocks (before overclocking)?

    Thanks!
  • bobbozzo - Tuesday, March 18, 2014 - link

    "it wasn't clear on the first page that this is talking about workstation CPUs."

    As opposed to servers.
  • Ian Cutress - Tuesday, March 18, 2014 - link

    1) I had it that way originally but it broke the table layout due to being too wide. I made a compromise and hoped people would follow the table in good faith.
    2) Generally Xeon in the name means anything Workstation and above. People use Xeons for a wide variety of uses - high end for workstaitons, or low end for servers, or vice versa.
    3) Individual core counts maybe, but when looking at 8c or 12c chips in the same power bracket, the frequency is still being pushed to more stringent requirements (thus lower yields/bin counts) vs. voltages. Then again, the E3-1290 does go to 4.0 GHz anyway, so in terms of absolute frequencies you can say (some) Xeons at least match the consumer parts.

Log in

Don't have an account? Sign up now