Professional Performance: Windows

Agisoft Photoscan – 2D to 3D Image Manipulation: link

Agisoft Photoscan creates 3D models from 2D images, a process which is very computationally expensive. The algorithm is split into four distinct phases, and different phases of the model reconstruction require either fast memory, fast IPC, more cores, or even OpenCL compute devices to hand. Agisoft supplied us with a special version of the software to script the process, where we take 50 images of a stately home and convert it into a medium quality model. This benchmark typically takes around 15-20 minutes on a high end PC on the CPU alone, with GPUs reducing the time.

Agisoft Photoscan 1.0.0

Photoscan, on paper, would offer more possibilities for faster memory to make a difference. However it would seem that the most memory dependent stage (stage 3) is actually a small part of the overall calculation and was absorbed by the natural variation in the larger stages, giving at most a 1.1% difference between times.

Cinebench R15

Cinebench R15 - Single Thread

Cinebench R15 - MultiThread

Cinebench is historically CPU dependent, giving a 2% difference from JEDEC to peak results.

3D Particle Movement

3DPM is a self-penned benchmark, taking basic 3D movement algorithms used in Brownian Motion simulations and testing them for speed. High floating point performance, MHz and IPC wins in the single thread version, whereas the multithread version has to handle the threads and loves more cores.

3D Particle Movement: Single Threaded

3D Particle Movement: MultiThreaded

3DPM is also relatively memory agnostic for DDR4 on Haswell-E, showing that DDR4-2133 is good enough.

Professional Performance: Linux

Built around several freely available benchmarks for Linux, Linux-Bench is a project spearheaded by Patrick at ServeTheHome to streamline about a dozen of these tests in a single neat package run via a set of three commands using an Ubuntu 14.04 LiveCD. These tests include fluid dynamics used by NASA, ray-tracing, molecular modeling, and a scalable data structure server for web deployments. We run Linux-Bench and have chosen to report a select few of the tests that rely on CPU and DRAM speed.

C-Ray: link

C-Ray is a simple ray-tracing program that focuses almost exclusively on processor performance rather than DRAM access. The test in Linux-Bench renders a heavy complex scene offering a large scalable scenario.

Linux-Bench c-ray 1.1 (Hard)

Natural variation gives a 4% difference, although the faster and more dense memory gave slower times.

NAMD, Scalable Molecular Dynamics: link

Developed by the Theoretical and Computational Biophysics Group at the University of Illinois at Urbana-Champaign, NAMD is a set of parallel molecular dynamics codes for extreme parallelization up to and beyond 200,000 cores. The reference paper detailing NAMD has over 4000 citations, and our testing runs a small simulation where the calculation steps per unit time is the output vector.

Linux-Bench NAMD Molecular Dynamics

NAMD showed little difference between our memory kits, peaking at 0.7% above JEDEC.

NPB, Fluid Dynamics: link

Aside from LINPACK, there are many other ways to benchmark supercomputers in terms of how effective they are for various types of mathematical processes. The NAS Parallel Benchmarks (NPB) are a set of small programs originally designed for NASA to test their supercomputers in terms of fluid dynamics simulations, useful for airflow reactions and design.

Linux-Bench NPB Fluid Dynamics

Despite the 4x8 GB results going south of the border, the faster memory does give a slight difference in NPB, peaking at 4.3% increased performance for the 3000+ memory kits.

Redis: link

Many of the online applications rely on key-value caches and data structure servers to operate. Redis is an open-source, scalable web technology with a b developer base, but also relies heavily on memory bandwidth as well as CPU performance.

Linux-Bench Redis Memory-Key Store, 100x

When tackling a high number of users, Redis performs up to 17% better using 2800+ memory, indicating our best benchmark result.

Memory Scaling on Haswell-E: CPU Real World Memory Scaling on Haswell: Single GTX 770 Gaming
Comments Locked

120 Comments

View All Comments

  • Flunk - Thursday, February 5, 2015 - link

    "There is one other group of individuals where super-high frequency memory on Haswell-E makes sense – the sub-zero overclockers."

    Yeah, I'm sure the 200 people on the planet who care about that are a real big market...

    Nice article overall though. I don't know why, but I was expecting more from DDR4. It looks like there is little reason to upgrade right now. Although I expect we'll all end up being forced into it by Intel.
  • Antronman - Thursday, February 5, 2015 - link

    There's a lot of consumers who want high clocked memory just because they want it.

    And there's more than 200 extreme overclockers on the planet.
  • galta - Thursday, February 5, 2015 - link

    The reason to upgrade today is not DDR4 per se, but 5xxx CPUs, and you might want these CPUs because of the extra cores, extra pci lanes, both, or just because you want it and can pay for it.
    These discussions over RAM get me tired. Rocks on the streets know that:
    a) fast memory makes close to no difference in real world, especially today with overclocking being so much more friendly than it was in the past
    b) whenever a new standard is introduced, it performs poorly when compared to previous standard. It was like this with DDR3 back in 2008 and it's the same today, but today you probably have less than 200 people saying they miss DDR2.
    Let's discuss more interesting and reasonable subjects.
  • Murloc - Thursday, February 5, 2015 - link

    200? You're severely underestimating the number of people who do that.

    Also why do car companies make cars that are going to be driven by just a few sheiks?

    With rams it's probably even easier given that you just have to bin chips and there are people who buy them just because they want the best. That's why they put increasingly cooler heatsinks and packages on the more pricey sticks. Not because they really need additional cooling in non-extreme use cases.
  • FlushedBubblyJock - Sunday, February 15, 2015 - link

    Because the elite cater to the elite, and the clique' is small and expensive, and leeches off the masses for the advantage and opulent and greedy lifestyle and media hype and self aggrandizement.
    They can fly each other around the world for huge parties and giveaway gatherings called global contests and spend enormous sums and feel very important.
  • imaheadcase - Thursday, February 5, 2015 - link

    Wait a tick, DDR2 is 800+mhz. That is what its default to on both my systems.
  • imaheadcase - Thursday, February 5, 2015 - link

    You put 200-533 MHz. My is actually at 936mhz for the overclock to.
  • ZeDestructor - Thursday, February 5, 2015 - link

    DDR = Double Data Rate, i.e: two operations are done per clock cycle. Thus the frequency is 400, but the effective frequency is 800. Same applies for DDR1-DDR4.

    GDDR5 is crazier: 4 operations per clock cycle, so 1750MHz works out to 7000MHz effective.
  • Murloc - Thursday, February 5, 2015 - link

    so basically what we knew all along: many enthusiasts are just wasting their money. The same goes for size although few people who build PCs are that stupid when it comes to this, it's mostly gamers who buy pre-built PCs who fall into this trap (it's not like they have much of a choice anyway, everybody is selling computers with lots of RAM and a pricey CPU bottlenecked by a weak GPU because it makes them money).
  • fredv78 - Thursday, February 5, 2015 - link

    seems to me most benchmarks are within the error margin (which is usually up to 3% and ideally should be quoted)

Log in

Don't have an account? Sign up now