Compute

Shifting gears, let’s take a look at compute performance on Pascal.

Overall, we’re not expecting a significant difference in compute performance compared to Maxwell 2 for standard compute benchmarks. The fundamental architecture hasn’t changed – the CUDA cores, register files, and caches still behave as before - so there’s little reason for compute performance to shift. GP104 for all intents and purposes should perform like a higher clocked and slightly wider Maxwell 2, similar to what we’ve seen in most games.

However in the long run there is potential for Pascal to show some improvements. The architecture’s improved scheduling features are geared in part towards HPC users, and instruction level preemption means that compute kernels can now be a lot more aggressive on consumer systems since they can be paused so easily. That said, to really leverage any of these improvements, applications utilizing GPU compute need to have work that benefits from better scheduling and be written with Pascal in mind, and for consumer workloads the latter is likely a long way off.

Starting us off for our look at compute is LuxMark3.1, the latest version of the official benchmark of LuxRender. LuxRender’s GPU-accelerated rendering mode is an OpenCL based ray tracer that forms a part of the larger LuxRender suite. Ray tracing has become a stronghold for GPUs in recent years as ray tracing maps well to GPU pipelines, allowing artists to render scenes much more quickly than with CPUs alone.

Compute: LuxMark 3.1 - Hotel

As with games, when it comes to LuxMark, the GTX 1080 is uncontested; this is the first high performance FinFET GPU in action. That said, I’m surprised by how close some of these results cluster. Though GTX 1080 is not a full generational replacement for GTX 980 Ti, normally it outperforms the Big Maxwell card by more than this. Instead we’re looking at a lead of just 10%, notably less than a simple extrapolation of CUDA core counts and frequencies would tell us to expect (GTX 1080 has almost 50% more FLOPs).

That said, GTX 1070 still places very close to GTX 980 Ti – albeit below it – so what we’re seeing isn’t just Pascal being a laggard. Especially since as a consequence of this, GTX 1080 only beats GTX 1070 by 12%. In any case, this may be a case of early drivers, particularly as OpenCL has not been an NVIDIA priority for the last couple of years. Alternatively, as strange as it may be, I’m not ready to rule out LuxMark being CPU limited. It’s something that we’ll have to keep an eye on.

For our second set of compute benchmarks we have CompuBench 1.5, the successor to CLBenchmark. CompuBench offers a wide array of different practical compute workloads, and we’ve decided to focus on face detection, optical flow modeling, and particle simulations.

Compute: CompuBench 1.5 - Face Detection

Compute: CompuBench 1.5 - Optical Flow

Compute: CompuBench 1.5 - Particle Simulation 64K

Depending on which sub-test we’re looking at, CompuBench is all over the place. In Face Detection the GTX 1080 takes a commanding lead, with GTX 1070 easily slotting into second place. On the other hand we have Optical Flow, which NVIDIA cards have traditionally struggled with, where even GTX 1080 can’t unseat Radeon Fury X. Finally in the middle we have the 64K Particle Simulation, which has GTX 1080 in the lead again, but not unlike LuxMark, it also has some interesting clustering going on.

Ultimately each test stresses our GPU collection in different ways, which as we can see greatly influences how the results pan out. Face Detection has always played well to NVIDIA’s strengths, and on a generational basis we get solid scaling from Maxwell 2 to Pascal. Even Optical Flow, which seems to favor raw FLOPs more than anything else, still shows very good gains with Pascal.

Particle Simulation is the outlier in this regard; Pascal’s generational gains are not insignificant, but they’re less than what we’d expect. Furthermore GTX 1080 and GTX 1070 are very closely clustered together despite their much larger difference in FLOPs. This may mean we’re looking at a CPU or driver bottleneck, or possibly some sort of internal path bottleneck. GTX 1080 has more FLOPs and a similar advantage in memory bandwidth, but once you get on chip things get much closer. If nothing else this goes to show that compute benchmarks are much more architecture sensitive than games, which is why we can’t make very broad generalizations for all compute workloads.

Moving on, our 3rd compute benchmark is the next generation release of FAHBench, the official Folding @ Home benchmark. Folding @ Home is the popular Stanford-backed research and distributed computing initiative that has work distributed to millions of volunteer computers over the internet, each of which is responsible for a tiny slice of a protein folding simulation. FAHBench can test both single precision and double precision floating point performance, with single precision being the most useful metric for most consumer cards due to their low double precision performance. Each precision has two modes, explicit and implicit, the difference being whether water atoms are included in the simulation, which adds quite a bit of work and overhead. This is another OpenCL test, utilizing the OpenCL path for FAHCore 21.

Compute: Folding @ Home Single Precision

Compute: Folding @ Home Double Precision

In single precision performance, to the surprise of no one the GTX 1080 is solidly in the lead, followed up by the GTX 1070. On a generational basis performance gains are decent, but at 44% for GTX 1080 they aren’t quite as great as we’ve seen from the card elsewhere. Meanwhile the two Pascal cards are again closer than we’d expect, with GTX 1080 leading by only 10%.

As for double precision performance, we can see that even with the higher overall compute throughput of GP104, it still can’t make up for the fact that FP64 performance on the GPU is capped at 1/32 by virtue of so few FP64 CUDA cores, which puts even NVIDIA’s latest and greatest at a disadvantage here. But if nothing else, generational scaling versus Maxwell 2 looks very good, with performance gains closely tracking the theoretical increase in FLOPs.

Hitman Synthetics
Comments Locked

200 Comments

View All Comments

  • jcardel - Wednesday, July 27, 2016 - link

    This is excactly the same situation as me. I got a 770 sitting in my rig, and am looking hard at the 1070, maybe soon. Although my 770 is still up to the task in most games, I really play only blizzard games theese days and they are not hard on your hardware.

    My biggest issue is really that it is rather noisy, so I will be looking for a solution with the lowest DB.

    Great article, it was totally worth waiting for.. I only read this sort of stuff here so have been waiting till now for any 1080 review.

    Thanks!
  • D. Lister - Thursday, July 21, 2016 - link

    Nice job, Ryan. Good comeback. Keep it up.
  • Saeid92 - Thursday, July 21, 2016 - link

    What is 99th procentile framerate?
  • Ryan Smith - Thursday, July 21, 2016 - link

    If you sorted the framerate from highest to lowest, this would be the framerate of the slowest 1%. It's basically a more accurate/meaningful metric for minimum frame rates.
  • Eris_Floralia - Thursday, July 21, 2016 - link

    This is why I love Anandtech. Deep in reviews. Well I even wanted to be one of your editors if you have the plan to create a Chinese transtate version of these reviews.
  • daku123 - Thursday, July 21, 2016 - link

    Typo on FP16 Throughput page. In second paragraph, it should be Tegra X1 (not Tesla X1?).
  • Ryan Smith - Thursday, July 21, 2016 - link

    Eyup. Thanks!
  • Badelhas - Thursday, July 21, 2016 - link

    Great detailed review, as always. But I have to ask once again:
    why didnt you do some kind of VR Benchmarks? Thats what drives my choises now, to be honest.

    Cheers
  • Ranger1065 - Thursday, July 21, 2016 - link

    After over 2 months of reading GTX1080 reviews I felt a distinct lack of excitement
    as I read Anandtech kicking off their review of the finfet generation. Could it
    prove to be anything but an anticlimax?

    Sadly and unsurprisingly...NOT.

    It was however amusing to see the faithfull positively gushing praises for Anandtech
    now that the "greatly anticipated" review is finally out.

    Yes folks, 20 or so pages of (well written) information, mostly already covered by other tech sites,
    finally published, it's as if a magic wand has been waved, the information has been presented with
    that special Anandtech sauce, new insights have been illuminated and all is well in Anandtechland again.

    (AT LEAST UNTIL THE NEXT 2 MONTH DELAY.) LOL.

    I do like the way Anandtech presents the FPS charts.

    Back to sleep now Anandtech :)
  • mkaibear - Thursday, July 21, 2016 - link

    You've hit the nail on the head here Ranger.

    The info which is included within the article is indeed mostly already covered by other tech sites.

    Emphasis on the "mostly" and the plural "sites".

    Those of us who have jobs which keep us busy and have an interest in this sort of thing often don't have the time to trawl round many different sites to get reviews and pertinent technical data so we rely upon those sites which we trust to produce in-depth articles, even if they take a bit longer.

    As an IT Manager for (most recently) a manufacturing firm and then a school, I don't care about bleeding edge, get the new stuff as soon as it comes out, I care about getting the right stuff, and a two month delay to get a proper review is absolutely fine. If I need quick benchmarks I'll use someone like Hexus or HardOCP but to get a deep dive into the architecture so I can justify purchases to the Art and Media departments, or the programers is essential. You don't get that anywhere else.

Log in

Don't have an account? Sign up now