Final Words

Bringing this review to a close, NVIDIA’s latest product launch has given us quite a bit to digest. Not only are we looking at NVIDIA’s latest products for the high volume mainstream desktop video card market, but we’re doing so through the glasses of a new generation of GPUs. With the GeForce GTX 750 series we are seeing our first look at what the next generation of GPUs will hold for NVIDIA, and if these cards are an accurate indication of what’s to follow then we’re being setup for quite an interesting time.

Starting from an architectural point of view, it’s clear from the very start that Maxwell is both a refresh of the Kepler architecture and at the same time oh so much more. I think from a feature perspective it’s going to be difficult not to be a bit disappointed that NVIDIA hasn’t pushed the envelope here in some manner, leaving us with a part that as far as features go is distinctly Kepler. Complete support for Direct3D 11.1 and 11.2, though not essential, would have been nice to have so that 11.2 could be the standard for new video cards in 2014. Otherwise I’ll fully admit I don’t know what else to expect of Maxwell – the lack of a new Direct3D standard leaves this as something of a wildcard – but it means that there isn’t a real marquee feature for the architecture to evaluate and marvel at.

On the other hand, the lack of a significant feature changes means that it’s much easier to evaluate Maxwell next to Kepler in the area where NVIDIA did focus: efficiency. This goes for power efficiency resource/compute efficiency, and space efficiency. Utilizing a number of techniques NVIDIA set out to double their performance per watt versus Kepler – a design that was already power efficient by desktop GPU standards – and it’s safe to say that they have accomplished this. With a higher resource efficiency giving NVIDIA additional performance with less hardware, and power optimizations bringing that power consumption down by dozens of watts, NVIDIA has done what in previous generations would have taken a die shrink. The tradeoff is that NVIDIA doesn’t have that die shrink, so die sizes grow in the process, but even then the fact that they packed so much more hardware into GM107 for only a moderate increase in die size is definitely remarkable from an engineering perspective.

Efficiency aside, Maxwell’s architecture is something of an oddity at first, but given NVIDIA’s efficiency gains it’s difficult to argue with the outcome. The partitioning of the SMM means that we have partitions that feel a lot like GF100 SMs, which has NVIDIA going backwards in a sense due to the fact that significant resource sharing was something that first became big with Kepler. But perhaps that was the right move all along, as evidenced by what NVIDIA has achieved. On the other hand the upgrade of the compute feature set to GK110 levels is good news all around. The increased efficiency it affords improves performance alongside the other IPC improvements NVIDIA has worked in, plus it means that some of GK110’s more exotic features such as dynamic parallelism and HyperQ are now a baseline feature. Furthermore the reduction in register pressure and memory pressure all around should be a welcome development; compared to GK107 there are now more registers per thread, more registers per CUDA core, more shared memory per CUDA core, and a lot more L2 cache per GPU. All of which should help to alleviate memory related stalls, especially as NVIDIA is staying on the 128-bit bus.

With that in mind, this brings us to the cards themselves. By doubling their performance-per-watt NVIDIA has significantly shifted their performance both with respect to their own product lineup and AMD’s lineup. The fact that the GTX 750 Ti is nearly 2x as fast as the GTX 650 is a significant victory for NVIDIA, and the fact that it’s nearly 3x faster than the GT 640 – officially NVIDIA’s fastest 600 series card without a PCIe power plug requirement – completely changes the sub-75W market. NVIDIA wants to leverage GM107 and the GTX 750 series to capture this market for HTPC use and OEM system upgrades alike, and they’re in a very good position to do so. Plus it goes without saying that compared to last-generation cards such as the GeForce GTX 550 Ti, NVIDIA has finally doubled their performance (and halved their power consumption!), for existing NVIDIA customers looking for a significant upgrade from older GF106/GF116 cards.

But on a competitive basis things are not so solidly in NVIDIA’s favor. NVIDIA does not always attempt to compete with AMD on a price/performance basis in the mainstream market, as their brand and retail presence gives them something they can bank on even when they don’t have the performance advantage. In this case NVIDIA has purposely chosen to forgo chasing AMD for the price/performance lead, and as such for the price the GeForce GTX 750 cards are the weaker products. Radeon R7 265 holds a particularly large 19% lead over GTX 750 Ti, and in fact wins at every single benchmark. Similarly, Radeon R7 260X averages a 10% lead over GTX 750, and it does so while having 2GB of VRAM to GTX 750’s 1GB.

On a pure price/performance basis, the GTX 750 series is not competitive. If you’re in the sub-$150 market and looking solely at performance, the Radeon R7 260 series will be the way to go. But this requires forgoing NVIDIA’s ecosystem and their power efficiency advantage; if either of those matter to you, then the lower performance of the NVIDIA cards will be justified by their other advantages. With that said however, we will throw in an escape clause: NVIDIA has hard availability today, while AMD’s Radeon R7 265 cards are still not due for about another 2 weeks. Furthermore it’s not at all clear if retailers will hold to their $149 MSRP due to insane demand from cryptocoin miners; if that happens then NVIDIA’s competition is diminished or removed entirely, and NVIDIA wins on price/performance by default.

Wrapping things up, as excited as we get and as focused as we are on desktop cards, it’s hard not to view this launch as a preview of things to come. With laptop sales already exceeding desktop sales, it’s a foregone conclusion that NVIDIA will move more GM107 based video cards in mobile products than they will in desktops. With GK107 already being very successful in that space and GM107 doubling NVIDIA’s performance-per-watt – and thereby doubling their performance in those power-constrained devices – it means that GM107 is going to be an even greater asset in the mobile arena. To that end it will be very interesting to see what happens once NVIDIA starts releasing the obligatory mobile variants of the GTX 750 series, as what we’ve seen today tells us that we could be in for a very welcome jump in mobile performance.

Overclocking: When Headroom Exceeds Clockspeed Limits
POST A COMMENT

184 Comments

View All Comments

  • MrSpadge - Tuesday, February 18, 2014 - link

    To be fair GTX650Ti Boost consumes ~100 W in the real world. Still a huge improvement! Reply
  • NikosD - Tuesday, February 18, 2014 - link

    Hello.

    I have a few questions regarding HTPC and video decoding.

    Can we say that we a new video processor from Nvidia, a new name like VP6 or more like a VP5.x ?

    How Nvidia is calling the new video decoder ?

    Why don't you add a 4K60 fps clip in order to test soon to be released HDMI 2.0 output ?

    If you run a benchmark using DXVA Checker between VP5 and VP6 (?) how much faster is VP6 in H.264 1080p, 4K clips ?

    Thanks!
    Reply
  • Ryan Smith - Thursday, February 20, 2014 - link

    NVIDIA doesn't have a name for it; at least not one they're sharing with us. Reply
  • NikosD - Thursday, February 20, 2014 - link

    Thanks.
    Is it possible to try a 4K60fps with Maxwell ?

    I wonder if it can decode it in realtime...
    Reply
  • Flunk - Tuesday, February 18, 2014 - link

    I think these will be a lot more exciting in laptops. Even if they're no where near Nvidia's claimed 2x Kepler efficiency per watt. On the desktop it's not really that big a deal. The top-end chip will probably be ~40% faster than the 780TI but that will be a while. Reply
  • dylan522p - Tuesday, February 18, 2014 - link

    the 880 will be much more powerful than the 780ti. More than 40% even. They could literally die shrink and throw a few more SMX's and the 40% would be achieved. I would imagine either they are gonna have a HUGE jump (80% +) or they are gonna do what they did with Kepler and release a 200W Sku that is about 50% faster and when 20nm yields are good enough have the 900 series come with 250W Skus. Reply
  • npz - Tuesday, February 18, 2014 - link

    For HTPC or general video offloading support, you'd be better off with a powerful cpu. The article states rendering issues in certain circumstances using the GPU and there is no mention of reference frames, 10-bit or colorspace support beyond YUV 4:2:0 so I will assume these to be unrectified by the fixed function decoders/encoders (at least until HEVC since these are part of its higher profiles). Reply
  • npz - Tuesday, February 18, 2014 - link

    I also don't trust the quality of NVENC at all, especially at lower bitrates vs x264 (and especially with x264 psycho visual optimizations). And it the lower bitrates is where such hw encoders will be most useful for real time streaming. x264's fast preset can already do greater than realtime and produces better results. Reply
  • Egg - Tuesday, February 18, 2014 - link

    I'm pretty sure that on HTPCs the idea is to use madVR, which is GPU assisted. Reply
  • npz - Tuesday, February 18, 2014 - link

    madVR only helps with rendering and from these tests, full GPU offloaded rendering (scaling, colorspace conversion to RGB, deinterlacing, post-processing, video mixing etc) is broken for some specs. It's useful when it works though. But I'd rather have a good cpu and basic gpu good enough the final portion of the rendering pipeline, that's guaranteed to play anything in any format/spec. Reply

Log in

Don't have an account? Sign up now