The AMD Vega GPU Architecture Teaser: Higher IPC, Tiling, & More, Coming in H1’2017by Ryan Smith on January 5, 2017 9:00 AM EST
As AMD was in the process of ramping up for the Polaris launch last year, one of the unexpected but much appreciated measures they took was to release a bare-bones GPU architecture roadmap for the next few years. AMD has traditionally held their cards very close to their proverbial chest on what they’re working on next, typically only announcing a new architecture weeks before it’s set to launch in retail products. Polaris itself was a departure from that, as it was announced a good 5 months in advance, but last year’s roadmap was the first honest-to-goodness look we’ve had at AMD’s long-term plans in a long time.
What did that map show us? After 2016’s Polaris would come Vega, and after that would be Navi. As a high-level roadmap it didn’t show us much – in fact other than a timeframe, the only detail attached to Vega was “HBM2” - but it was enough to help understand one of the things AMD would be doing architecturally to set Vega apart from Polaris. As for the timeframe itself, that was ambiguous at best in AMD’s roadmap. But now as we draw closer to the launch of Vega, the picture has become clearer. AMD will be hitting a yearly cadence with Vega. The first chip, which tapped out last year, will be launching in the first half of this year (H1’17).
To that end, with Vega’s launch not too far over the horizon, AMD is ready to start talking about what will be their next GPU architecture. Last year at this time we got our first real glimpse into Polaris and what would become the Radeon RX 480/470/460, and this year AMD is back again with a teaser of things to come with Vega.
Setting The Stage: Expectations Management; Less How & More Why
Before we dive into any architectural details, perhaps it’s best we first set the stage. This goes for both what to expect of today’s announcement, and to better understand what AMD is doing and why.
First and foremost, today’s detail release is a teaser, not a deep dive, or even a preview. AMD is only releasing a few details about Vega, and those are being kept at a high level. In fact it’s fair to say that there’s just enough information to answer little and raise even more questions; just what a proper teaser should be.
Why? Well part of the reason is that we’re still months off from the launch of Vega. I believe it’s fair to say that by announcing a first-half of the year launch date when we’re already in 2017 is a strong indicator that Vega will not launch until later in that window, likely some time in Q2. So we’re still a good three to five months out from the launch of Vega, which means AMD doesn’t want to (or need to) release too many details this far out. Rather they can trickle out chosen details for maximum impact.
At the same time the AMD of 2017 has more they can focus on in the high-performance space than just GPUs. Ryzen launches soon, and they also have other products on the horizon such as the Radeon Instinct accelerators. Polaris received as much detail as it did because it was all AMD really had to talk about, and they needed to recover from a rough 2015 where AMD’s at-the-time power efficiency woes were brought into full focus. But now Vega can share the stage with Ryzen and other products, and that lets AMD be more selective about what they say.
All of which is something I would argue is a good thing. At the end of the day Polaris was an optimized version of the GCN 1.2 (aka GCN 3) architecture for GlobalFoundries’ 14nm FinFET process. The resulting GPUs were solid competitors in the mainstream and value markets, improving on AMD’s power efficiency in a way they badly needed. But they weren’t high-end parts; they didn’t excite like those parts did, and for technology enthusiasts they didn’t significantly change the architecture itself (in fact GCN 4 was ISA compatible with GCN 3, something that doesn’t happen a lot in the GPU space). AMD talked big about Polaris – perhaps too big – and I do think it hurt them in some circles once it became clearer that this was AMD catching up. Which is not to say that AMD’s marketing arm won’t talk big about Vega as well, but they need not ride the technology angle so hard. Vega is a launch that can be more natural and more subdued, especially as at this point we know AMD is aiming big with a much-needed new generation of high-end parts.
In any case, as AMD isn’t riding the technology angle quite as hard in this year’s teaser, they are spending a bit more time explaining the market and some of the logic behind Vega’s design. For its teasing debut, Vega is little less discussion of “how,” and a little more conversation of “why”.
So what is AMD looking to do with Vega? Besides aiming for the high-end of the market, AMD is looking at how the market for GPUs has changed in the last half-decade, and what they need to do to address it. Machine learning is one part of that, being a market that has practically sprung up overnight to become a big source of revenue for GPUs. This is where the previously announced Radeon Instinct will fit in.
But more than that, it’s about fundamental shifts in how workloads are structured. GPU performance growth has far outpaced GPU memory capacity. Scene geometry complexity has continued to grow. Newer rendering methods have significantly changed GPU memory access patterns.
To that end, AMD is looking to address all of these factors with Vega. Which is not to say that this is everything – this is a teaser, after all – but this is where AMD is starting. Where they are going to be with their next generation architecture and how they believe it will address the changes in the market. So without further ado, let’s take a teasing look at what the future has in store for AMD’s GPUs.
Post Your CommentPlease log in or sign up to comment.
View All Comments
FireSnake - Thursday, January 5, 2017 - linkve.ga is not accessible .... no Ryzen servers yet, haha :)
Michael Bay - Thursday, January 5, 2017 - linkDesktop graphics game is lost. AMD should just close all non-gpu divisions and outright reorient towards producing specific ASIC solutions for miners, this newfangled machine bullshit and such.
DATS WHERE DEM MONES BE BRAH
rpmrg - Thursday, January 5, 2017 - linkHere's another armchair analyst.
Michael Bay - Thursday, January 5, 2017 - linkAt least try to hide your envy next time, honey.
lobz - Thursday, January 5, 2017 - linkyeah rpmrg dude, don't you ever be envious of his armchair...
ddriver - Thursday, January 5, 2017 - linkPeople are confusing "armchair analyst" with "idiot" :)
While it is disappoint for me, as I am predominantly interested in compute, vega has the makings of a good gaming GPU, disappointing for me as this comes from sacrificing high-precision performance for the sake of efficiency and optimizing for low-prevision performance, which is the precision which game graphics utilize.
extide - Thursday, January 5, 2017 - linkWell, the only bit about DP FP on the slides mentioned that it was 'configurable' -- so we don't know what we will see 1/2, 1/3, 1/4... ? Rate vs SP DP? AMD has always been more gratuitous with not disabling that higher end compute on consumer parts -- but we wont know what's actually in the silicon until later. I wouldn't be surprised to see at least one 1/2 rate DP FP part ship at some point, whether it be sold as a consumer card or only as a pro card, only time will tell.
eachus - Saturday, January 14, 2017 - linkAMD has stated that Vega 20 (due in 2H18) will have 1/2 DPFP, while Vega 10 will be 1/16th. The configurable means that different chips can have different double precision rates. There is at least one additional Vega chip in the works (Vega 11) plus Navi. However, I think that 1/2 is as fast as we will see. ;-)
Michael Bay - Friday, January 6, 2017 - linkBut not in your case. ^_^
Alexvrb - Friday, January 6, 2017 - linkWhy not, his armchair is the brains of the operation! Seriously, he's totes right though... they should just throw away GPU revenue. Take chips that could otherwise be sold in several product categories, and only sell them as "newfangled machine bullshit" chips. Plus some low-margin ephemeral ASICs for miners. See, it's brilliant. Cuts down volume which saves on shipping costs. Plus it means less wafers so you don't get those pesky economies of scale benefits. Nvidia would obviously go this route too but they just don't have the balls to stop selling graphics cards. Excelsior!!