AMD Radeon HD 7970 Review: 28nm And Graphics Core Next, Together As One
by Ryan Smith on December 22, 2011 12:00 AM EST- Posted in
- GPUs
- AMD
- Radeon
- ATI
- Radeon HD 7000
A Quick Refresher: Graphics Core Next
One of the things we’ve seen as a result of the shift from pure graphics GPUs to mixed graphics and compute GPUs is how NVIDIA and AMD go about making their announcements and courting developers. With graphics GPUs there was no great need to discuss products or architectures ahead of time; a few choice developers would get engineering sample hardware a few months early, and everyone else would wait for the actual product launch. With the inclusion of compute capabilities however comes the need to approach launches in a different manner, a more CPU-like manner.
As a result both NVIDIA and AMD have begun revealing their architectures to developers roughly six months before the first products launch. This is very similar to how CPU launches are handled, where the basic principles of an architecture are publically disclosed months in advance. All of this is necessary as the compute (and specifically, HPC) development pipeline is far more focused on optimizing code around a specific architecture in order to maximize performance; whereas graphics development is still fairly abstracted by APIs, compute developers want to get down and dirty, and to do that they need to know as much about new architectures as possible as soon as possible.
It’s for these reasons that AMD announced Graphics Core Next, the fundamental architecture behind AMD’s new GPUs, back in June of this year at the AMD Fusion Developers Summit. There are some implementation and product specific details that we haven’t known until now, and of course very little was revealed about GCN’s graphics capabilities, but otherwise on the compute side AMD is delivering on exactly what they promised 6 months ago.
Since we’ve already covered the fundamentals of GCN in our GCN preview and the Radeon HD 7970 is primarily a gaming product we’re not going to go over GCN in depth here, but I’d encourage you to read our preview to fully understand the intricacies of GCN. But if you’re not interested in that, here’s a quick refresher on GCN with details pertinent to the 7970.
As we’ve already seen in some depth with the Radeon HD 6970, VLIW architectures are very good for graphics work, but they’re poor for compute work. VLIW designs excel in high instruction level parallelism (ILP) use cases, which graphics falls under quite nicely thanks to the fact that with most operations pixels and the color component channels of pixels are independently addressable datum. In fact at the time of the Cayman launch AMD found that the average slot utilization factor for shader programs on their VLIW5 architecture was 3.4 out of 5, reflecting the fact that most shader operations were operating on pixels or other data types that could be scheduled together
Meanwhile, at a hardware level VLIW is a unique design in that it’s the epitome of the “more is better” philosophy. AMD’s high steam processor counts with VLIW4 and VLIW5 are a result of VLIW being a very thin type of architecture that purposely uses many simple ALUs, as opposed to fewer complex units (e.g. Fermi). Furthermore all of the scheduling for VLIW is done in advance by the compiler, so VLIW designs are in effect very dense collections of simple ALUs and cache.
The hardware traits of VLIW mean that for a VLIW architecture to work, the workloads need to map well to the architecture. Complex operations that the simple ALUs can’t handle are bad for VLIW, as are instructions that aren’t trivial to schedule together due to dependencies or other conflicts. As we’ve seen graphics operations do map well to VLIW, which is why VLIW has been in use since the earliest pixel shader equipped GPUs. Yet even then graphics operations don’t achieve perfect utilization under VLIW, but that’s okay because VLIW designs are so dense that it’s not a big problem if they’re operating at under full efficiency.
When it comes to compute workloads however, the idiosyncrasies of VLIW start to become a problem. “Compute” covers a wide range of workloads and algorithms; graphics algorithms may be rigidly defined, but compute workloads can be virtually anything. On the one hand there are compute workloads such as password hashing that are every bit as embarrassingly parallel as graphics workloads are, meaning these map well to existing VLIW architectures. On the other hand there are tasks like texture decompression which are parallel but not embarrassingly so, which means they map poorly to VLIW architectures. At one extreme you have a highly parallel workload, and at the other you have an almost serial workload.
Cayman, A VLIW4 Design
So long as you only want to handle the highly parallel workloads VLIW is fine. But using VLIW as the basis of a compute architecture is going is limit what tasks your processor is sufficiently good at. If you want to handle a wider spectrum of compute workloads you need a more general purpose architecture, and this is the situation AMD faced.
But why does AMD want to chase compute in the first place when they already have a successful graphics GPU business? In the long term GCN plays a big part in AMD’s Fusion plans, but in the short term there’s a much simpler answer: because they have to.
In Q3’2011 NVIDIA’s Professional Solutions Business (Quadro + Tesla) had an operating income of 95M on 230M in revenue. Their (consumer) GPU business had an operating income of 146M, but on a much larger 644M in revenue. Professional products have much higher profit margins and it’s a growing business, particularly the GPU computing side. As it stands NVIDIA and AMD may have relatively equal shares of the discrete GPU market, but it’s NVIDIA that makes all the money. For AMD’s GPU business it’s no longer enough to focus only on graphics, they need a larger piece of the professional product market to survive and thrive in the future. And thus we have GCN.
292 Comments
View All Comments
CeriseCogburn - Sunday, March 11, 2012 - link
We'll have to see if amd "magically changes that number and informs Anand it was wrong like they did concerning their failed recent cpu.... LOLThat's a whole YEAR of lying to everyone trying to make their cpu look better than it's actual fail, and Anand shamefully chose to announce the number change "with no explanation given by amd"... -
That's why you should be cautious - we might find out the transistor count is really 33% different a year from now.
piroroadkill - Thursday, December 22, 2011 - link
Only disappointing if you:a) ignored the entire review
b) looked at only the chart for noise
c) have brain damage
Finally - Thursday, December 22, 2011 - link
In Eyefinity setups the new generation shines: http://tinyurl.com/bu3wb5cwicko - Thursday, December 22, 2011 - link
I think the price is disappointing. Everything else is nice though.CeriseCogburn - Sunday, March 11, 2012 - link
The drivers suckRussianSensation - Thursday, December 22, 2011 - link
Not necessarily. The other possibility is that being 37% better on average at 1080P (from this Review) over HD6970 for $320 more than an HD6950 2GB that can unlock into a 6970 just isn't impressive enough. That should be d).piroroadkill - Friday, December 23, 2011 - link
Well, I of course have a 6950 2GB that unlocked, so as far as I'm concerned, that has been THE choice since the launch of the 6950, and still is today.But you have to ignore cost at launch, it's always high.
CeriseCogburn - Thursday, March 8, 2012 - link
I agree RS, as these amd people are constantly screaming price percentage increase vs performance increase... yet suddenly applying the exact combo they use as a weapon against Nvidia to themselves is forbidden, frowned upon, discounted, and called unfair....Worse yet, according to the same its' all Nvidia's fault now - that amd is overpriced through the roof...LOL - I have to laugh.
Also, the image quality page in the review was so biased toward amd that I thought I was going to puke.
Amd is geven credit for a "perfect algorythm" that this very website has often and for quite some time declared makes absolutely no real world difference in games - and in fact, this very reviewer admitted the 1+ year long amd failure in this area as soon as they released "the fix" - yet argued everyone else was wrong for the prior year.
The same thing appears here.
Today we find out the GTX580 nvidia card has much superior anti-shimmering than all prior amd cards, and that finally, the 7000 high end driver has addressed the terrible amd shimmering....
Worse yet, the decrepit amd low quality impaired screens are allowed in every bench, with the 10% amd performance cheat this very site outlined them merely stated we hope Nvidia doesn't so this too - then allowed it, since that year plus ago...
In the case of all the above, I certainly hope the high end 797x cards aren't CHEATING LIKE HECK still.
For cripe sakes, get the AA stuff going, stop the 10% IQ cheating, and get our bullet physics or pay for PhysX, and stabilize the drivers .... I am sick of seeing praise for cheating and failures - if they are (amd) so great let's GET IT UP TO EQUIVALENCY !
Wow I'm so mad I don't have a 7970 as supply is short and I want to believe in amd for once... FOR THE LOVE OF GOD DID THEY GET IT RIGHT THIS TIME ?!!?
slayernine - Thursday, December 22, 2011 - link
Holy fan boys batman!This comment thread reeks of nvidia fans green with jealousy
Hauk - Thursday, December 22, 2011 - link
LOL, Wreckage first!Love him or hate him, he's got style..