The AMD Radeon R9 Fury Review, Feat. Sapphire & ASUSby Ryan Smith on July 10, 2015 9:00 AM EST
A bit over two weeks ago AMD launched their new flagship video card, the Radeon R9 Fury X. Based on the company’s new Fiji GPU, the R9 Fury X brought with it significant performance improvements to AMD’s lineup, with AMD’s massive Fiji greatly increasing the card’s shading resources. Meanwhile Fiji also marked the introduction of High Bandwidth Memory (HBM) in to consumer products, giving the R9 Fury X a significant leg up in memory bandwidth. Overall AMD put together a very impressive card, however at $649 it fell just short of the GeForce GTX 980 Ti AMD needed it to beat.
Meanwhile alongside the announcement of the R9 Fury X, AMD announced that there would be three other Fiji-based cards. These include the R9 Fury, the R9 Nano, and a yet-to-be-named dual-GPU Fiji card. The first of these remaining cards to launch would be the R9 Fury, the obligatory lower-tier sibling to AMD’s flagship R9 Fury X. Today we will be taking a look at the first of those remaining cards, the R9 Fury, which launches next week.
While R9 Fury X remains the fastest Fiji card – and by virtue of being introduced first, the groundbreaking card – the impending launch of the R9 Fury brings with it a whole slew of changes that make it an interesting card in its own right, and a very different take on a Fiji product altogether. From a performance standpoint it is a lower performing card, featuring a cut-down Fiji GPU, but at the same time it is $100 cheaper than the R9 Fury X. Meanwhile in terms of construction, unlike the R9 Fury X, which is only available in its reference closed loop liquid cooling design, the R9 Fury is available as semi-custom and fully-custom cards from AMD’s board partners, built using traditional air coolers, making this the first air cooled Fiji card. As a result the R9 Fury at times ends up being a very different take on Fiji, for all of the benefits and drawbacks that comes with.
|AMD GPU Specification Comparison|
|AMD Radeon R9 Fury X||AMD Radeon R9 Fury||AMD Radeon R9 290X||AMD Radeon R9 290|
|Memory Clock||1Gbps HBM||1Gbps HBM||5Gbps GDDR5||5Gbps GDDR5|
|Memory Bus Width||4096-bit||4096-bit||512-bit||512-bit|
|Typical Board Power||275W||275W||250W||250W|
|Manufacturing Process||TSMC 28nm||TSMC 28nm||TSMC 28nm||TSMC 28nm|
|Architecture||GCN 1.2||GCN 1.2||GCN 1.1||GCN 1.1|
Starting things off, let’s take a look at the specifications of the R9 Fury. As we mentioned in our R9 Fury X review, we have known since the initial R9 Fury series launch that the R9 Fury utilizes a cut-down Fiji GPU, and we can now reveal just how it has been cut down. As is usually the case for these second-tier cards, the R9 Fury features both a GPU with some functional units disabled and a slightly reduced clockspeed, allowing AMD to recover partially defective GPUs while easing up on the clockspeed requirements.
The Fiji GPU in the R9 Fury ends up having 56 of 64 CUs enabled, which brings down the total stream processor count from 4,096 to 3,584. This in turn ends up being the full extent of the R9 Fury’s disabled functional units, as AMD has not touched the front-end or back-end, meaning the number of geometry units and the number of ROPs remained unchanged.
Also unchanged is the memory subsystem. All Fiji-based cards, including the R9 Fury, will be shipping with a fully enabled memory subsystem, meaning we’re looking at 4GB of HBM attached to the GPU over a 4096-bit memory bus. With Fiji topping out at just 4GB of memory in the first place – one of the drawbacks faced by the $650 R9 Fury X – cutting back on memory here to a smaller capacity is not a real option for AMD, so every Fiji card will come with that much memory.
As for clockspeeds, R9 Fury takes a slight trim on the GPU clockspeed. The reference clockspeed for the R9 Fury is a flat 1000MHz, a 5% reduction from the R9 Fury X. On the other hand the memory clock remains unchanged at 500MHz DDR, for an effective memory rate of 1Gbps/pin.
All told then, on paper the performance difference between the R9 Fury and R9 Fury X will stand to be between 0% and 17%; that is, the R9 Fury will be up to 17% slower than the R9 Fury X. In the best case scenario for the R9 Fury of a memory bandwidth bottleneck, it has the same 512GB/sec of memory bandwidth as the R9 Fury X. At the other end of the spectrum, in a shader-bound scenario, the combination of the reduction in shader hardware and clockspeeds is where the R9 Fury will be hit the hardest, as its total FP32 throughput drops from 8.6 TFLOPs to 7.17 TFLOPs. Finally in the middle, workloads that are front-end or back-end bound will see a much smaller drop since those units haven’t been cut-down at all, leading to just a 5% performance drop. As for the real world performance drop, as we’ll see it’s around 7%.
Power consumption on the other hand is going to be fairly similar to the R9 Fury X. AMD’s official Typical Board Power (TBP) for the R9 Fury is 275W, the same as its older sibling. Comparing the two products, the R9 Fury sees some improvement from the disabled CUs, however as a second-tier part it uses lower quality chips overall. Meanwhile the use of air cooling means that operating temperatures are higher than the R9 Fury X’s cool 65C, and as a result power loss from leakage is higher as well. At the end of the day this means that the R9 Fury is going to lose some power efficiency compared to the R9 Fury X, as any reduction in power consumption is going to be met with a larger decrease in performance.
Moving on, let’s talk about the cards themselves. With the R9 Fury X AMD has restricted vendors to selling the reference card, and we have been told it will be staying this way, just as it was for the R9 295X2. On the other hand for R9 Fury AMD has not even put together a complete reference design, leaving the final cards up to their partners. As a result next week’s launch will be a “virtual” launch, with all cards being semi or fully-custom.
Out of the gate the only partners launching cards are Sapphire and Asus, AMD’s closest and largest partners respectively. Sapphire will be releasing stock and overclocked SKUs based on a semi-custom design that couples the AMD reference PCB with Sapphire’s Tri-X cooler. Asus on the other hand has gone fully-custom right out of the gate, pairing up a new custom PCB with one of their DirectCU III coolers. Cards from additional partners will eventually hit the market, but not until later in the quarter.
The R9 Fury will be launching with an MSRP of $549, $100 below the R9 Fury X. This price puts the R9 Fury up against much different competition than its older sibling; instead of going up against NVIDIA’s GeForce GTX 980 Ti, the closest competition will be the older GeForce GTX 980. The official MSRP on that card is $499, so the R9 Fury is more expensive, but in turn AMD is promising better performance than the GTX 980. Otherwise NVIDIA’s partners serve to fill that $50 gap with their higher-end factory overclocked GTX 980 cards.
Finally, today’s reviews of the R9 Fury are coming slightly ahead of the launch of the card itself. As previously announced, the card goes on sale on Tuesday the 14th, however the embargo on the reviews is being lifted today. AMD has not officially commented on the launch supply, but once cards do go on sale, we’re expecting a repeat of the R9 Fury X launch, with limited quantities that will sell out within a day. After that, it seems likely that R9 Fury cards will remain in short supply for the time being, also similar to the R9 Fury X. R9 Fury X cards have come back in stock several times since the launch, but have sold out within an hour or so, and there’s currently no reason to expect anything different for R9 Fury cards.
|Summer 2015 GPU Pricing Comparison|
|Radeon R9 Fury X||$649||GeForce GTX 980 Ti|
|Radeon R9 Fury||$549|
|$499||GeForce GTX 980|
|Radeon R9 390X||$429|
|Radeon R9 290X
Radeon R9 390
|$329||GeForce GTX 970|
|Radeon R9 290||$250|
|Radeon R9 380||$200||GeForce GTX 960|
|Radeon R7 370
Radeon R9 270
|$130||GeForce GTX 750 Ti|
|Radeon R7 360||$110|
Post Your CommentPlease log in or sign up to comment.
View All Comments
Shadow7037932 - Friday, July 10, 2015 - linkYes! Been waiting for this review for a while.
Drumsticks - Friday, July 10, 2015 - linkIndeed! Good that it came out so early too :D
I'm curious @anandtech in general, given the likely newer state of the city/X's drivers, do you think that the performance deltas between each fury card and the respective nvidia will swing further or into AMD's favor as they solidify their drivers?
Samus - Friday, July 10, 2015 - linkSo basically if you have $500 to spend on a video card, get the Fury, if you have $600, get the 980 Ti. Unless you want something liquid cooled/quiet, then the Fury X could be an attractive albeit slower option.
Driver optimizations will only make the Fury better in the long run as well, since the 980Ti (Maxwell 2) drivers are already well optimized as it is a pretty mature architecture.
I find it astonishing you can hack off 15% of a cards resources and only lose 6% performance. AMD clearly has a very good (but power hungry) architecture here.
witeken - Friday, July 10, 2015 - linkNo, not at all. You must look at it the other way around: Fury X has 15% more resources, but is <<15% faster.
0razor1 - Friday, July 10, 2015 - linkSmart , you :) :D This thing is clearly not balanced. That's all there is to it. I'd say x for the WC at 100$ more make prime logic.
thomascheng - Saturday, July 11, 2015 - linkBalance is not very conclusive. There are games that take advantage of the higher resources and blows past the 980Ti and there are games that don't and therefore slower. Most likely due to developers not having access to Fury and it's resources before. I would say, no games uses that many shading units and you won't see a benefit until games do. The same with HBM.
FlushedBubblyJock - Wednesday, July 15, 2015 - linkWhat a pathetic excuse, apologists for amd are so sad.
AMD got it wrong, and the proof is already evident.
No, NONE OF US can expect anandtech to be honest about that, nor it's myriad of amd fanboys,
but we can all be absolutely certain that if it was nVidia whom had done it, a full 2 pages would be dedicated to their massive mistake.
I've seen it a dozen times here over ten years.
When will you excuse lie artists ever face reality and stop insulting everyone else with AMD marketing wet dreams coming out of your keyboards ?
Will you ever ?
redraider89 - Monday, July 20, 2015 - linkAnd you are not an nividia fanboy are you? Hypocrite.
redraider89 - Monday, July 20, 2015 - linkTypical fanboy, ignore the points and go straight to name calling. No, you are the one people shold be sad about, delusional that they are not a fanboy when they are.
redraider89 - Monday, July 20, 2015 - linkProof that intel and nvidia wackos are the worst type of people, arrogant, snide, insulting, childish. You are the poster boy for an intel/nvidia sophomoric fanboy.