Meet The Sapphire Tri-X R9 Fury OC

Today we’ll be looking at Fury cards from both Sapphire and Asus. We’ll kick things off with Sapphire’s card, the Tri-X R9 Fury OC.

Radeon R9 Fury Launch Cards
  ASUS STRIX R9 Fury Sapphire Tri-X R9 Fury Sapphire Tri-X R9 Fury OC
Boost Clock 1000MHz /
1020MHz (OC)
1000MHz 1040MHz
Memory Clock 1Gbps HBM 1Gbps HBM 1Gbps HBM
VRAM 4GB 4GB 4GB
Maximum ASIC Power 216W 300W 300W
Length 12" 12" 12"
Width Double Slot Double Slot Double Slot
Cooler Type Open Air Open Air Open Air
Launch Date 07/14/15 07/14/15 07/14/15
Price $579 $549 $569

Sapphire is producing this card in two variants, a reference clocked version and a factory overclocked version. The version we’ve been sampled is the factory overclocked version, though other than some basic binning to identify cards that can handle being overclocked, the two cards are physically identical.

As far as Sapphire’s overclock goes, it’s a mild overclock, with the card shipping at 1040MHz for the GPU while the memory remains unchanged at 1Gbps. As we discussed in our R9 Fury X review, Fiji cards so far don’t have much in the way of overclocking headroom, so AMD’s partners have to take it easy on the factory overclocks. Sapphire’s overclock puts the upper-bound of any performance increase at 4% – with the real world gains being smaller – so this factory overclock is on the edge of relevance.

Getting down to the nuts and bolts then, Sapphire’s card is a semi-custom design, meaning Sapphire has paired an AMD reference PCB with a custom cooler. The PCB in question is AMD’s PCB from the R9 Fury X, so there’s little new to report here. The PCB itself measures 7.5” long and features AMD’s 6 phase power design, which is designed to handle well over 300W. For overclockers there is still no voltage control options available for this board design, though as Sapphire has retained AMD’s dual BIOS functionality there’s plenty of opportunity for BIOS modding.

The real story here is Sapphire’s Tri-X cooler, which gets the unenviable job of replacing AMD’s closed loop liquid cooler from the R9 Fury X. With a TBP of 275W Sapphire needs to be able to dissipate quite a bit of heat to keep up with Fiji, which has led to the company using one of their Tri-X coolers. We’ve looked at a few different Tri-X cards over the years, and they have been consistently impressive products. For the Tri-X R9 Fury, Sapphire is aiming for much the same.

Overall the Tri-X cooler used on the Tri-X R9 Fury ends up being quite a large cooler. Measuring a full 12” long it runs the length of the PCB and then some, meanwhile with that much copper and aluminum it’s not a light card either. The end result is that with such a large cooler the card is better defined as a PCB mounted on a cooler than a cooler mounted on a PCB, an amusing juxtaposition from the usual video card. As a result of this Sapphire has gone the extra mile to ensure that the PCB can support the cooler; there are screws in every last mounting hole, there’s a full-sized backplate to further reinforce the card, and the final 4.5” of the cooler that isn’t mounted to the PCB has its own frame to keep that secure as well.

Moving to the top of the card, the Tri-X R9 Fury features three of Sapphire’s 90mm “Aerofoil” fans, the company’s larger, dual ball bearing fans. These fans are capable of moving quite a bit of air even when moving at relatively low speeds, and as a result the overall card noise is kept rather low even under load, as we’ll see in full detail in our benchmark section.

Meanwhile Sapphire has also implemented their version of zero fan speed idle on the Tri-X R9 Fury, dubbed Intelligent Fan Control, which allows the card to turn off its fans entirely when their cooling capacity isn’t needed. With such a large heatsink the Fiji GPU and supporting electronics don’t require active cooling when idling, allowing Sapphire to utilize passive cooling and making the card outright silent at idle. This is a feature a number of manufacturers have picked up on in the last couple of years, and the silent idling this allows is nothing short of amazing. For Sapphire’s implementation on the Tri-X R9 Fury, what we find is that the fans finally get powered up at around 53C, and power down when the temperature falls below 44C.

Sapphire Tri-X R9 Fury Zero Fan Idle Points
  GPU Temperature Fan Speed
Turn On 53C 27%
Turn Off 44C 23%

Helping the cooling effectiveness of the Tri-X quite a bit is the length of the fans and heatsink relative to the length of the PCB. With the 4.5” of overhang, the farthest fan is fully beyond the PCB. That means that all of the air it pushes through the heatsink doesn’t get redirected parallel to the card – as is the case normally for open air cards – but rather the hot air goes straight through the heatsink and past it. For a typical tower case this means that hot air goes straight up towards the case’s exhaust fans, more efficiently directing said hot air outside of the case and preventing it from being recirculated by the card’s fans. While this doesn’t make a night & day difference in cooling performance, it’s a neat improvement that sidesteps the less than ideal airflow situation the ATX form factor results in.

Moving on, let’s take a look at the heatsink itself. The Tri-X’s heatsink runs virtually the entire length of the card, and is subdivided into multiple segments. Connecting these segments are 7 heatpipes, ranging in diameter between 6mm and 10mm. The heatpipes in turn run through both a smaller copper baseplate that covers the VRM MOSFETs, and a larger copper baseplate that covers the Fiji GPU itself. Owners looking to modify the card or otherwise remove the heatsink will want to take note here; we’re told that it’s rather difficult to properly reattach the heatsink to the card due to the need to perfectly line up the heatsink and mate it with the GPU and the HBM stacks.

The Tri-X R9 Fury’s load temperatures tend to top out at 75C, which is the temperature limit Sapphire has programmed the card for. As with the R9 Fury X and the reference Radeon 290 series before that, Sapphire is utilizing AMD’s temperature and fan speed target capabilities, so while the card will slowly ramp up the fan to 75C, once it hits that temperature it will more greatly ramp up the fan to keep the temperature at or below 75C.

Moving on, since Sapphire is using AMD’s PCB, this means the Tri-X also inherits the former’s BIOS and lighting features. The dual-BIOS switch is present, and Sapphire ships the card with two different BIOSes. The default BIOS (switch right) uses the standard 300W ASIC power limit and 75C temperature target. Meanwhile the second BIOS (switch left) Increases the power and temperature limits to 350W and 80C respectively, for greater overclocking limits. Note however that this doesn’t change the voltage curve, so Fury cards in general will still be held back by a lack of headroom at stock voltages. As for the PCB’s LEDs, Sapphire has retained those as well, though they default to blue (sapphire) rather than AMD red.

Finally, since this is the AMD PCB, display I/O remains unchanged. This means the Tri-X offers 3x DisplayPorts along with a single HDMI 1.4 port.

Wrapping things up, the OC version we are reviewing today will retail for $569, $20 over AMD’s MSRP. The reference clocked version on the other hand will retail at AMD’s MSRP of $549, the only launch card that will be retailing at this price. Finally, Sapphire tells us that the OC version will be the rarer of the two due to its smaller run, and that the majority of Tri-X R9 Fury cards that will be on sale will be the reference clocked version.

The AMD Radeon R9 Fury Review Meet The ASUS STRIX R9 Fury
Comments Locked

288 Comments

View All Comments

  • Midwayman - Friday, July 10, 2015 - link

    I'd love to see these two go at it again once dx12 games start showing up.
  • Mugur - Saturday, July 11, 2015 - link

    Bingo... :-). I bet the whole Fury lineup will gain a lot with DX12, especially the X2 part (4 + 4 GB won't equal 4 as in current CF). The are clearly CPU limited at this point.
  • squngy - Saturday, July 11, 2015 - link

    I don't know...

    Getting dx12 performance at the cost of dx11 performance sounds like a stupid idea this soon before dx12 games even come out.

    By the time a good amount of dx12 games come out there will probably be new graphics cards available.
  • thomascheng - Saturday, July 11, 2015 - link

    They will probably circle around and optimize things for 1080p and dx11, once dx12 and 4k is at a good place.
  • akamateau - Tuesday, July 14, 2015 - link

    DX12 games are out now. DX12 does not degrade DX11 performance. In fact Radeon 290x is 33% faster than 980 Ti in DX12. Fury X just CRUSHES ALL nVIDIA silicon with DX12 and there is a reason for it.

    Dx11 can ONLY feed data to the GPU serially and sequencially. Dx12 can feed data Asynchronously, the CPU send the data down the shader pipeline WHEN it is processed. Only AMD has this IP.
  • @DoUL - Sunday, July 19, 2015 - link

    Kindly provide link to a single DX12 game that is "out now".

    In every single review of the GTX 980 Ti there is this slide of DX12 feature set that the GTX 980 Ti supports and in that slide in all the reviews "Async Compute" is right there setting in the open, so I'm not really sure what do you mean by "Only AMD has this IP"!

    I'd strongly recommend that you hold your horses till DX12 games starts to roll out, and even then, don't forget the rocky start of DX11 titles!

    Regarding the comparison you're referring to, that guy is known for his obsession with mathematical calculations and synthetic benchmarking, given the differences between real-world applications and numbers based on mathematical calculations, you shouldn't be using/taking his numbers as a factual baseline for what to come.
  • @DoUL - Sunday, July 19, 2015 - link

    My Comment was intended as a reply to @akanateau
  • OldSchoolKiller1977 - Sunday, July 26, 2015 - link

    You are an idiotic person, wishful think and dreams don't make you correct. As stated please provide a link to these so called DX12 games and your wonderful "Fury X just CRUCHES ALL NVidia" statement.
  • Michael Bay - Sunday, July 12, 2015 - link

    As long as there is separate RAM in PCs, memory argument is moot, as contents are still copied and executed on in two places.
  • akamateau - Tuesday, July 14, 2015 - link

    Negative. Once Graphic data is processed and sent to the shaders it next goes to VRAM or video ram.

    System ram is what the CPU uses to process object draws. Once the objects are in the GPU pipes system ram is irrelevant.

    IN fact that is one of AMD's stacked memory patents. AMD will be putting HBM on APU's to not only act as CPU cache but HBM video ram as well. They have patents for programmable HBM using FPGA's and reconfigurable cache memory HBM as well.

    Stacked memory HBM can also be on the cpu package as a replacement for system ram. Can you imagine how your system would fly with 8-16gb of HBM instead of system ram?

Log in

Don't have an account? Sign up now