Feature Article

AMD Radeon R9 Fury X Review

GameSpot may receive revenue from affiliate and advertising partnerships for sharing this content and from purchases through links.

A new 4K flagship, locked and loaded.

To some people, AMD has fallen behind Nvidia because of its lack of response to cards like the GTX 980 Ti and GTX Titan X, but that stops now, with the red team unleashing its newest flagship GPU: the Radeon R9 Fury X.

It's an interesting chip that impresses in several key departments. AMD's familiar architecture has been upgraded in every important area, and the corporation has ditched air-cooling for a separate water-cooling module that it promises can keep the beefy chip chilled.

This amount of hardware doesn't come cheap, however, which means you'll have to shell out $650 for the Fury X. That's similar cash to the Nvidia GeForce GTX 980 Ti, which is the Fury’s main 4K rival.

No Caption Provided

Radeon Fury X Specs

AMD’s new GPU is called Fiji, and it turbo-charges the existing Graphics Core Next architecture to help the Fury compete with the 980 Ti.

The core is constructed from four Shader Engines, with each divided into 16 Compute Units--and each of those is rammed with 64 stream processors. That means the Fury X is built with 4,096 of those stream processors, which is a huge jump from the 2,816 included in the GTX 980 Ti and the Radeon R9 290X, which was AMD’s previous flagship.

The new GPU is clocked to 1,050MHz, which is about 50MHz more than the GTX 980 Ti, and the chip is built from a 8.9 billion transistors--which, if you're counting, is almost one billion more than the Nvidia GPU.

"AMD hasn’t just turbo-charged its major components to build the Fury X, it has also deployed a new memory system."

AMD hasn’t just turbo-charged its major components to build the Fury X, it has also deployed a new memory system. The new chips are called High Bandwidth Memory and replace the more familiar GDDR5, and they change the fundamental structure of memory to improve performance: the chips are stacked vertically as well as horizontally, and they’re accessed using a far wider bus than any previous graphics memory system.

No Caption Provided

Those changes improve performance and increase efficiency so much that AMD only has to run the Fury X’s memory at 1,000MHz and it still manages to far exceed the GTX 980 Ti in bandwidth. AMD’s new card can deliver a theoretical throughput of 512GB/s, while the Nvidia card offers 336GB/s.

The structure of other components remains the same. A single Graphics Command Processor still controls the chip from above, and each Shader Engine retains its dedicated texture and rasteriser units. The 28nm manufacturing process also remains in place.

Innovation isn’t limited to the inside of the card. Traditional air-cooling has been ditched in favour of a water-cooling unit that’s similar to the pre-built devices available for processors. It’s a double-edged sword; the cooler is modestly sized and helps the card measure just 195mm, which is almost 100mm shorter than other top cards, but the cooler does require a 120mm mount inside of a chassis. That shouldn't be hard to find in most enclosures, but it's still something to bear in mind.

The water-cooling bodes well for keeping the chip chilled, and we don’t expect the Fury X to prove hotter than its rivals either; its top power requirement, of 275W, is competitive. To get the Fury X running you’ll need a PSU with two eight-pin connectors, which is a little more than the eight- and six-pin connectors needed with the GTX 980 Ti.

There’s one major area where the Fury X can’t compete, though, and that’s partner cards, as in, the modified versions of GPUs that often appear with overclocked cores, different cooling configurations and extra features. At the moment AMD isn’t opening the Fury X to board partner modification, which means you have to buy its reference design instead.

That makes it important to pay attention to the Fury’s ports. We’ve no qualms about its trio of DisplayPort connectors, but the HDMI 1.4 output omits a couple of features when compared to HDMI 2.0; There’s no 4K support at 60fps or 21:9 aspect ratio support using HDMI 1.4, for instance.

Undoubtedly, the Fury X is a good-looking card. The exterior is made from metal with illuminated logos, and a row of eight LEDs indicate GPU-load in either red or blue. The main plate can be removed with four screws, and AMD has appeased modders by making the design of that plate available for 3D modelling, so it can be replaced by custom designs.

Radeon R9 Fury X vs Titan X vs Others

GPU

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GeForce GTX Titan X

CUDA Cores

4,096

2,816

2,816

3,072

Base Clock

1,050MHz

1,000MHz

1,000MHz

1,000MHz

GPU Boost Clock

N/A

N/A

1,075MHz

1,088.5MHz

Memory

4GB

4GB/8GB

6GB

12GB

Memory Data Rate

1,000MHz

5,000MHz

7,010MHz

7,010MHz

Memory Bandwidth

512GB/s

320GB/s

336GB/s

336GB/s

Memory Interface

4,096-bit

512-bit

384-bit

384-bit

ROPs

64

64

96

96

TDP

275W

290W

250W

250W

Fabrication Process

28nm

28nm

28nm

28nm

Radeon R9 Fury X Performance (1440p)

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GTX Titan X

Heaven @ Ultra, 8X AA

42.7

33

51

52.1

3D Mark Fire Strike Extreme

7,276

4,973

7,405

7,598

Crysis 3 @ Very High

64.6

42.3

62.6

64.6

Tomb Raider, Ultimate, FXAA

131.1

N/A

126

N/A

Bioshock Infinite @ Ultra DOF, AA

97.5

62.4

103

100.5

Battlefield 4 @ Ultra, 4X MSAA, HBAO

63.8

46.1

73.9

74.3

Batman: Arkham Origins @ Very High

125

92

149

168

Metro: Last Light @ Very High

85.5

52

77

73.1

Shadow of Mordor @ Ultra

83.7

N/A

85.2

N/A

Radeon R9 Fury X Performance (4K)

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GTX Titan X

Heaven @ Ultra, 8X AA

20.1

14.7

21.9

23.1

3D Mark Fire Strike Ultra

3,943

2,658

3,888

3,959

Crysis 3 @ Very High

31.9

22.1

29.8

31.1

Tomb Raider, Ultimate, FXAA

62.7

N/A

59

N/A

Bioshock Infinite @ Ultra DOF, AA

52.5

29.8

54.7

54.7

Battlefield 4 @ Ultra, 4X MSAA, HBAO

34

23.9

37.2

38.4

Batman: Arkham Origins @ Very High

74

44

82

81

Metro: Last Light @ Very High

45

30

42

39.9

Shadow of Mordor @ Ultra

48

N/A

47.3

N/A

Editor’s note: The GTX 980 Ti and R9 290X benchmarks are based on previous tests which did not account for Tomb Raider or Shadow of Mordor.

There’s no mistaking the fact that the Fury X is one of the fastest cards on the market right now--our 1080p benchmarks aren’t displayed on these graphs, but that’s because the results are elementary. If you’re playing games at 1,920 x 1,080 the Fury X is overkill.

At 1440p, the situation isn’t as clear-cut. The Fury X can easily handle any game at this higher resolution, but it faces competition from the GTX 980 Ti. The AMD card proved faster in Crysis 3, Tomb Raider and Metro: Last Light, but it was never far ahead--its averages of 64fps, 131fps and 85.5fps were only a handful of frames beyond the Nvidia card.

However, the GTX 980 Ti was meaningfully faster in our four other test games--in some cases, it was ten frames or more faster than the AMD card. The GTX 980 Ti’s 1440p victory was further emphasised by our theoretical tests, in which the Nvidia card proved quicker than the Fury X in both Unigine Heaven and 3D Mark Fire Strike.

"The Fury X proved faster than the GTX 980 Ti in Crysis 3, Tomb Raider and Metro: Last Light, but it was never far ahead... the 980 Ti was meaningfully faster in our four other test games."

The battle was similarly close at 4K but, crucially, the Fury X fought back. Its 3D Mark result of 3,943 is better than the GTX 980 Ti could manage, and it was a handful of frames quicker in Crysis 3, Tomb Raider, Metro and Shadow of Mordor. It’s a close-run thing, with its best victory coming with three-frame leads in two games.

Nvidia’s card took the crown in Bioshock Infinite, Batman Arkham Origins and Battlefield 4, but only by similarly slim margins. Its best result came from Batman, where it was six frames better than the Fury X.

The Fury X fell behind at 1440p and just about took an overall victory at 4K, but neither card proved dominant, and the Fury X’s improved 4K performance is undermined elsewhere. Nvidia’s card is a few frames back in some of our 4K tests, but the GTX 980 Ti is available in dozens of overclocked configurations, and most of those will help the green team’s hardware catch up. That’s something not available with the Fury X, which is only manufacturerd as a reference design.

Meanwhile, The Fury X performed well in thermal tests. The Fury’s water-cooling unit saw the core hit a maximum temperature of 65°C, which is far cooler than the 82°C GTX 980 Ti, but the Fury X wasn’t as frugal as its rival: our AMD-powered rig drew 369W from the mains, but the GTX 980 Ti machine needed just 330W.

Click on the thumbnails below to view in full-screen
Click on the thumbnails below to view in full-screen

Verdict

The revisions made to AMD’s Graphics Core Next architecture means the Fury X delivers a huge leap over its predecessor, and it’s got plenty of innovation on board – the water-cooling unit is impressively effective, and the revised memory system delivers incredible bandwidth with a smaller, slower amount of RAM.

AMD's latest chip is consistently quick, but it can’t quite overhaul the GTX 980 Ti. It just about loses out at 1440p, and at 4K its slight lead will be wiped out by many of the overclocked Nvidia cards that have entered the market. While AMD’s latest card is an undoubted 4K contender, especially if the card’s smaller size appeals, it’s only able to draw level with Nvidia rather than take a clear lead.

Got a news tip or want to contact us directly? Email news@gamespot.com


Back To Top
191 Comments  RefreshSorted By 
GameSpot has a zero tolerance policy when it comes to toxic conduct in comments. Any abusive, racist, sexist, threatening, bullying, vulgar, and otherwise objectionable behavior will result in moderation and/or account termination. Please keep your discussion civil.

Avatar image for davillain
DaVillain

59009

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

DaVillain  Moderator

After watching JaysTwoCents video R9 390, I'll point out that the card isn't all that bad but that thing can barley go 4K but that doesn't mean it's a bad card if your in the market for something decent for a low low. As for AMD R9 Fury X, I'm not worry about competing Nvidia nor doing 4K. Nice review GS.

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@davillain-: I'm really considering on sending my 970 back and getting the 390x or Fury pro. Nvidia driver team as been lazy as shit the last couple months, broken driver after broken driver. I had to roll my drivers back to the April GTA 5 ones, just so i don't BSOD as much.

Upvote • 
Avatar image for mordeaniis
Mordeaniis

161

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

@rolento25: No problems here, most of the issues with GTA seemed to be game related for me (ie, their updates fixed more issues than driver updates did), and BSOD is rarely if every a driver caused issue. RMA might be in order.

Upvote • 
Avatar image for gord0nd
Gord0nD

262

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

@rolento25 said:

@davillain-: I'm really considering on sending my 970 back and getting the 390x or Fury pro. Nvidia driver team as been lazy as shit the last couple months, broken driver after broken driver. I had to roll my drivers back to the April GTA 5 ones, just so i don't BSOD as much.

I would really try to figure out the BSOD issues. That is not a driver issue. If the drivers do make a difference it is happening, then its just triggering a problem that is already there. Im not critiquing, or defending your want to change cards, I just don't think your problem with go away if you do.

Upvote • 
Avatar image for Zironn1
Zironn1

110

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

4k

Upvote • 
Avatar image for robertoenrique
robertoenrique

1191

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

I bet it's the shitty AMD drivers that hold this GPU back compared to the 980ti and the Titan X.

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@robertoenrique: Shitty AMD drivers? Do you have one of the 900 series Nvidia cards? Nvidia drivers have been the worst i've seen from both companys. Nvidia drivers with the Blue screen of death.

Upvote • 
Avatar image for Phelaidar
Phelaidar

1533

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@rolento25: AMD problem is definitely software related. It's not the first time they release a beast of a GPU and it seems to run worse than it should. FuryX should be competing with the Titan.

I Don't like their support. Most games run worse on amd months after launch and their driver release window is everywhere.

Upvote • 
Avatar image for robertoenrique
robertoenrique

1191

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@Phelaidar: THIS

Upvote • 
Avatar image for TalSet_11
TalSet_11

135

Forum Posts

0

Wiki Points

0

Followers

Reviews: 66

User Lists: 0

@rolento25: I switched from AMD to Nvidia three years ago precisely because of bsod issues. First had a 6790, then a 7950, then switched. Had a 670 and now a 980. I haven't had a bsod on either of my gaming rigs since then. AMD cards are good, and i certainly agree with what others are saying about the new Nvidia drivers being disappointing (especially in any game using GameWorks), but the power per watt on the 900-series cards is damn impressive.

Upvote • 
Avatar image for leviathanwing
leviathanwing

6384

Forum Posts

0

Wiki Points

0

Followers

Reviews: 34

User Lists: 0

@rolento25: historically... but this is the here and now and right now amd has some seriously crappy drivers. look above and see or go spec check amd vs nvidia cards. similar specs shouldnt have such disparity between them.

Upvote • 
Avatar image for TalSet_11
TalSet_11

135

Forum Posts

0

Wiki Points

0

Followers

Reviews: 66

User Lists: 0

@leviathanwing: I think that their driver game started to fall apart during their Mantle API experiment, which was a complete disaster

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@TalSet_11: How was Mantle a disaster? It did was AMD wanted... to push MS into making DX12

Upvote • 
Avatar image for mordeaniis
Mordeaniis

161

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

Edited By Mordeaniis

@rolento25 said:

@TalSet_11: How was Mantle a disaster? It did was AMD wanted... to push MS into making DX12

No it didn't, DX12 wasn't pushed around by AMD.

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@leviathanwing: The here and now? do you have a AMD card? i just came from one and haven't had driver issues in the last few years. The here and now is the poor Nvidia 900 series drivers.

Upvote • 
Avatar image for leviathanwing
leviathanwing

6384

Forum Posts

0

Wiki Points

0

Followers

Reviews: 34

User Lists: 0

@rolento25: i roll with 290xs... and i know that two 970s will outdo mine... seeing it happen in the household in fact. so please take your weird weird fanboy attitude and shove on.

Upvote • 
Avatar image for 40mdf0
40mdf0

1218

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

Edited By 40mdf0

This and its lesser sibling the Fury really need price cuts to be competitive. I'd pick a Fury up if it was $50-$100 less.

My only real gripe is the lack of HDMI 2.0, as I connect all of my equipment to an AV receiver and thus prefer that standard.

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Oh its AMD yawn

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@bigruss51: Enjoying the Nvidia 900 series BSOD drivers?

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By bigruss51

@rolento25: working just fine

Upvote • 
Avatar image for TalSet_11
TalSet_11

135

Forum Posts

0

Wiki Points

0

Followers

Reviews: 66

User Lists: 0

Edited By TalSet_11

@rolento25: I've been on a 980 for almost six months. No bsod's here. What are you referring to?

Upvote • 
Avatar image for rolento25
Rolento25

2833

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By Rolento25

@TalSet_11: Go read the forums if you want more info.

Upvote • 
Avatar image for 3cupsofflower
3CupsofFlower

1227

Forum Posts

0

Wiki Points

0

Followers

Reviews: 6

User Lists: 5

@bigruss51: WTF? Even graphics card companies have fanboys.

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@3cupsofflower: Believe it. I got those awesome NVidia and GeForce stickers stuck on everything.

Upvote • 
Avatar image for deactivated-642321fb121ca
deactivated-642321fb121ca

7142

Forum Posts

0

Wiki Points

0

Followers

Reviews: 20

User Lists: 0

@bigruss51: Nvidia fanboy, yawn.

Upvote • 
Avatar image for Acillatem1993
Acillatem1993

1103

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@Random_Matt: I love how nvidia fanboys are just hating for no reason lol

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@Acillatem1993: I don't hate. Just not interested in anything AMD comes out with. Majority of games run better with NVidia cards. Fact.

Upvote • 
Avatar image for Acillatem1993
Acillatem1993

1103

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@bigruss51: You're not interested, yet you bothered to express your disliking. That right there is hate, even if you dont realize it.

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By bigruss51

@Acillatem1993: So you're a psychologist now. I am not interested in AMD cards but I am interested in leaving a comment, dumbass. Now there's hate. See the difference?

Upvote • 
Avatar image for Acillatem1993
Acillatem1993

1103

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@bigruss51: Now you're even insulting. Let the hate flow.

Upvote • 
Avatar image for bigruss51
bigruss51

207

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@Acillatem1993: *$%*^&^&*)*%^$!*_-+|^%... !)-{?&$^

Upvote • 
Avatar image for el_duderino8
el_duderino8

89

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@Random_Matt: Man I miss that 'Like' button ;)

Upvote • 
Avatar image for deactivated-642321fb121ca
deactivated-642321fb121ca

7142

Forum Posts

0

Wiki Points

0

Followers

Reviews: 20

User Lists: 0

Looking forward to the Fury Pro.

Upvote • 
Avatar image for oflow
oflow

5185

Forum Posts

0

Wiki Points

0

Followers

Reviews: 40

User Lists: 0

Now if there were only some pc games worth playing with this beast...

Upvote • 
Avatar image for TalSet_11
TalSet_11

135

Forum Posts

0

Wiki Points

0

Followers

Reviews: 66

User Lists: 0

@oflow: I can't even wrap my head around this bizarre statement.

Upvote • 
Avatar image for leviathanwing
leviathanwing

6384

Forum Posts

0

Wiki Points

0

Followers

Reviews: 34

User Lists: 0

@oflow: pc games? plenty... necessitating this? not many... in fact unless youre 4k gaming not any. but divinity, pillars, wasteland 2.... e.t.c plenty fine good games to be had.

Upvote • 
Avatar image for oflow
oflow

5185

Forum Posts

0

Wiki Points

0

Followers

Reviews: 40

User Lists: 0

Edited By oflow

@leviathanwing: that was kinda my point. i guess I worded it wrong.

Both Divinity and Wasteland 2 are coming to console so I don't even need a pc to play them. But I stand corrected none the less.

Upvote • 
Avatar image for TalSet_11
TalSet_11

135

Forum Posts

0

Wiki Points

0

Followers

Reviews: 66

User Lists: 0

@oflow: I've played a ton of Divinity and Elite: Dangerous on PC. I have no idea how those experiences are supposed to carry over to a controller, especially E:D. A controller can't do even a third of the inputs that I've mapped to my HOTAS. But maybe they've figured it out. Time will tell.

Upvote • 
Avatar image for oflow
oflow

5185

Forum Posts

0

Wiki Points

0

Followers

Reviews: 40

User Lists: 0

Edited By oflow

@TalSet_11: I actually played E:D on pc and I'm now playing it Xbox and I personally actually like it better on console than pc. (I'm not a flight stick person I played the PC version with an Xbox controller as well)

I like the console versions of Diablo 3 and Elder Scrolls Online better than the pc versions as well, especially D3 I think it's actually far better on console than pc so I don't really see an issue with these games translating over.

Upvote • 
Avatar image for leviathanwing
leviathanwing

6384

Forum Posts

0

Wiki Points

0

Followers

Reviews: 34

User Lists: 0

@oflow: i know ill be a second time buyer... because good games and good devs deserve it. got to put those steam sale/psn flash sale savings to good use anyhow.

Upvote •