Feature Article

AMD Radeon R9 Fury X Review

GameSpot may receive revenue from affiliate and advertising partnerships for sharing this content and from purchases through links.

A new 4K flagship, locked and loaded.

To some people, AMD has fallen behind Nvidia because of its lack of response to cards like the GTX 980 Ti and GTX Titan X, but that stops now, with the red team unleashing its newest flagship GPU: the Radeon R9 Fury X.

It's an interesting chip that impresses in several key departments. AMD's familiar architecture has been upgraded in every important area, and the corporation has ditched air-cooling for a separate water-cooling module that it promises can keep the beefy chip chilled.

This amount of hardware doesn't come cheap, however, which means you'll have to shell out $650 for the Fury X. That's similar cash to the Nvidia GeForce GTX 980 Ti, which is the Fury’s main 4K rival.

No Caption Provided

Radeon Fury X Specs

AMD’s new GPU is called Fiji, and it turbo-charges the existing Graphics Core Next architecture to help the Fury compete with the 980 Ti.

The core is constructed from four Shader Engines, with each divided into 16 Compute Units--and each of those is rammed with 64 stream processors. That means the Fury X is built with 4,096 of those stream processors, which is a huge jump from the 2,816 included in the GTX 980 Ti and the Radeon R9 290X, which was AMD’s previous flagship.

The new GPU is clocked to 1,050MHz, which is about 50MHz more than the GTX 980 Ti, and the chip is built from a 8.9 billion transistors--which, if you're counting, is almost one billion more than the Nvidia GPU.

"AMD hasn’t just turbo-charged its major components to build the Fury X, it has also deployed a new memory system."

AMD hasn’t just turbo-charged its major components to build the Fury X, it has also deployed a new memory system. The new chips are called High Bandwidth Memory and replace the more familiar GDDR5, and they change the fundamental structure of memory to improve performance: the chips are stacked vertically as well as horizontally, and they’re accessed using a far wider bus than any previous graphics memory system.

No Caption Provided

Those changes improve performance and increase efficiency so much that AMD only has to run the Fury X’s memory at 1,000MHz and it still manages to far exceed the GTX 980 Ti in bandwidth. AMD’s new card can deliver a theoretical throughput of 512GB/s, while the Nvidia card offers 336GB/s.

The structure of other components remains the same. A single Graphics Command Processor still controls the chip from above, and each Shader Engine retains its dedicated texture and rasteriser units. The 28nm manufacturing process also remains in place.

Innovation isn’t limited to the inside of the card. Traditional air-cooling has been ditched in favour of a water-cooling unit that’s similar to the pre-built devices available for processors. It’s a double-edged sword; the cooler is modestly sized and helps the card measure just 195mm, which is almost 100mm shorter than other top cards, but the cooler does require a 120mm mount inside of a chassis. That shouldn't be hard to find in most enclosures, but it's still something to bear in mind.

The water-cooling bodes well for keeping the chip chilled, and we don’t expect the Fury X to prove hotter than its rivals either; its top power requirement, of 275W, is competitive. To get the Fury X running you’ll need a PSU with two eight-pin connectors, which is a little more than the eight- and six-pin connectors needed with the GTX 980 Ti.

There’s one major area where the Fury X can’t compete, though, and that’s partner cards, as in, the modified versions of GPUs that often appear with overclocked cores, different cooling configurations and extra features. At the moment AMD isn’t opening the Fury X to board partner modification, which means you have to buy its reference design instead.

That makes it important to pay attention to the Fury’s ports. We’ve no qualms about its trio of DisplayPort connectors, but the HDMI 1.4 output omits a couple of features when compared to HDMI 2.0; There’s no 4K support at 60fps or 21:9 aspect ratio support using HDMI 1.4, for instance.

Undoubtedly, the Fury X is a good-looking card. The exterior is made from metal with illuminated logos, and a row of eight LEDs indicate GPU-load in either red or blue. The main plate can be removed with four screws, and AMD has appeased modders by making the design of that plate available for 3D modelling, so it can be replaced by custom designs.

Radeon R9 Fury X vs Titan X vs Others

GPU

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GeForce GTX Titan X

CUDA Cores

4,096

2,816

2,816

3,072

Base Clock

1,050MHz

1,000MHz

1,000MHz

1,000MHz

GPU Boost Clock

N/A

N/A

1,075MHz

1,088.5MHz

Memory

4GB

4GB/8GB

6GB

12GB

Memory Data Rate

1,000MHz

5,000MHz

7,010MHz

7,010MHz

Memory Bandwidth

512GB/s

320GB/s

336GB/s

336GB/s

Memory Interface

4,096-bit

512-bit

384-bit

384-bit

ROPs

64

64

96

96

TDP

275W

290W

250W

250W

Fabrication Process

28nm

28nm

28nm

28nm

Radeon R9 Fury X Performance (1440p)

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GTX Titan X

Heaven @ Ultra, 8X AA

42.7

33

51

52.1

3D Mark Fire Strike Extreme

7,276

4,973

7,405

7,598

Crysis 3 @ Very High

64.6

42.3

62.6

64.6

Tomb Raider, Ultimate, FXAA

131.1

N/A

126

N/A

Bioshock Infinite @ Ultra DOF, AA

97.5

62.4

103

100.5

Battlefield 4 @ Ultra, 4X MSAA, HBAO

63.8

46.1

73.9

74.3

Batman: Arkham Origins @ Very High

125

92

149

168

Metro: Last Light @ Very High

85.5

52

77

73.1

Shadow of Mordor @ Ultra

83.7

N/A

85.2

N/A

Radeon R9 Fury X Performance (4K)

Radeon R9 Fury X

Radeon R9 290X

GTX 980 Ti

GTX Titan X

Heaven @ Ultra, 8X AA

20.1

14.7

21.9

23.1

3D Mark Fire Strike Ultra

3,943

2,658

3,888

3,959

Crysis 3 @ Very High

31.9

22.1

29.8

31.1

Tomb Raider, Ultimate, FXAA

62.7

N/A

59

N/A

Bioshock Infinite @ Ultra DOF, AA

52.5

29.8

54.7

54.7

Battlefield 4 @ Ultra, 4X MSAA, HBAO

34

23.9

37.2

38.4

Batman: Arkham Origins @ Very High

74

44

82

81

Metro: Last Light @ Very High

45

30

42

39.9

Shadow of Mordor @ Ultra

48

N/A

47.3

N/A

Editor’s note: The GTX 980 Ti and R9 290X benchmarks are based on previous tests which did not account for Tomb Raider or Shadow of Mordor.

There’s no mistaking the fact that the Fury X is one of the fastest cards on the market right now--our 1080p benchmarks aren’t displayed on these graphs, but that’s because the results are elementary. If you’re playing games at 1,920 x 1,080 the Fury X is overkill.

At 1440p, the situation isn’t as clear-cut. The Fury X can easily handle any game at this higher resolution, but it faces competition from the GTX 980 Ti. The AMD card proved faster in Crysis 3, Tomb Raider and Metro: Last Light, but it was never far ahead--its averages of 64fps, 131fps and 85.5fps were only a handful of frames beyond the Nvidia card.

However, the GTX 980 Ti was meaningfully faster in our four other test games--in some cases, it was ten frames or more faster than the AMD card. The GTX 980 Ti’s 1440p victory was further emphasised by our theoretical tests, in which the Nvidia card proved quicker than the Fury X in both Unigine Heaven and 3D Mark Fire Strike.

"The Fury X proved faster than the GTX 980 Ti in Crysis 3, Tomb Raider and Metro: Last Light, but it was never far ahead... the 980 Ti was meaningfully faster in our four other test games."

The battle was similarly close at 4K but, crucially, the Fury X fought back. Its 3D Mark result of 3,943 is better than the GTX 980 Ti could manage, and it was a handful of frames quicker in Crysis 3, Tomb Raider, Metro and Shadow of Mordor. It’s a close-run thing, with its best victory coming with three-frame leads in two games.

Nvidia’s card took the crown in Bioshock Infinite, Batman Arkham Origins and Battlefield 4, but only by similarly slim margins. Its best result came from Batman, where it was six frames better than the Fury X.

The Fury X fell behind at 1440p and just about took an overall victory at 4K, but neither card proved dominant, and the Fury X’s improved 4K performance is undermined elsewhere. Nvidia’s card is a few frames back in some of our 4K tests, but the GTX 980 Ti is available in dozens of overclocked configurations, and most of those will help the green team’s hardware catch up. That’s something not available with the Fury X, which is only manufacturerd as a reference design.

Meanwhile, The Fury X performed well in thermal tests. The Fury’s water-cooling unit saw the core hit a maximum temperature of 65°C, which is far cooler than the 82°C GTX 980 Ti, but the Fury X wasn’t as frugal as its rival: our AMD-powered rig drew 369W from the mains, but the GTX 980 Ti machine needed just 330W.

Click on the thumbnails below to view in full-screen
Click on the thumbnails below to view in full-screen

Verdict

The revisions made to AMD’s Graphics Core Next architecture means the Fury X delivers a huge leap over its predecessor, and it’s got plenty of innovation on board – the water-cooling unit is impressively effective, and the revised memory system delivers incredible bandwidth with a smaller, slower amount of RAM.

AMD's latest chip is consistently quick, but it can’t quite overhaul the GTX 980 Ti. It just about loses out at 1440p, and at 4K its slight lead will be wiped out by many of the overclocked Nvidia cards that have entered the market. While AMD’s latest card is an undoubted 4K contender, especially if the card’s smaller size appeals, it’s only able to draw level with Nvidia rather than take a clear lead.

Got a news tip or want to contact us directly? Email news@gamespot.com


Back To Top
191 Comments  RefreshSorted By 
GameSpot has a zero tolerance policy when it comes to toxic conduct in comments. Any abusive, racist, sexist, threatening, bullying, vulgar, and otherwise objectionable behavior will result in moderation and/or account termination. Please keep your discussion civil.

Avatar image for ECH71
ECH71

933

Forum Posts

0

Wiki Points

0

Followers

Reviews: 6

User Lists: 0

Only 4GB of memory after releasing 8GB 290X? What is happening?!

*pulls out his hair with both hands*

Upvote • 
Avatar image for Coseniath
Coseniath

3183

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@ECH71: Since they choose to use HBM so early, they could only use HBM v1.0.

HBM1 is limited to 4GB. So actually they are forced to use 4GB only.

Upvote • 
Avatar image for KohPhiPhi2
KohPhiPhi2

37

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

So, basically, they've caught up with Ndivia in terms of performance (not surpassing them though), while they have managed to keep temperatures to significantly lower levels. All that at about the price range as the Ndivia counterparts.

Seems like a good deal to me. Glad to see AMD back in the game.

Upvote • 
Avatar image for Talavaj
Talavaj

265

Forum Posts

0

Wiki Points

0

Followers

Reviews: 42

User Lists: 0

Edited By Talavaj

Disappointing, AMD usually beats Nvidia in value but due to the recent price cuts on Nvidia high end gpus who would bother with this over 980 ti ?

Completely blew it wit this one.

Upvote • 
Avatar image for Hector_01
Hector_01

227

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

@Talavaj: price cuts??????????? In Australia a 980ti is 1100 bucks. That is just a ripoff. The fury x is currently 999 in aussie so it is actually cheaper. I wouldn't buy either just a waste of money.

Upvote • 
Avatar image for naryanrobinson
naryanrobinson

1272

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

For those saying that AMD makes products that run like garbage, too hot and too loud,

this card is just about on par with the competing NVIDIA card in performance and thermals, and beats it hands down in quietness.

Upvote • 
Avatar image for deactivated-671700589c1c6
deactivated-671700589c1c6

98

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

@naryanrobinson: Also forgetting that the Fury X is watercooled whilst the 980ti runs on air... That's a pretty considerable difference in my opinion.

Upvote • 
Avatar image for OmegaGear
OmegaGear

737

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

Edited By OmegaGear

@naryanrobinson: AMD and nVidia have both had their share of loud and power hungry cards. It just so happens that the last run of AMD cards stuck out. Nvidia had the same issue with the 480s. Those things ran hotter at idle than modern cards run at load.

Upvote • 
Avatar image for kalpesh_78
kalpesh_78

52

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

@naryanrobinson: I disagree. Fury is marginally quieter than the 980 Ti. Also... most of the cards are far more silent when they are new. Check back on the card noise a year later.

Upvote • 
Avatar image for mulder_000
mulder_000

200

Forum Posts

0

Wiki Points

0

Followers

Reviews: 6

User Lists: 0

I have a silly question since I haven't been following the GPU scene for a couple of years. I was just wondering what is the best card for 1920 X 1080 (1 monitor) gaming? I'm not interested in 4K. Nor Sli or multi-monitor gaming.

Is a GTX 970 4GB a good enough card to last 2-3 years? I'm still using an AMD 6870 1 GB card and it still plays almost every game on High settings. But, I'd like to do a bit of catch-up. Thanks.

Upvote • 
Avatar image for Reuwsaat
Reuwsaat

1088

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

@mulder_000: How to choose a video card? Simple answer that will always work:

First, what do you need a gaming video card for? For gaming performance. And what is that, to put it as simple as possible? FPS. Hit YouTube and search for the video card name + resolution + benchmark. Yep, there you go, you'll see for yourself, unchanged and unpaid, just how does that card runs your content. No need for website reviews and you get to see as close as possible to what you will get when you pick it or not, and as bonus you can also see the rig the guy is running too, so you might want to check on the rest of the hardware too.

So? Do you really need to care about how much memory does it have? Or how high is the clock? No, because you don't know how the technlogy of the card employs its resources, specially now with this new memory from AMD, it doesn't use the memory as it used to, so it doesn't matter if it's "just 4Gb", it'll work just as good because of the way it access its memory. The only thing you might want to check out is the DirectX version, always opt for the latest.

Why isn't the answer more technical? Technical jargon go away with experience, see for yourself if I'm bsing you or not, it will always work, guaranteed.

Upvote • 
Avatar image for deactivated-5b116b1818dbb
deactivated-5b116b1818dbb

34

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@mulder_000: Be careful gtx 970 has hardware flaw where a sever performance hit can occur especially in the future where more memory intensive games will be used. I seriously don't recommend 970 if your not planning to replace the card soon.

I honestly recommend R9 380(4 gig version) who don't want to spend much on a graphics card, it has a very close architecture to the fiji core which is in the FuryX and are much more efficient then the r9 200 series.

Upvote • 
Avatar image for kalpesh_78
kalpesh_78

52

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

@Lehikex: Yup. I agree on that too. 980 Ti is much better though. Future proof for at least a couple of years.

Upvote • 
Avatar image for zatorys
zatorys

638

Forum Posts

0

Wiki Points

0

Followers

Reviews: 26

User Lists: 0

@Lehikex said:

@mulder_000: Be careful gtx 970 has hardware flaw where a sever performance hit can occur especially in the future where more memory intensive games will be used. I seriously don't recommend 970 if your not planning to replace the card soon.

I honestly recommend R9 380(4 gig version) who don't want to spend much on a graphics card, it has a very close architecture to the fiji core which is in the FuryX and are much more efficient then the r9 200 series.

i agree with him

Upvote • 
Avatar image for ank000
ank000

57

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@mulder_000: Nvidia GTX 970 will certainly last you 2-3 years if you plan to play at FHD. You can also look at r9 390(non-X) 8GB which is priced similarly to 970 and is slightly more powerful but uses significantly more power.

P.S : GTX 970 is the best value higher mid range card under $350 which is considered a 1440p gaming card, so it will definitely last you for long time if you only plan on using it with FHD monitor.

Upvote • 
Avatar image for 1blackone
1blackone

109

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

Edited By 1blackone

@ank000: BUT since he mentioned he had an AMD previously, im gonna recommend an R9 290x as it will probably be more compatible with his rig. As you can see on the chart, it will play anything that is currently out and yet to be released at 1080p at +60 or close to it, It would supposedly be more compatible with your MOBO which will allow some additional tuning, plus the MSI R9 for 299$ which will suit you just fine for 1080p gaming for the next few years

THE MAIN THING to remember is that future games like Star Citizen really wont attack your CPU, and even your GPU: itll go for your RAM more than anything else. Make sure you get a card with all the features u like at around the 300$, get 16 GB of really good ram, and you should be set for a long while.

Upvote • 
Avatar image for Reuwsaat
Reuwsaat

1088

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

@1blackone said:

@ank000: BUT since he mentioned he had an AMD previously, im gonna recommend an R9 290x as it will probably be more compatible with his rig. As you can see on the chart, it will play anything that is currently out and yet to be released at 1080p at +60 or close to it, It would supposedly be more compatible with your MOBO which will allow some additional tuning, plus the MSI R9 for 299$ which will suit you just fine for 1080p gaming for the next few years

THE MAIN THING to remember is that future games like Star Citizen really wont attack your CPU, and even your GPU: itll go for your RAM more than anything else. Make sure you get a card with all the features u like at around the 300$, get 16 GB of really good ram, and you should be set for a long while.

Define how any VGA can be more "compatible with his rig"? You could get AMD CPU, AMD RAM, AMD SSD and still get GeForce, it has nothing to do with anything, it's just a PCI-E x16 slot.

Upvote • 
Avatar image for 1blackone
1blackone

109

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

@Reuwsaat: Sure thing, I'll "define" (you mean explain?)

Having an AMD GPU & CPU on a supported mobo has the potential to give a few sizable benefits. That's what the "will allow some additional tuning" hinted at.

One way is that having all AMD could give a one access to APIs like Mantle(now Vulkan) that could drastically boost a rig performance in running a game or program. Another way is when it comes to Overclocking. This way, you don't have to have both Nvidia control panel (or evga/whatever OC program) possibly conflicting with CCC. everything can all be tweaked in one program "under one roof", which also cuts down on the amounts of process your OS is running.

I hope I explained the benefits of having all Nvidia or all AMD in a PC...

Upvote • 
Avatar image for Reuwsaat
Reuwsaat

1088

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

Edited By Reuwsaat

@1blackone: Mantle/Vulkan support is negligible on the market today, and about overclocking, the % of people who actually do that is even more negligible against the number of those who don't, specially when we are talking about gpu overclocking; I honestly don't think he would be overclocking either, considering his question. Also, he did not mention at any time that his CPU was AMD, at least not at this thread, so you're assuming it.

Upvote • 
Avatar image for 1blackone
1blackone

109

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

Edited By 1blackone

@Reuwsaat: Yes, I did assume it. That's what the "probably" and "supposedly" was meant to infer. Those words inherently mean that you are assuming something...

API support is small TODAY, but we don't build PCs or 24 hour periods, its about the long run. 3-4 years from know that could change, so one would be wise to keep that option as viable as possible. Same for OC'ing: who knows what he might want to do in the future, and OC'ing is getting easier to get into all the time. just a simple bios tweak here, a program there, an online tutorial and an afternoon and bam, you're overclocking. Again, its about keeping options OPEN, and personally, i think having multiple brands in a rig potentially limit options, in the long run.

If you'd be so kind, would you just conclude this thread with a "Yeah, well, that's just, like, your opinion, man." post so we can both move on?

Upvote • 
Avatar image for ank000
ank000

57

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

Edited By ank000

@1blackone: I didn't recommend 290X only because even though its competent today an will be for sometimes to come its still a 2 year old GPU. Also its power consumption and temp under load is way more than that of GTX 970.

Another reason for me to not recommending a R9 290x is because it runs really hot, increasing the temp inside the cabinet which in return might reduce the lifespan of other parts due to heat unless extra fan is installed. So I thought its better to spend $30 more on GPU.

Also I don't think GPU has anything to do with other parts from other brands. Be it AMD or Nvidia it doesn't affect in any significant way.

Upvote • 
Avatar image for mulder_000
mulder_000

200

Forum Posts

0

Wiki Points

0

Followers

Reviews: 6

User Lists: 0

@1blackone: Thanks for the advice. My MOBO, unfortunately, only can handle a max of 12 GB.

The MSI R290X go for over $350 here in Canada. (Damn Canadian $) I will have to, at some point, upgrade my MOBO and CPU. I have 12GB memory now and as I said I am running games quite fine for the time being.

Upvote • 
Avatar image for lordossiss
LordOssiss

399

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

I figure this is a decent enough place to ask:
What kind of graphics card should I/ could I get with a small form factor PC?

Upvote • 
Avatar image for Alurit
Alurit

1002

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By Alurit

@lordossiss: depends on how small you want to go.

if you want to go with a miniITX that can fit a full length card and an AIO liquid cooler, then a gtx 980ti with an AIO on it will give you the best performance.

if you cannot fit both but can fit a fury x then the fury x will give you the most performance, even if you can fit an air cooled gtx 980ti, since it cannot boost high in a small form factor case (linustechtips made a video about it on youtube), even a reference cooler that blows out of the air instead of inside like an aftermarket one.

if you can only fit a small card then the best itx sized card is the gtx 970 mini from gigabyte or asus.

if you cannot afford a 970 then a gtx 960 or a r9 285, there's mini versions of both, they are similar in performance, with 285 might having a small edge, but using more power thus exhausting a bit more hot air into the case, the the difference is small between the two.

if you really have to go budget then you're left with the 750ti. that card being about 20-30% faster then a ps4 on average, with some unoptimised games they are being similar. that little beast uses so less power and thus running so cool that there is even air cooled versions but you can't use that in a small case, and it has worse performance then an actively cooled one.

Upvote • 
Avatar image for ank000
ank000

57

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@lordossiss: Wait for AMD R9 Nano. It is said to be around 6 inch and perfect for mini- ITX build. Alternatively you can opt for AMD Fury (non-X) which will be available soon .

Also, Nvidia GTX 970 from gigabyte for mini-ITX is also an option.

Upvote • 
Avatar image for dave70292
dave70292

92

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

The R9 Fury Non X is better deal for $100 cheaper

Upvote • 
Avatar image for streamline
streamline

2258

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Since this is water cooled, noise may be lower: It would be good to compare how loud it is maxed out to its competitors. Also, at $650, how well does it do against two cards at about $325?

Upvote • 
Avatar image for mordeaniis
Mordeaniis

161

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 5

Edited By Mordeaniis

Nothing "Stops now." It's comparable to the competition, it isn't replacing it as the new king. In fact it lacks some pretty important things: HDMI 2.0 is a big sticking point for a freakin' 4K card. nVidia has that in cards that aren't even super amazing at 4k and has for a while, there's no reason for this card not to have it. Hope you don't want to run a 4K TV with this.

Upvote • 
Avatar image for deactivated-60b838d2a137f
deactivated-60b838d2a137f

2184

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 5

@mordeaniis: You could just use DisplayPort though, no?

Upvote • 
Avatar image for jhcho2
jhcho2

5103

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 2

Might as well just get a 980 Ti

Upvote • 
Avatar image for Suikogaiden
Suikogaiden

1472

Forum Posts

0

Wiki Points

0

Followers

Reviews: 3

User Lists: 0

Edited By Suikogaiden

No games even play in a native 4k resolution so just wasting your money. Morons will buy this anyway though but who knows in 10 years when 4k is actually in use it might be worth buying.

Upvote • 
Avatar image for Kunakai
Kunakai

912

Forum Posts

0

Wiki Points

0

Followers

Reviews: 4

User Lists: 0

Edited By Kunakai

@Suikogaiden: Nonsense.

Upvote • 
Avatar image for motopramaus
motopramaus

212

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 5

Edited By motopramaus

@Suikogaiden: What does that mean ? I play games at 4K on my PC, you think it's upscaling ? This isn't a console, this is rendering 4K resolution on the GPU.

Upvote • 
Avatar image for 1blackone
1blackone

109

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

@motopramaus: Hmm...cutting through the hyperbole, maybe he means a new-ish 3d high-poly game with native 4k textures that is pegged at 60+ plus FPS? If he means that, then yes he's right, that level of tech wont hit "mid-level" money PC ranges for a several years--and remember: other than CD Projekt, devs make games for the mid-level rigs, not the uber 3x SLI'd monster rigs.

So yes, maybe not 10 years, but still far off enough to consider letting a few Black Fridays to come by before you purchase anything (monitor, card, whatev) saying its "4k!"

Upvote • 
Avatar image for Kunakai
Kunakai

912

Forum Posts

0

Wiki Points

0

Followers

Reviews: 4

User Lists: 0

@1blackone: If that's what he's talking about then he shouldn't be referring to it as the native resolution. As far as I'm aware most recent AAA games output at 4k, most 4k monitors display at 4k.

Upvote • 
Avatar image for DanielL5583
DanielL5583

1221

Forum Posts

0

Wiki Points

0

Followers

Reviews: 166

User Lists: 0

Rectify the CUDA cores count and replace it with the stream processors.

AMD does not use CUDA processing; that's an NVIDIA technology.

Upvote • 
Avatar image for ratchet500000
ratchet500000

50

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By ratchet500000

@DanielL5583: The article also compares core frequency (MHz) across two GPUs of different architecture, which is inaccurate on so many levels. By their logic, a single core CPU from a decade ago clocked at 3 Ghz is more powerful than a modern quad-core clocked at 2.5 Ghz. Whoever wrote this is so misinformed, but what can you expect from gamespot?

Upvote • 
Avatar image for ov3r_kill_br0ny
Ov3r_Kill_Br0ny

97

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

I have the R9 290X and it still runs almost all games maxed out at 60FPS. Can not wait to upgrade to R9 490 next year with a Zen CPU.

Upvote • 
Avatar image for DrunkenPunk800
DrunkenPunk800

1824

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

No thanks. I like AMD processors, but I buy Nvidia GPU's.

Upvote • 
Avatar image for deactivated-5b116b1818dbb
deactivated-5b116b1818dbb

34

Forum Posts

0

Wiki Points

0

Followers

Reviews: 1

User Lists: 0

@DrunkenPunk800: AMD is doing much better in the GPU industry then the CPU. maybe next year they'll catch up with the new Zen processor

Upvote • 
Avatar image for Lash540
Lash540

489

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

It's funny, Nvidia releases 750ti , it gets the full treatment, even a video review.

AMD releases the first HBM graphics card and they get errors galore in the text and no video.

*facepalm

Upvote • 
Avatar image for Kunakai
Kunakai

912

Forum Posts

0

Wiki Points

0

Followers

Reviews: 4

User Lists: 0

Edited By Kunakai

@Lash540: What's funny about it? The circumstances of both reviews are entirely different, from the people writing them to the target audience to the times they were written. I don't think the 980 ti even got a written review.

Upvote • 
Avatar image for Lash540
Lash540

489

Forum Posts

0

Wiki Points

0

Followers

Reviews: 0

User Lists: 0

Edited By Lash540

@Kunakai: It's just funny to me, I always see alot of good quality articles about Nvidia things, but not the same treatment for the AMD stuff. That is what's funny to me.

Upvote • 
Avatar image for urbangamez
urbangamez

3511

Forum Posts

0

Wiki Points

0

Followers

Reviews: 2

User Lists: 0

good show by amd, hopefully they keep improving the tech and the drivers.

Upvote • 
Avatar image for Yams1980
Yams1980

2866

Forum Posts

0

Wiki Points

0

Followers

Reviews: 27

User Lists: 0

Edited By Yams1980

the specs on paper make the fury look like it should destroy the nvidia cards but it doesn't at all. In many cases it actually falls behind the 980ti.

the drivers just need work ? maybe future games this card will become powerful who knows. i still think the 980ti is the card you should get if your willing to spend the money and want the most efficient and powerful card... and obviously the 970 for sensible users who game at 1080p and who know better things are near and don't need to burn away 500 dollars more for 15-20% speed gain.

The card is tiny though i like that, i wonder how big it would been if they had cooled it with fans?

Upvote • 
Avatar image for Pelezinho777
Pelezinho777

1520

Forum Posts

0

Wiki Points

0

Followers

Reviews: 15

User Lists: 0

Edited By Pelezinho777

Can it play Arkham Knight at 20fps on low?

Upvote • 
Avatar image for kenundrum7
kenundrum7

380

Forum Posts

0

Wiki Points

0

Followers

Reviews: 49

User Lists: 0

Edited By kenundrum7

@Pelezinho777: That made me chuckle.

Upvote •