AMD Reveals Radeon VII: High-End 7nm Vega Video Card Arrives February 7th for $699by Ryan Smith on January 9, 2019 1:00 PM EST
As it turns out, the video card wars are going to charge into 2019 quite a bit hotter than any of us were expecting. Moments ago, as part of AMD’s CES 2019 keynote, CEO Dr. Lisa Su announced that AMD will be releasing a new high-end, high-performance Radeon graphics card. Dubbed the Radeon VII (Seven), AMD has their eyes set on countering NVIDIA’s previously untouchable GeForce RTX 2080. And, if the card lives up to AMD’s expectations, then come February 7th it may just as well do that.
Today’s announcement is interesting in that it’s just as much about technology as it is the 3D chess that is the market positioning fights between AMD and NVIDIA. Technically AMD isn’t announcing any new GPUs here – regular readers will correctly guess that we’re talking about Vega 20 – but the situation in the high-end market has played out such that there’s now a window for AMD to bring their cutting-edge Vega 20 GPU to the consumer market, and this is a window AMD is looking to take full advantage of.
At a high level then, the Radeon VII employs a slightly cut down version of AMD’s Vega 20 GPU. With 60 of 64 CUs enabled, it actually has a few less CUs than AMD’s previous flagship, the Radeon RX Vega 64, but it makes up for the loss with much higher clockspeeds and a much more powerful memory and pixel throughput backend. As a result, AMD says that the Radeon VII should beat their former flagship by anywhere between 20% and 42% depending on the game (with an overall average of 29%), which on paper would be just enough to put the card in spitting distance of NVIDIA’s RTX 2080, and making it a viable and competitive 4K gaming card.
|AMD Radeon Series Specification Comparison|
|AMD Radeon VII||AMD Radeon RX Vega 64||AMD Radeon RX 590||AMD Radeon R9 Fury X|
|Memory Clock||2.0Gbps HBM2||1.89Gbps HBM2||8Gbps GDDR5||1Gbps HBM|
|Memory Bus Width||4096-bit||2048-bit||256-bit||4096-bit|
|Single Precision Perf.||13.8 TFLOPS||12.7 TFLOPS||7.1 TFLOPS||8.6 TFLOPS|
|Manufacturing Process||TSMC 7nm||GloFo 14nm||GloFo/Samsung 12nm||TSMC 28nm|
|GPU||Vega 20||Vega 10||Polaris 30||Fiji|
|GCN 4||GCN 3|
Diving into the numbers a bit more, if you took AMD’s second-tier Radeon Instinct MI50 and made a consumer version of the card, the Radeon VII is almost exactly what it would look like. It has the same 60 CU configuration paired with 16GB of HBM2 memory. However the Radeon VII’s boost clock is a bit higher – 1800MHz versus 1746MHz – so AMD is getting the most out of those 60 CUs. Still, it’s important to keep in mind that from a pure FP32 throughput standpoint, the Vega 20 GPU was meant to be more of a sidegrade to Vega 10 than a performance upgrade; on paper the new card only has a 9% compute throughput advantage. So it’s not on compute throughput where Radeon VII’s real winning charm lies.
Instead, the biggest difference between the two cards is on the memory backend. Radeon Vega 64 (Vega 10) 2 HBM2 memory channels running at 1.89Gbps each, for a total of 484GB/sec of memory bandwidth. Radeon VII (Vega 20) doubles this and then some to 4 HBM2 memory channels, which also means memory capacity has doubled to 16GB. And then there’s the clockspeed boost on top of this to 2.0Gbps for the HBM2 memory. As a result Radeon VII has a lot memory bandwidth to feed itself, from the ROPs to the stream processors. Given these changes and AMD’s performance estimates, I think this lends a lot of evidence to the idea that Vega 10 was underfed – it needed more memory bandwidth keep its various processing blocks working at full potential – but that’s something we’ll save for the eventual review.
Past that, as this is still a Vega architecture product, it’s the Vega we all know and love. There are no new graphical features here, so even if AMD has opted to shy away from putting Vega in the name of the product, it’s going to be comparable to those parts as far as gaming is concerned. The Vega 20 GPU does bring new compute features – particularly much higher FP64 compute throughput and new low-precision modes well-suited for neural network inferencing – but these features aren’t something consumers are likely to use. Past that, AMD will be employing some mild product segmentation here to avoid having the Radeon VII cannibalize the MI50 – the Radeon VII does not get PCIe 4.0 support, nor does it get Infinity Link support –
The other wildcard for the moment is TDP. The MI50 is rated for 300W, and while AMD’s event did not announce a TDP for the card, I fully expect AMD is running the Radeon VII just as hard here, if not a bit harder. Make no mistake: AMD is still having to go well outside the sweet spot on their voltage/frequency curve to hit these high clockspeeds, so AMD isn’t even trying to win the efficiency race. Radeon VII will be more efficient than Radeon Vega 64 – AMD is saying 25% more perf at the same power – but even if AMD hits RTX 2080’s performance numbers, there’s nothing here to indicate that they’ll be able to meet its efficiency. This is another classic AMD play: go all-in on trying to win on the price/performance front.
Accordingly, the Radeon VII is not a small card. The photos released show that it’s a sizable open-air triple fan cooled design, with a shroud that sticks up past the top of the I/O bracket. Coupled with the dual 8-pin PCIe power plugs on the rear of the card, and it’s clear AMD intends to remove a lot of heat. Both AMD and NVIDIA have now gone with open-air designs for their high-end cards on this most recent generation, so it’s an interesting development, and one that favors AMD given their typically higher TDPs.
Vendor performance claims should always be taken with a grain of salt, but for the moment this is what we have. If AMD manages to reach RTX 2080 performance, then I expect this to be another case of where the two cards are tied on average but are anything but equal; there will be games where AMD falls behind, games where they do tie the RTX 2080, and then even some games they pull ahead in. These scenarios are always the most interesting for reviewers, but they’re also a bit trickier for consumers since it means there’s no clear-cut winner.
All told then, the competitive landscape is going to be an interesting one. AMD’s own proposition is actually fairly modest; with a $699 price tag they’re launching at the same price as the RTX 2080, over four months after the RTX 2080. They are presumably not going to be able to match NVIDIA’s energy efficiency, and they won’t have feature parity since AMD doesn’t (yet) have its own DirectX Raytracing (DXR) implementation.
But what AMD does have, besides an at least competitive price and presumably competitive performance in today’s games, is a VRAM advantage. Whereas NVIDIA didn’t increase their VRAM amounts between generations, AMD is for this half-generation card, giving them 16GB of VRAM to RTX 2080’s 8GB. Now whether this actually translates into a performance advantage now or in the near future is another matter; AMD has tried this gambit before with the Radeon 390 series, where it didn’t really pay off. On the other hand, the fact that NVIDIA’s VRAM capacities have been stagnant for a generation means that AMD is delivering a capacity increase “on schedule” as opposed to ahead of schedule. So while far from guaranteed, it could work in AMD’s favor. Especially as, given the performance of the card, AMD intends for the Radeon VII to be all-in on 4K gaming, which will push memory consumption higher.
Finally on the gaming front, not content to compete on just performance and pricing, AMD will also be competing on gaming bundles. The Radeon VII will be launching with a 3 game bundle, featuring Resident Evil 2, Devil May Cry 5, and The Division 2. NVIDIA of course launched their own Anthem + Battlefield V bundle at the start of this week, so both sides are now employing their complete bags of tricks to attract buyers and to prop up the prices of their cards.
Speaking of pricing, perhaps the thing that surprises me the most is that we’re even at this point – with AMD releasing a Vega 20 consumer card. When they first announced Vega 20 back in 2018, they made it very clear it was going to be for the Radeon Instinct series only. That the new features of the Vega 20 GPU were better suited for that market, and more importantly as a relatively large chip (331mm2) for this early in the life of TSMC’s 7nm manufacturing node, yields were going to be poor.
So that AMD is able to sell what are admittedly defective/recovered Vega 20s in a $699 card, produce enough of them to meet market demand, and still turn a profit on all of this is a surprising outcome. I simply would not have expected AMD to get a 7nm chip out at consumer prices this soon. All I can say is that either AMD has pulled off a very interesting incident of consumer misdirection, or the competitive landscape has evolved slowly enough that Vega 20 is viable where it otherwise wouldn’t have been. Or perhaps it’s a case of both.
Shifting gears for a second, while I’ve focused on gaming thus far, it should be noted that AMD is going after the content creation market with the Radeon VII as well. This is still a Radeon card and not a Radeon Pro card, but as we’ve seen before, AMD has been able to make a successful market out of offering such cards with only a basic level of software developer support. In this case AMD is expecting performance gains similar to the gaming side, with performance improving the more a workload is pixel or memory bandwidth bound.
Wrapping things up, the Radeon VII will be hitting the streets on February 7th for $699. At this point AMD has not announced anything about board partners doing custom designs, so it looks like this is going to be a pure reference card launch. As always, stay tuned and we should know a bit more information as we get closer to the video card’s launch date.
Post Your CommentPlease log in or sign up to comment.
View All Comments
rarson - Monday, January 21, 2019 - linkMy guess is that AMD saw the raster performance of the 2080 wasn't so spectacular and realized that their 7nm Vega could get reasonably close at around the same price. I suspect that AMD is still losing money on these even at $699 due to the high cost of HBM2. I also suspect that either Navi isn't quite as ready as they expected or they realized they needed to allocate more resources towards console chips. Either way, it's clearly a stop-gap, but if performance lives up to the claims, I would say it's not a very bad one. Had they stuck to the original plan, I imagine we would have had to wait significantly longer for a new card but that it would probably be a Navi-based one.
Cellar Door - Wednesday, January 9, 2019 - linkTo be honest - this is NOT gonna fly at $699. Why would anyone pick this over the 2080 that has the raytrascing tech and DLSS.
AMD seemed to miss their opportunity here.
marees - Wednesday, January 9, 2019 - linkRay tracing not efficient at 4k
and DLSS equivalent to 1800p super sampling with TAA
so people who have 4k and/or high frequency monitors have a choice now
FMinus - Wednesday, January 9, 2019 - linkBecause raytracing is a buzzword nothing more nothing less, maybe in 2025 it will mean something, now it doesn't.
webdoctors - Wednesday, January 9, 2019 - link699? I thought the AMD fanboys said it'll be 399 and they'd never rip them off unlike the other guys. What happened to the comments about competition would bring down prices? Trolls haven't woke up yet?
But wow, 16 GB of HBM2......
KateH - Wednesday, January 9, 2019 - linkAMD fanatics said it would be Navi at $399 with 1080/2070-level performance (not a fangirl here but i spend too much time reading tech rumors). This is a bit of a curveball
GreenReaper - Wednesday, January 9, 2019 - linkThere probably will be one like that... just not yet.
mapesdhs - Thursday, January 10, 2019 - linkI can't see it selling at this price, it's just too high. Doesn't matter how good it is, AMD needs something far cheaper to gain both market share and mindshare. Tech sits will also bitch about power/heat/noise even though most gamers really don't care at this level. People simply don't buy AMD often enough, even when it makes perfect sense to do so. AMD isn't going to get back into the GPU game by targeting what now passes for high end these days. Roughly the same as a 2080 at about the same cost won't work. It needs to be waaaay cheaper to really make tech sites yell yes, buy this thing, and also be such a gap that NVIDIA can't simply lower its prices to counter the difference without making it very clear their own products never had to be so expensive in the first place. Doing it like this, NVIDIA can respond merely by making the 2080 slightly cheaper, and it doesn't have to be much of a margin to have the desired effect. Worse, there are people who openly say they wanted AMD to release something decent just so they could get a cheaper NVIDIA card (these are the NPCs who help to maintain market stagnation; if people don't buy AMD based on such a deluded minset then AMD just isn't going to bother with this market at all, why should they?).
Paul ("not an apple fan" on youtube) has talked about these issues at length, worth checking out.
A product like this needed to be $500 or below. Until the NPC zoids stop defaulting to NVIDIA whenever things are more or less equal, this situation won't change, and NVIDIA can continue to screw the consumer with overpriced tech passed off as advances which are just the table scraps from the compute market (what we're getting now are not gaming cards, they're not designed for *gaming*, as The Good Old Gamer explained so well some months ago).
Arbie - Friday, January 11, 2019 - linkI suppose "AMD fanatics" is just your shorthand for "folks trying to predict what graphics cards AMD would be able to release in 2019", and somehow not meant to be offensive.
sing_electric - Wednesday, January 9, 2019 - linkAMD and Nvidia have really different architectures - we'll see if the GCN successor (next year?) starts looking more like Nvidia but I doubt it.
AMD DOES support GPU-enhanced ray tracing on Vulkan, and there's nothing "magical" about the cores in Nvidia's RTX series that enable it - they're just optimized. We've seen that GCN is actually surprisingly good at doing a lot of compute tasks, and so it wouldn't surprise me if AMD could support hardware DirectX raytracing on existing GPUs in the future.
We *still* don't know how DLSS works, but it seems like it's just an AI algorithm that allows most of a scene to be rendered at below the advertised resolution while upscaling select parts (and applying TAA on others?), so again, it's not "nothing" but it's just software around which Nvidia may have optimized parts of its hardware. There is a chance AMD could announce something similar, again, on existing hardware, though I'd kind of doubt it.