cancel
Showing results for 
Search instead for 
Did you mean: 

General Discussions

AMD Announces CDNA, RDNA2 Architectures, Significant Leap in Performance-per-Watt

AMD made several GPU-themed announcements at its Analyst Day today. While we’ll have coverage of the wider event, we wanted to focus specifically on the GPU announcements the company has made. AMD announced a pair of new architectures: CDNA and RDNA.

AMD-GPU-1

RDNA2: More Than Big Navi

AMD didn’t say a ton about RDNA2, but they did make some announcements about the kind of improvements users should expect. When AMD launched Navi 10, there were some discontented mutterings about the fact that the new 7nm GPUswere only on-par with Nvidia’s 12nm parts, as far as performance-per-watt was concerned. While this represented a very real improvement over GCN, AMD hadn’t delivered enough improvement to close the gap in a single bound.

According to the company, RDNA2 will deliver a substantial additional improvement in performance per watt.

It’s always wise to take claims like this with a grain of salt, but fortunately, we can look back to our original 5700 and 5700 XT review to see how Navi compared with GCN. The Radeon 5700 offers nearly identical performance to the Vega 64, but it draws substantially less power:

5700-GPU-Power (1)

If you divide the Vega 64’s measured 347W by 1.5x, you’d expect the RDNA-based 5700 to draw 231W, compared with the 256W we observed. Our observed improvement, therefore, is somewhat smaller than AMD’s assumption, with a 1.36x improvement compared with a claimed 1.5x.

Keep in mind, however, that AMD undoubtedly measures performance-per-watt in a suite of titles, while we gather data in just one game / resolution. Yes, companies are going to pick data that puts their best foot forward, but our single data point is meant to be representative, not definitive. Regardless, AMD is claiming it can deliver the equivalent benefit of a full node shrink with further architectural improvements. That may be reasonable, given that the company didn’t have nearly as much time to develop RDNA after GCN as it had to develop Ryzen following Piledriver. AMD’s own timeline explicitly recalls that comparison:

A further 1.5x – 1.36x improvement in performance-per-watt would give AMD a decisive advantage versus Turing. We have absolutely no information on how it would compare against Ampere, and therefore won’t speculate. Even if Nvidia retained overall performance-per-watt leadership, however, delivering two generational improvements of 1.5x (or even 1.36x) in a row would indicate AMD was taking the issue far more seriously than it has in the past. GCN became modestly more power efficient over its long life, but AMD rode the ragged edge of its TDP ratings to maximize performance against Nvidia from virtually the moment the architecture launched in 2012 to its final dGPU iteration as the Radeon VII in early 2019.

The bottom line is this: Best-case, AMD will be on much stronger competitive footing with Nvidia as far as performance-per-watt. Worst-case — assuming AMD delivers on these figures but Nvidia gains more from the 7nm shrink — is that AMD will still have a solid story to tell about this aspect of its business.

AMD gave no information on segmentation for RDNA2 but emphasized that it was an enthusiast-class architecture. It’s obviously intended to take the fight to Nvidia’s upper-tier products and AMD has stated it won’t be calling the chip “big Navi,” because the level of advances from RDNA to RDNA2 deserve a better name. It’s been quite some time since AMD sunk much effort into quickly iterating on a GPU architecture, but there was a time when the company was known for this kind of rapid-fire iteration. The HD 2000 series hit market (or, more accurately, splatted into it) in 2007, to absolutely no one’s delight. AMD rapidly evolved the dead-letter HD 2000 family into the more-respectable HD 3000 family by late-year, launched the competitive HD 4000 family by mid-2008, and had the first DX11 GPUs in-market by September, 2009.

A fairly rapid iteration between RDNA (July 2019) and RDNA2 (?, 2020), in other words, would be more the historical exception rather than the rule.

CDNA: Data Center Driven

Although AMD didn’t explicitly make this connection during the event, the emergence of CDNA as a separate architecture may explain why we’ve seen no formal ROCm support for Navi 10 under Linux. (ROCm is AMD’s open source GPGPU computing platform that translates CUDA into code AMD GPUs can run).

AMD’s compute-centric roadmap starts with GCN in 2019 (Radeon Instinct MI50 and MI60), progresses through CDNA, and arrives at “CDNA2” in 2022. CDNA2, therefore, would be the architecture that supports the El Capitan supercomputer. CDNA is a compute-centric version of RDNA, but AMD didn’t give specifics about the changes between the two families beyond that statement. Sometimes it’s possible to draw conclusions about what one company will do by examining the behavior of a competitor, but Nvidia has actually used several different strategies for its high-end Tesla GPUs compared with its consumer cards. There have been times when Team Green deployed significantly different architectures for Tesla compared with GeForce and times when the company tapped the same core design for parts in both families.

Click to enlarge. This slide is cropped differently to maximize visibility of the small text elements. Gray = Unsupported, Dark blue = early, and bright blue = full production.

One major feature of CDNA? Fully connected, cache-coherent architecture between CPU and GPU, which CDNA will introduce. AMD did not specifically say if CDNA corresponds exactly to RDNA, while CDNA2 corresponds to RDNA2. In theory, CDNA might have a similar relationship to RDNA that Nvidia’s old GK104 had to GK110. Both chips implemented the Kepler architecture, but GK110 supported several features that GK104 didn’t, in addition to packing more GPU cores and supporting faster double-precision floating-point performance.

AMD Announces CDNA, RDNA2 Architectures, Significant Leap in Performance-per-Watt - ExtremeTech 

0 Likes
4 Replies

Bit more information from Guru3D's article. They said AMD announced RNDA 2 (Navi 2X) cards will not be released until the END of this year. This. Sucks.

https://www.guru3d.com/news-story/amd-talks-rdna2raytracingroadmaps-and-a-possible-teaser-of-big-navi.html

Lots of lead time allows people time to scrape it up to get a new video card. Not everyone is as rich as Bill Gates etc.

0 Likes

Usually I ignore your inane, usually irrelevant responses to threads, but I can't this time.

Which, if you had a modicum of common sense, you would know this means that prices will remain high for the foreseeable future, not being cut to clear inventory due to the next generation of products, and being effectively reduced due to lower tier, lower price next generation hardware having the same or higher performance than higher tier, higher price current generation hardware.

This is not a good thing, this is not "lots of lead time", this is extremely bad news for anyone who requires a new graphics card in a time where prices of current generation hardware are still in excess of 175% of the previous generation cards they replaced because AMD is not interested in a price war which leaves nVidia with no reason to cut prices, which leaves previous generation hardware, 3-4 years old already, as the most accessible to most of the market, the $200 market, and those cards are only 1920x1080 class. This leaves nVidia as the only manufacturer if you want to experience ray tracing, which will cost you, minimum, $300 for the lowest end offering,  only leaves nVidia as the only manufacturer of a mid range graphics card even remotely capable of 4k60 average in even SOME games, the 2070 Super which is priced north of $500, and leaves nVidia as the only manufacturer of a graphics card that can guarantee a 4k60 experience, but even then WITHOUT ray tracing, the 2080Ti priced upwards of $1900.

0 Likes

I have seen snippets of news about unsold video cards. Slashing prices will clear leftovers that is basic business.

Ray Tracing at 4K is so expensive as to be prohibitive. Even the RTX 2080 is hard pressed.

The last GeForce card I acquired was a GTX 1060 which is Pascal. I do not have any Turing cards. At the prices wanted they can keep them. As I complained on my site, RX 5700 XT cards are way overpriced and as long as I see ripoff pricing, to hell with them all.

My RX 480 is still working fine. All I play right now is Halo which is not very demanding.

0 Likes