Discussion RDNA 5 / UDNA (CDNA Next) speculation

Page 7 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

stayfrosty

Junior Member
Apr 4, 2024
18
50
51
Isn't a design with only 2 retscale AID's kind of small? I thought with CoWoS going 5.5x reticle they would scale bigger. The design doesn't seem that much larger than the MI355x (which also has 2x retscale AID's)
 

basix

Member
Oct 4, 2024
87
177
66
If you extract things like media engines, Infinity Fabric IO/PHY, PCIe and reduce the amount of chip-to-chip interfaces you get much more usable Die area. At the same time, energy efficiency gets improved. Especially if you compare it to a 6x Base Die configuration (Die-to-Die interconnects, increased bandwidth requirements, pJ/bit). And I don't know if AMD sticks to CoWoS or changes back to EFB like used on MI250X.

The current Base Die is 13 x 29mm according to AMD. If you go beyond 800mm2 and exctract before mentioned IP-Blocks, you might land at 1.3x more usable area. Combined with N4 instead of N6 and 6um SoIC instead of 9um SoIC the gains are quite decent.
The same applies to the Top Die. You can now increase its Die Size by 1.3x (more stuff can be kept on same chip) and N3 greatly increases transistor density.

Another benefit: You do not need two Base Die version (normal and mirrored).

If you go forward to MI400 you could extend that to 12x HBM stacks and possible 3x Base Die, I don't know. But increased usable Die area and no need for Base Die mirroring stays as benefits. It would then somehow look like Clearwater Forest from Intel. But with HBM4 and much bigger Die as additions.

And maybe as an extra:
You could more easily switch or extend your Infinity Fabric Phy with e.g. LPDDR5/6 PHY and add much more and cheaper memory capacity
 

stayfrosty

Junior Member
Apr 4, 2024
18
50
51
If you go forward to MI400 you could extend that to 12x HBM stacks and possible 3x Base Die, I don't know. But increased usable Die area and no need for Base Die mirroring stays as benefits. It would then somehow look like Clearwater Forest from Intel. But with HBM4 and much bigger Die as additions.
Agree with all you said and it absolutely makes sense for MI355x. But i'm just a bit disappointed that apparently according to the article linked in the previous page that MI400 will stay with a 2x Base Die setup. So in absolute terms the design will not be THAT much bigger than MI355x. Well, except if they start stacking compute chiplets, but I doubt that.
 
Reactions: Tlh97 and bearmoo

basix

Member
Oct 4, 2024
87
177
66
MI400 AID/MID are N3P
Interesting.

Even more space for "stuff"

Agree with all you said and it absolutely makes sense for MI355x. But i'm just a bit disappointed that apparently according to the article linked in the previous page that MI400 will stay with a 2x Base Die setup. So in absolute terms the design will not be THAT much bigger than MI355x. Well, except if they start stacking compute chiplets, but I doubt that.
Still, if N3P for AID and MID is true, ~800mm2 AID and let's say ~200...250mm2 for each MID ist ist still much more area than MI300X and also Blackwell. And then add the fact, that compared to Blackwell you add another +600mm2 or so N2 XCDs on each AID. This thing is HUGE.

Additionally, you can scale that design to 3x AIDs, 12x XCDs and 12x HBM-Stacks. Maybe as an MI400 refresh one year later?
 
Last edited:

marees

Senior member
Apr 28, 2024
948
1,263
96
Is CDNA Next the same as UDNA ?

We are seeing strong customer interest in the MI400 series for large-scale training and inference deployments and remain on track to launch in 2026.
The CDNA next architecture takes another major leap enabling powerful rackscale solutions that tightly integrate networking CPU and GPU capabilities at the silicon level to support Instinct solutions at data center scale. We designed CDNA next to deliver leadership AI and HPC flops while expanding our memory capacity and bandwidth advantages and supporting an open ecosystem of scale-up and scale-out networking products.

 

RnR_au

Platinum Member
Jun 6, 2021
2,344
5,626
106
UDNA doesn't exist according to one persistent rumour on this forum. Was UDNA even mentioned in the Earnings Call? Maybe usage of 'CDNA Next' shows that UDNA is indeed dead.

edit: no... no mention of UDNA from a quick search on that transcript page.
 

marees

Senior member
Apr 28, 2024
948
1,263
96
UDNA doesn't exist according to one persistent rumour on this forum. Was UDNA even mentioned in the Earnings Call? Maybe usage of 'CDNA Next' shows that UDNA is indeed dead.

edit: no... no mention of UDNA from a quick search on that transcript page.
Updated the thread title now...
 

soresu

Diamond Member
Dec 19, 2014
3,689
3,026
136
UDNA doesn't exist according to one persistent rumour on this forum. Was UDNA even mentioned in the Earnings Call? Maybe usage of 'CDNA Next' shows that UDNA is indeed dead.

edit: no... no mention of UDNA from a quick search on that transcript page.
I would suggest that UDNA is real, but simply represents a long term unified product stack strategy, rather than a more absolute single µArch per generation like nVidia has.

Think of it like RDNA and CDNA sharing the same ISA for general CU + matrix units, while each has differing layout or ratio for these, and CDNA probably remains headerless.

This way software should be far more portable between the 2 separate SKU markets, while still allowing for domain specific µArch differences.

Sort of like GCN days when the enthusiast SKU had vastly higher FP64 perf vs all lower segments, something like 1/2 rate vs 1/16th rate.
 
Last edited:
Reactions: scineram

Joe NYC

Platinum Member
Jun 26, 2021
2,941
4,314
106
If Medusa Halo has a WoW stacked GPU then surely the flagship dGPU will too.

If by WoW stacked, you mean Wafer on Wafer stacked, has there been any leak / suggestion of that?

Because, even though it seemed the way Zen 5 + V--Cache are configured, it seemed it was that way to enable WoW stacking, in the end it seems that AMD is NOT doing WoW stacking on Zen 5.

It would be neat if Medusa Halo used stacking, to put IOD functionality on a cheaper die and then a smaller iGPU on more advanced node. It might be justified for "Halo" product, even if it doesn't trickle down to mainstream products right away.
 

Joe NYC

Platinum Member
Jun 26, 2021
2,941
4,314
106
Yeah, the new one is seemingly.
--------XCDx4--XCDx4--------
MID----AID-------AID----MID
Presumably it will all be hooked up like N4C would've been.

I wonder if the MID serves as a base for HBM4 memory, including logic portion of HBM4.

If AMD is talking about memory being one of the differentiating factors going forward, that would be a possible approach.
 

basix

Member
Oct 4, 2024
87
177
66
I do not see HBM as part of MID (Multimedia Interface Die). HBM as close as possible to compute is the best idea, therefore attached to AID.

You can differentiate in other ways:
  • HBM elevated a little (EFB?) and directly attached on AID at their edges (so no HBM signal routing through the interposer), which roughly halves pJ/bit
  • Stacked SRAM / V-Cache
  • HBM4 could use PIM
  • MID could feature additional LPDDR5/6 channels (cheaper memory and higher total memory capacity)
 
Last edited:

Joe NYC

Platinum Member
Jun 26, 2021
2,941
4,314
106
I do not see HBM as part of MID (Multimedia Interface Die). HBM as close as possible to compute is the best idea, therefore attached to AID.

You can differentiate in other ways:
  • HBM elevated a little (EFB?) and directly attached on AID (so no HBM signal routing through the interposer), which roughly halves pJ/bit
  • Stacked SRAM / V-Cache
  • HBM4 could use PIM
  • MID could feature additional LPDDR5/6 channels (cheaper memory and higher total memory capacity)

The way I would picture it is, based on some info leaked
- 2 AIDs next to each other connected with fast link, maybe hybrid bond linked silicon bridge
- XCDs stacked on top of AIDs
- MID next to each AID, maybe connected with same bridge silicon bridge (or EFB?)
- HBM4 stacks (2 or 4?) stacked on top of MIO, with the base logic die being part of MID, PIM could be part of MID / HBM combo.

This would mean AID would no longer have memory controllers talking directly to HBM, it would just have a networking interface talking to neighboring bridges.

Then, custom product could be developed with different numbers of AIDs (compute) and MIDs (memory capacity)

One sticking point is if Mi400 is expected to use HBM4. I have not seen this question addressed on either NVidia or AMD side, which is what would be the first product to use HBM4
 

poke01

Diamond Member
Mar 8, 2022
3,383
4,626
106
So N48 is “boring”, will RDNA5 be exciting in terms of design and SKUs?
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |