Discussion RDNA4 + CDNA3 Architectures Thread

Page 123 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

DisEnchantment

Golden Member
Mar 3, 2017
1,672
6,153
136





With the GFX940 patches in full swing since first week of March, it is looking like MI300 is not far in the distant future!
Usually AMD takes around 3Qs to get the support in LLVM and amdgpu. Lately, since RDNA2 the window they push to add support for new devices is much reduced to prevent leaks.
But looking at the flurry of code in LLVM, it is a lot of commits. Maybe because US Govt is starting to prepare the SW environment for El Capitan (Maybe to avoid slow bring up situation like Frontier for example)

See here for the GFX940 specific commits
Or Phoronix

There is a lot more if you know whom to follow in LLVM review chains (before getting merged to github), but I am not going to link AMD employees.

I am starting to think MI300 will launch around the same time like Hopper probably only a couple of months later!
Although I believe Hopper had problems not having a host CPU capable of doing PCIe 5 in the very near future therefore it might have gotten pushed back a bit until SPR and Genoa arrives later in 2022.
If PVC slips again I believe MI300 could launch before it

This is nuts, MI100/200/300 cadence is impressive.



Previous thread on CDNA2 and RDNA3 here

 
Last edited:

Mahboi

Senior member
Apr 4, 2024
741
1,316
96
This is an interesting statement to me. I don't do any development on GPU's, and software is my second priority, so can you expand on this statement to help me understand a bit more of what is improving and how that will help game development? (Will it?)
Afraid I can't, I'm a hack.

But although more competent people surely will, I'll give you my 2c:
The history of gaming libs and indirectly GPU libs started very early in computing, most 3D rendering started in the early 90s (OpenGL 1.0 is a 1991 program and all the way to OpenGL 4.6, it is fully retrocompatible, meaning your 2015 OGL renderer is running C lines from 1991).

This means that a lot of "old" GPU stuff is really, really outdated. We're talking 10+ years before multithreading.
In the late 2000s NV started CUDA, and around 2014 AMD released Mantle which would become the core for Vulkan (and later DX12, which is a radical departure from DX 1-11, which is just like OpenGL fully retrocompatible).

Vulkan on the one hand and HIP/CUDA on the other all rebuilt the core of how a GPU operates from the software side, particularly dispatching/command buffer, batching calls, compute calls, etc. I can't give you any details (a real GPU programmer will certainly do better, ask Matias Goldberg or maybe Osvaldo Doederlein on Twitter and they'll probably know a lot more than me), but as a "broad rule", they've taken out a lot of simpler, single threaded, low optimised calls that divided the rendering process in Vertex shaders (make triangles), Texture shaders (give them textures) and Pixel shaders (I forgot what they do, I told you I'm a hack), and re-did them every frame, into a more versatile/modern architecture that batches compute calls, batches multiple shaders and reorders them, loads up a command buffer and dumps it all on the GPU at once...

NV's Mesh shaders (AMD's primitive shaders were very alike to them, but didn't win the fight for MS/Sony) essentially build 3D worlds differently, in a system that is much more parallel, batched and closer to the "compute model". It's a new paradigm on making graphics and can use GPUs far more freely/efficiently with batched calls, reorganisation of call order, and better parallelism (on CPU with multithreading and GPU with just using them more smartly).


To think I've read all of those and forgot like 60% of it...I spend way too much time learning stuff and moving on to learn more stuff instead of making money.
 

beginner99

Diamond Member
Jun 2, 2009
5,221
1,594
136
If that was true, then a 3-chiplet (or even 4) variant should have been relatively simple, and it would have in part satisfied the high end market. So I am quite skeptic about that.

I don't think so. 2 dies is much easier than 4 due to communication between the dies. it adds up quickly and you don't want a connection between every single die and then maybe lag becomes a problem. 2 probably also don't need special packing technology and it limits the amount of redundance. N44 would be a lot larger if it needed to support 2 or more connections vs just one.
 
Reactions: Tlh97 and coercitiv

marees

Member
Apr 28, 2024
98
64
46
Getting impatient to know potential release date of RDNA 4. Will it be put on hold until FSR 4/PS5 pro with AI upscaling ?

AMD used to leak like a sieve during Raja Koduri days
I wonder what Lisa Su did/is doing to the leakers 🤔
 

linkgoron

Platinum Member
Mar 9, 2005
2,333
857
136
Getting impatient to know potential release date of RDNA 4. Will it be put on hold until FSR 4/PS5 pro with AI upscaling ?

AMD used to leak like a sieve during Raja Koduri days
I wonder what Lisa Su did/is doing to the leakers 🤔
They moved to Intel and then left to a startup.
 

jpiniero

Lifer
Oct 1, 2010
14,820
5,433
136

Looks like Lisa's got more important products to talk about at Computex... the W7900 AI Edition!
 

branch_suggestion

Senior member
Aug 4, 2023
292
625
96

Looks like Lisa's got more important products to talk about at Computex... the W7900 AI Edition!
You laugh but it sets the foundations for N50.
 

Tuna-Fish

Golden Member
Mar 4, 2011
1,419
1,749
136

Looks like Lisa's got more important products to talk about at Computex... the W7900 AI Edition!

I don't really understand what specifically makes this a great AI product? Unless they managed to score 24Gb modules from somewhere, there seems to be nothing here but AI on the shroud and a slimmer cooler?
 

jpiniero

Lifer
Oct 1, 2010
14,820
5,433
136
I don't really understand what specifically makes this a great AI product? Unless they managed to score 24Gb modules from somewhere, there seems to be nothing here but AI on the shroud and a slimmer cooler?

I am wondering if 512-bit would be possible with N31.. but I think it's just a minor refresh so they could put AI in the name. Cuz you know, everything is AI now.
 

Aapje

Golden Member
Mar 21, 2022
1,466
2,031
106
I don't really understand what specifically makes this a great AI product? Unless they managed to score 24Gb modules from somewhere, there seems to be nothing here but AI on the shroud and a slimmer cooler?

Dual slot blower means that companies can put multiple of these card on a server motherboard and run jobs in parallel. Nvidia sells slim GPUs to companies for a similar reason.
 

dhruvdh

Junior Member
Apr 2, 2024
17
30
51
I don't really understand what specifically makes this a great AI product? Unless they managed to score 24Gb modules from somewhere, there seems to be nothing here but AI on the shroud and a slimmer cooler?
With two of these (and they want you to use two, because dual slot), you can run a Llama 3 70B class models in FP8 locally.

That means basically running it lossless. The Llama 3 70B model is much, much "smarter" than the original ChatGPT release.

If you want to have this model do something for you in a loop, be always listening, run generation on individual rows a CSV file with lots of rows - it increasingly becomes extremely expensive to use cloud.

Personally, I would want it for help with organizing my Zotero library. Add tags based on paper title, abstract.

I would also want to use it to ingest my bank/credit card statements, and categorize transactions, help turn them into text double-entry ledgers.

But the current W7900 is 4000$ and hard to buy. You would need 2. Let's see what they do with pricing.
 

SolidQ

Senior member
Jul 13, 2023
405
416
96
Mean no RDNA4 soon?
 
Reactions: marees

soresu

Platinum Member
Dec 19, 2014
2,921
2,142
136
Reactions: Tlh97 and blckgrffn

blckgrffn

Diamond Member
May 1, 2003
9,192
3,158
136
www.teamjuchems.com
Ot they could launch navi 48 first & navi 44 later
A staggered launch makes sense because you can keep your product in the news cycle for longer, but I don't think AMD wants to drag this out. Their best chance for the longest possible competitive relevance of these SKUs is to roll them out as quick as strategy and supply chain allows. This could be a rare window where they upstage the nvidia SKUs arrayed against their specific price points by a margin large enough to capture some mindshare.

I might be very wrong.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |