Discussion RDNA4 + CDNA3 Architectures Thread

Page 299 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

DisEnchantment

Golden Member
Mar 3, 2017
1,754
6,631
136





With the GFX940 patches in full swing since first week of March, it is looking like MI300 is not far in the distant future!
Usually AMD takes around 3Qs to get the support in LLVM and amdgpu. Lately, since RDNA2 the window they push to add support for new devices is much reduced to prevent leaks.
But looking at the flurry of code in LLVM, it is a lot of commits. Maybe because US Govt is starting to prepare the SW environment for El Capitan (Maybe to avoid slow bring up situation like Frontier for example)

See here for the GFX940 specific commits
Or Phoronix

There is a lot more if you know whom to follow in LLVM review chains (before getting merged to github), but I am not going to link AMD employees.

I am starting to think MI300 will launch around the same time like Hopper probably only a couple of months later!
Although I believe Hopper had problems not having a host CPU capable of doing PCIe 5 in the very near future therefore it might have gotten pushed back a bit until SPR and Genoa arrives later in 2022.
If PVC slips again I believe MI300 could launch before it

This is nuts, MI100/200/300 cadence is impressive.



Previous thread on CDNA2 and RDNA3 here

 
Last edited:

beginner99

Diamond Member
Jun 2, 2009
5,281
1,695
136
straight up denied AI is useful and thinks the whole thing is a bubble but lol.

To some degree it is a bubble. AI /LLM can be useful in some contexts but these are by now mostly already being used with integrations in office, email and software dev products. once the market has been established and isn't growing exponentially, the demand for new GPUs for AI will go down massively as growth becomes incremental.
 
Reactions: Saylick

SpudLobby

Golden Member
May 18, 2022
1,027
695
106
To be factually correct, there is no such thing as DeepSeek R1 30-32B model. What you are talking about are distills.
Forgot that. Fair enough, I’ve used the LlaMa ones (Haven’t tried Qwen yet). Though to add on, it actually reinforces my point here that they are distilled at the base instead of using V3, because Qwen and LlaMa — LlaMa especially, are just okay to good, not great, it’s just that they’re open and small. But fine tuning them on the reasoning here while prolonging outputs produces vastly better results just not possible at similar scales otherwise. I was just demonstrating the direction of effect here is the opposite of what people expect.

Still think local LLM’s are overrated if not outright hype. Batched inference is vastly more efficient and latency is easily low enough for cloud efficiency & speed gains to be worthwhile, or outright better capability. Spare your RAM, battery life, get real capability and more speed.



In some sense this is a win for AMD among others vs Nvidia in client — most people aren’t going to go crazy for local LLMs. Upscaling tech or other small AI models doing summarization or creative stuff sure, but that’s a meme on memory/compute compared to this.
 
Reactions: exquisitechar

SpudLobby

Golden Member
May 18, 2022
1,027
695
106
To some degree it is a bubble. AI /LLM can be useful in some contexts but these are by now mostly already being used with integrations in office, email and software dev products. once the market has been established and isn't growing exponentially, the demand for new GPUs for AI will go down massively as growth becomes incremental.
Nonzero bubble quite possible on the hardware end for Nvidia alone if not quite likely — and there may be a 5 steps forward but 1-2 steps back thing at some point, much like dotcom era. It’s not even close to crypto which is scams and speculation.

The statements by those here among other hardware types for years have been totally delusional re: the utility, commercial viability*, etc.

*Like OpenAI lost $5-6B last year per NYT, but with the rate of improvements in cost on the model side they could pretty much double to triple the price of API calls today and increase plus prices and it would still net positive. If the response to this is “well not after DeepSeek V3/R1 is hosted on American hyperscalers at performance (current MS one is like a trial, is slow)” etc, then like, okay. Thanks for demonstrating AI can absolutely be cheap. Just means the hyperscalers will do it and/or models can still improve. It isn’t going anywhere. Nvidia itself I am ambivalent on.

Gemini 2.0 Flash is also utterly dirt cheap, good and Flash Thinking is unlikely to be much different.
 

Keller_TT

Member
Jun 2, 2024
113
112
76
The bubble is not about AI being a scam, it's about companies being overvalued considering many are burning billions while digging for a breakthrough that is supposed to materialize into untold riches. Nvidia will be more than fine, they're selling the shovels.
AI, to the extent it is being hyped as if it's the greatest thing since the invention of the wheel, and the sheer greed and deception involved in marketing anything with an "AI" sticker, is both a scam and a massively inflated bubble.
GenAI can be optimized for efficiency but in terms of big leaps in performance, we're well into diminishing returns without being able to gather and train data at the scale of the universe and obliterating privacy. I don't expect the US to lead there. May be my most capitalistic country with Direct Democracy with a dedicated canton as Crypto Valley with regulatory freedom for building a decentralized tech stack will do something about it and the EU will flatter us with their imitation.
 

Heartbreaker

Diamond Member
Apr 3, 2006
4,652
6,107
136
AI, to the extent it is being hyped as if it's the greatest thing since the invention of the wheel, and the sheer greed and deception involved in marketing anything with an "AI" sticker, is both a scam and a massively inflated bubble.

Whatever level the AI "Bubble" is at, it's deflation to more realistic levels won't harm NVidia. NVidia's inflated stock price is more or less irrelevant to the company. It's not seeking new investment or needing to raise capital. It's a money printing machine.

Unlike Crypto Mining, AI has real value, it's not going to pop like a bubble. It's going to soft land.
 

basix

Member
Oct 4, 2024
40
75
51
GenAI can be optimized for efficiency but in terms of big leaps in performance, we're well into diminishing returns without being able to gather and train data at the scale of the universe and obliterating privacy.
Well, DeepSeek R1 has just recently shown a new avenue regarding model training, which for the majority relies on synthetic data. And synthetic data can be pushed much more easily towards "the scale of the universe".
The same applies for Chain of Thought. This one primarily scales with compute and not so much with additional data.

And if you follow AI research, there are big leaps each 3 months. Maybe not regarding model accuracy (these take somewhat longer) but regarding model efficiency, RAM consumption and inferencing speed. These are just not as prominent to the public like e.g. a GPT4 Release. But they will be very visible regarding the pricetag of using them.

And there are many, many useful things you can do outside of LLM. Very application specific DNN which yield in huge quality and/or performance gains.
 
Reactions: xpea

PJVol

Senior member
May 25, 2020
792
776
136
AI, to the extent it is being hyped as if it's the greatest thing since the invention of the wheel, and the sheer greed and deception involved in marketing anything with an "AI" sticker, is both a scam and a massively inflated bubble.
Regarding the topic, I absolutely don't mind if amd (or nvidia) slapping AI stickers literally all over the graphics card, and as long as this doesn't affect the price I'm fine.
 
Reactions: igor_kavinski

Keller_TT

Member
Jun 2, 2024
113
112
76
AI has real value, it's not going to pop like a bubble. It's going to soft land.
It is a bubble when it is unsustainably expensive and energy devouring, and oversold the way it is with a mythical halo, and its limits are soon apparent.
NV of course will more than survive when the market corrects itself, and it would be like dot-com burst for some, and just like being brought down to earth for NV and some others (I'm waiting for it to be the former for a very nefarious unOpen company).
Well, DeepSeek R1 has just recently shown a new avenue regarding model training, which for the majority relies on synthetic data. And synthetic data can be pushed much more easily towards "the scale of the universe".
The same applies for Chain of Thought. This one primarily scales with compute and not so much with additional data.

And if you follow AI research, there are big leaps each 3 months. Maybe not regarding model accuracy (these take somewhat longer) but regarding model efficiency, RAM consumption and inferencing speed. These are just not as prominent to the public like e.g. a GPT4 Release. But they will be very visible regarding the pricetag of using them.

And there are many, many useful things you can do outside of LLM. Very application specific DNN which yield in huge quality and/or performance gains.
I'm not talking just LLM at all. I mean the breadth of GenAI in general, which encompasses a whole lot more. Synthetic data has its limits. I'm talking about real-time, personalized AI, and a number of areas where there is a hard constraint on relevant and quality data.

And you're emphasizing what I just said. DeepSeek R1 has opened eyes in terms of computational efficiency. But that is specific to LLMs and they'll for sure evolve much more. The expectation is other GenAI areas too become much more efficient. That is through both hardware and software architectures, and algorithms. I'm involved with the Swiss AI industry-institute connect who focus on Client and Edge AI too, and a really broad base of what AI can bring without imposing itself. So, I know where the limits are, and the political implications that can't be overlooked.
 
Reactions: DAPUNISHER

basix

Member
Oct 4, 2024
40
75
51
And you're emphasizing what I just said. DeepSeek R1 has opened eyes in terms of computational efficiency. But that is specific to LLMs and they'll for sure evolve much more. The expectation is other GenAI areas too become much more efficient. That is through both hardware and software architectures, and algorithms. I'm involved with the Swiss AI industry-institute connect who focus on Client and Edge AI too, and a really broad base of what AI can bring without imposing itself. So, I know where the limits are, and the political implications that can't be overlooked.
Well, hi then to a fellow countryman.

I am working in engineering and AI has so many use cases and benefits without really compromising privacy. There it is mostly about physics & chemistry (which have no privacy) and coding (which is more a licensing topic regarding training data).
It really depends on the use case. Everything connected and related to people is much more nuanced and complicated (social and political implications as you say). Medical data is a good example for that. Potentially very useful, but with huge concerns on the other side.
And regarding leaps and breakthroughs it is always a question, how much money is spent and how many bright people are working on it. And for many very specific things there might be a breakthrough but you simply do not take notice of it (because not on your radar, so many of them).

I fully agree with you that AI as term is hugely inflated and overhyped. The same applies for respective stock market assets, Nvidia is just one of many.
Nevertheless: AI has very nice use cases and benefits. It will stay. It will get better. I said already 2017 or so that AI will be the next big thing. And here we are, much has changed. Much has improved. Much has still not been achieved yet (fully self-driving cars anyone?).
 
Reactions: igor_kavinski

Keller_TT

Member
Jun 2, 2024
113
112
76
Well, hi then to a fellow countryman.
Grüezi 👋
It will get better. I said already 2017 or so that AI will be the next big thing. And here we are, much has changed. Much has improved. Much has still not been achieved yet (fully self-driving cars anyone?).
Well, that's a whole new can of worms with "fully self-driving cars". Apart from feasibility in cities, I know there are a number of legal hurdles about liabilities and the rules based on which self-driving cars might decide life and death decisions.
That belongs to the application of deontic logic, and I learned about some truly fascinating research into it from a source that would blow your mind: A particular branch of exegesis of ancient Indian sacred compositions called Mimamsa and its sub-school that went to absurd lengths on linguistics and determining injunctions and morality with which it had to be approached. The Indians kicked ass when it came to linguistics and analytic philosophy, and I know there's a paper that's popular among Indians in the AI field from an old NASA scientist about how Sanskrit was the world's first generative grammar that was specifically adapted for knowledge representation, and has a body of literature that far exceeds that of Greek and Latin combined.
Prof. Agata Ciabattoni of TU Wien has gone deep into that area with applicability to self-driving cars here:
I attended an AI workshop there that was fully organized around this, and we had some truly astounding discussions.
 
Reactions: moinmoin and basix

beginner99

Diamond Member
Jun 2, 2009
5,281
1,695
136
GenAI can be optimized for efficiency but in terms of big leaps in performance, we're well into diminishing returns without being able to gather and train data at the scale of the universe and obliterating privacy
I'm firmly in the "AutoGPT" camp. the advantage of what was developed is to be able (not always but pretty well) understand what the user is asking / demanding. What is less useful is what is being done with that capability.

The productivity boost will come when AI can do automation on text prompt / voice prompt. Like at work the ordering process is a complete and utter shit show. The people that have to use it regularly tend to need like 10 minutes per simple order. because it's slow and often fails and has to be tried multiple times. Imagine you can just tell an AI to order for you in the internal system. This of course is not plug and play anymore and will require software development so not so easy and cheap to implement but that would actual give productivity benefits.
 
Reactions: coercitiv

RnR_au

Platinum Member
Jun 6, 2021
2,204
5,271
106
Nvidia will be more than fine, they're selling the shovels.
Nvidia is selling the shovels most folk are using in training the models. Anyone can build shovels for inference. The former as a market will be much smaller than the latter once inference really takes hold in the economy and its becomes normalised across all industries.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |