Discussion Zen 5 Speculation (EPYC Turin and Strix Point/Granite Ridge - Ryzen 9000)

Page 710 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

poke01

Platinum Member
Mar 8, 2022
2,004
2,542
106
The better question if you are going solder RAM, you could save board space and have bigger busses if you use on-package but what are the disadvantages of soldering memory to the board vs MoP?

I can’t think of any for STX halo.
 
Reactions: FlameTail

adroc_thurston

Diamond Member
Jul 2, 2023
3,492
5,055
96
I am a straight up bubble huffer. I say that AI functions will be the main reason for sales of computing devices by 2040. I still find it bizarre there’s so many people on various text site forums don’t see that AI is THE future of computing in society.
matrix math isn't all that useful.
I believe the next 30 years will yield a bigger technological and social transformation than than the period from the dawn of civilization to 2024.
xir we're launching TLAMs at Taiwan in a few years.
You sure?
I can’t think of any for STX halo.
MoP means SKU spam.
Suboptimal for a new swimlane new part new-new-new.
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,912
3,524
136
I am a straight up bubble huffer. I say that AI functions will be the main reason for sales of computing devices by 2040. I still find it bizarre there’s so many people on various text site forums don’t see that AI is THE future of computing in society.

To be sure, new algorithms are needed. The current ones are just stopgaps, as impressive as they may be at these small simple tasks. But, just as current computers are many orders of magnitude more complex and capable than a Commodore 64, neural networks are going to follow the same trajectory, at about triple the speed.

Research is already underway to boost reasoning capacity by 100 to 1000 fold. Those algorithms will be here in a few short years, ones that will make large language models feel like stone age technology. Critically, those won’t be the last ones.

I believe the next 30 years will yield a bigger technological and social transformation than than the period from the dawn of civilization to 2024.

Yeah, I’m huffing big time.
this shows a complete lack of understanding about the hardware your algorithms are running on and the evolution of hardware and what the future looks like.

Your Computations are already running on hyper targeted and optimised hardware that has the biggest BOM's out side of mainframes we have ever seen. Also who exactly is making any money from AI outside Nvida? Im watching this daily as i want to move before the A** falls out of the market.
 

The Hardcard

Member
Oct 19, 2021
199
288
106
matrix math isn't all that useful.

and yet is the main focus of every major chipmaker on the planet, as well as nearly all of the minor ones. From CEOs, CTOs, PhD fellows all the way down to interns, it is the most important aspect of every architecture and chip at every stage from concept through design, production and validation. Hmmmmmmmm.

xir we're launching TLAMs at Taiwan in a few years.
You sure?
Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.

this shows a complete lack of understanding about the hardware your algorithms are running on and the evolution of hardware and what the future looks like.

Your Computations are already running on hyper targeted and optimised hardware that has the biggest BOM's out side of mainframes we have ever seen. Also who exactly is making any money from AI outside Nvida? Im watching this daily as i want to move before the A** falls out of the market.
Well obviously, I think the lack of understanding is fully in your camp.

Llama 3 405B has for the last week been used and tested by thousands of people who have been heavily using LLMs for two to three years now. So far it consistently demonstrates effectively equal capabilities as ChatGPT 4o. And it will run on 2 Macbook Pros connected with a single Thunderbolt cable.

More importantly, LLMs are not the AI revolution. They are the precursor.
 

adroc_thurston

Diamond Member
Jul 2, 2023
3,492
5,055
96
and yet is the main focus of every major chipmaker on the planet, as well as nearly all of the minor ones. From CEOs, CTOs, PhD fellows all the way down to interns, it is the most important aspect of every architecture and chip at every stage from concept through design, production and validation. Hmmmmmmmm.
First time?
Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.
That's not how it works.
Gigafabs are Taiwan-only and R&D expertise is also pretty much non-transferrable.
 
Reactions: Joe NYC and maddie

csbin

Senior member
Feb 4, 2013
884
529
136
oh, nooooooooo



Translated added -

As to why the equivalent delay is 1 and not 0.5, this is a major problem I'm having at the moment.The current version of the microcode seems that a single thread cannot see two decoders no matter what, that is, after the op$ is released or turned off, the front end directly becomes 4-wide and can only take 1 per cycle (regardless of whether there is a branch jump or not). This is obviously inconsistent with AMD's propaganda that a single thread can use two decoders, and more investigation is needed.
Mod DAPUNISHER
 

branch_suggestion

Senior member
Aug 4, 2023
373
831
96
>RDNA double-issue flashbacks.
Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.
As it turns out, the software is lagging the hardware and holding it back, as is tradition for AMD.
Whether there is a new AGESA in the next week or so that magically enables the core to use dual decoders for compatible 1t workloads, well that would make any review delay justified.
Yes I am coping.
 

misuspita

Senior member
Jul 15, 2006
497
591
136
Last edited:
Reactions: Gideon

mostwanted002

Member
Jun 16, 2023
33
66
61
mostwanted002.page
Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.
As it turns out, the software is lagging the hardware and holding it back, as is tradition for AMD.
Whether there is a new AGESA in the next week or so that magically enables the core to use dual decoders for compatible 1t workloads, well that would make any review delay justified.
Yes I am coping.
^we are coping.

:copiumhuff:
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,912
3,524
136
and yet is the main focus of every major chipmaker on the planet, as well as nearly all of the minor ones. From CEOs, CTOs, PhD fellows all the way down to interns, it is the most important aspect of every architecture and chip at every stage from concept through design, production and validation. Hmmmmmmmm.


Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.


Well obviously, I think the lack of understanding is fully in your camp.

Llama 3 405B has for the last week been used and tested by thousands of people who have been heavily using LLMs for two to three years now. So far it consistently demonstrates effectively equal capabilities as ChatGPT 4o. And it will run on 2 Macbook Pros connected with a single Thunderbolt cable.

More importantly, LLMs are not the AI revolution. They are the precursor.
So one model that cant do anything useful is more efficient then another model that cant do anything useful....

how many more 10x's of everything do we need to get somewhere useful.

We kind of had 10x scaling in hardware moving from CPU SIMD to MIMD to GEMM engines to hardware aware sparsity, memory compression etc etc. Now we have massive clusters with chips at retile limits that are limited by the speed of light across derivatives of Clos fabrics In a word where CMOS scaling is dead.

I've seen one good use case of AI that works today in my field (high-end tech) and its not replacing a single job or driving large efficiencies. It will just give better situational awareness during failure conditions.
 

gdansk

Platinum Member
Feb 8, 2011
2,843
4,239
136
Somehow 9% improvement in INT 1T doesn't sound as bad if it's still (apparently) 4-wide decode.
Didn't think they could get any more out of that.
 

MS_AT

Senior member
Jul 15, 2024
210
504
96
Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.
As it turns out, the software is lagging the hardware and holding it back, as is tradition for AMD.
Whether there is a new AGESA in the next week or so that magically enables the core to use dual decoders for compatible 1t workloads, well that would make any review delay justified.
Yes I am coping.
I am afraid it's about bad marketing message and miscommunication. The materials were mentioning that decoders are statically partitioned in SMT mode. Now traditionally when you wanted to turn off SMT, you went to BIOS and disabled it. Now the question is, is the SMT mode static when enabled [If SMT is on in the BIOS is the core always in SMT mode] or is it dynamic like the interviews are leading us to believe.
 

CouncilorIrissa

Senior member
Jul 28, 2023
521
2,002
96

MS_AT

Senior member
Jul 15, 2024
210
504
96
🤔 It seems that the AMD employee who submitted the patch knew it was 4-wide all along.
We were bamboozled by Mike Clark yet again.
It might be other things too. These were early patches, they might have wanted just to get znver5 option added that would not dramatically break the situation for people using --march=native rather than to give accurate representation of the core. They might have wanted not to share everything or they might have forgot to update. Don't forget that CPUs are designed to handle less than ideal code [OoO, branch prediction etc.] so this won't have terrible effect. It would be much worse if somebody forgot to turn on all available instructions sets as that would hamper the generated code more.
 
Reactions: Nothingness

tsamolotoff

Member
May 19, 2019
176
306
136
To be sure, new algorithms are needed.
The real AI has never been tried, just wait 2-3 weeks (insert 💲💲 here). Well, seriously, what kind of use the current transformers have apart from replacing the politicians (as they can lie and get even more delusional than the most flamboyant political figures around the world). And also, why do we need it integrated into general-purpose CPUs and GPUs at all, it could be just a separate addon board or card and we won't need to sacrifice 16 mb of cache and employ a weird dual ccd setup for this deadweight silicon
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |