Question Zen 6 Speculation Thread

Page 93 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

adroc_thurston

Diamond Member
Jul 2, 2023
5,437
7,608
96
I wonder what the packaging technology will be for Venice.
There's a whole mix there.
Very likely, different from client, which means sharing of CCDs will no longer be taking place.
Kinda?
MLID said that on future roadmaps, AMD is not differentiating between classic and dense cores for servers.
sure as hell they are lol.
From the press release (and the tweet), it mentions the chip was "brought up" which I imply being back at AMD and powered on.
Indeed.
SF 4nm IOd seems likely.
No? No. TSMC is the only foundry with good analog.
I wonder how many IOd and CCd variants there are planned then.
CCDs not many, other things yeah.
 
Reactions: marees

Saylick

Diamond Member
Sep 10, 2012
3,874
8,985
136
From the press release (and the tweet), it mentions the chip was "brought up" which I imply being back at AMD and powered on.

So, perhaps, 1 year of debugging, validation and hopefully H1 or mid 2026 launch.
It's targeting Q3 2026, I believe.
 
Reactions: Joe NYC

dr1337

Senior member
May 25, 2020
454
733
136
OEMs are pushing AI because Wall Street orgasms at the mere thought of it. Doesn't mean the OEM's actual customers are interested.
Its really the other way around, millions of people are using GPT and other commercially/closed models. Where as most LLM questions/and even lots of image stuff will run just fine on a 16gb GPU from 2020 using distilled models. Meanwhile look at apple AI, they critically depend on GPT simply because they don't have any models or public facing tech of their own.

So in this world where the real OEMs are HP, Dell, Lenovo, ect. And their customers have a pressing need for high bandwidth matrix calculations, and this job can either be done by a giant GPU chip using special GDDR ram, or you can put an NPU into a SOC, save tons of power, use the same ram chips already on the mobo, and still get good enough performance. And if everyone has AI on their machine, it really makes it easy for developers to make AI based apps and programs they can sell to people. This is both microsoft and apple's angle.

Right now it's awful for me to try and run AI on my non-accelerated CPU and I fully have to commit my GPU and lots of power for good performance. Where as NPUs on CPUs really open up the doors to things like games using AI for npcs, or in digital media where the NPU can be used to free up compute for things like effects and transitions. For development it really blows the hinges because it means running one AI into another, locally, suddenly becomes feasible and thats where things really shine. I for one really enjoy using different system prompts on the same model to get more varied results. ie, one programmer bot thinks its a webdev, the other I tell its john carmack, now give me code for X. Would be super nice to get both results at the same time for comparison.

Also look at the GPU market lately, it's not like there's a crypto rush right now, those sales and scalping is legit pressure from the AI market. And if AMD can leverage even a tiny bit of that onto desktop CPU, thats becomes a win for potential volume and margins.
 
Reactions: OneEng2

MS_AT

Senior member
Jul 15, 2024
570
1,203
96
NPU into a SOC, save tons of power, use the same ram chips already on the mobo, and still get good enough performance.
NPU is not magical, it will steal MemBW from the system. So your concurrent use cases example is not so rosy. And AMD in their lemonade server is sticking to using NPU on strix point for prompt processing and gpu for token generation citing power efficiency
I for one really enjoy using different system prompts on the same model to get more varied results. ie, one programmer bot thinks its a webdev, the other I tell its john carmack, now give me code for X. Would be super nice to get both results at the same time for comparison.
Hmm, that should be already doable, just use a batched mode, batch two inputs together, it will actually improve the compute to memBW needs. Unless you want to use 2 different models.
 
Reactions: Joe NYC

dr1337

Senior member
May 25, 2020
454
733
136
To make memes. Which is fun but silly that it's driving the entire financial system right now.
Don't companies like adobe and apple make a non-insignificant amount of their money on selling new software? Its really not just people on social media using AI, there's even larger streamers that have an LLM+TTS generator respond to random chatter comments and everyone loves it. Content aware fill has been a thing since like 2015 and people in commercial media still use it all the time.
Hmm, that should be already doable, just use a batched mode, batch two inputs together, it will actually improve the compute to memBW needs. Unless you want to use 2 different models.
Takes the same amount of time, you only get so many tokens/iterations per second regardless of how you organize the workload. Idk unless like you're running a bunch of tiny models on a huge accelerator but that's not what I'm talking about. Its much better to have models be as big as possible and thats why you don't see anyone commercially trying to run/sell swarms of 4gb distills.
NPU is not magical, it will steal MemBW from the system.
There are instances where its better when the entire system can fully share resources instead of one component dependent on another. The main bottleneck really is the acceleration of the FP instructions in this case. You can even see this in Apple benchmarks, they do have tons of memory bandwidth esp compared to GPUs but the tokens per second isn't on the same level. SOC type NPUs aren't as powerful as the huge cards/OAM designs and that's really all just down to transistor value. But are they are way more efficient than current CPU instruction sets and that makes them desirable.

And in this day in age, the difference between pcie3 and pcie5 is so minimal for GPUs and other devices even like the 5090, that an SOC class NPU wouldn't make a difference for system BW 99.8% of the time.
 
Last edited:

MS_AT

Senior member
Jul 15, 2024
570
1,203
96
Takes the same amount of time, you only get so many tokens/iterations per second regardless of how you organize the workload. Idk unless like you're running a bunch of tiny models on a huge accelerator but that's not what I'm talking about. Its much better to have models be as big as possible and thats why you don't see anyone commercially trying to run/sell swarms of 4gb distills.
Yes, it takes the same amount of time with caveats, but you get your
Would be super nice to get both results at the same time for comparison.
I wasn't talking about the absolute time, but if you have a model that fits in your GPU or CPU if you have good membW, then you can get that almost for free.
You can even see this in Apple benchmarks, they do have tons of memory bandwidth esp compared to GPUs but the tokens per second isn't on the same level.
Which tokens per second. The generation usually is matching the available memory bandwidth, the prompt processing sucks due to lack of compute. It also matters at what kind of context size you compare. All popular outlets I have seen that started using LLMs in benchmarks fail to mention this.
But are they are way more efficient than current CPU instruction sets and that makes them desirable.
I thought we were comparing to GPUs
And in this day in age, the difference between pcie3 and pcie5 is so minimal for GPUs and other devices even like the 5090, that an SOC class NPU wouldn't make a difference for system BW 99.8% of the time.
The difference is not minimal, but the BW difference between PCIe5 and GDDR is so big, that everyone tries to keep the workload on the GPU to avoid transfers. And if you wanted to run inference on NPU fully it would need to have access to full memBW. In other words, NPUs make sense only on chips which have already high memBW, as using NPU on a desktop socket might give you better time to first token but won't give you noticeable increase in token generation over what CPU alone could do. And provided you have dGPU in the system that has enough memory to fit the model, NPU is pointless as offloading prompt processing will run into PCIe 5 bottleneck, unless the model is small enough, but if it is small enough, you can run it on GPU

On mobile SoC the situation in small 14 inch models where you cannot fit a dGPU, then things look more favourable for the NPU, especially since you get LPDDR with higher bandwidth to capacity ratio that what you can achieve on desktop sockets and all functional units share the same physical memory.
 

eek2121

Diamond Member
Aug 2, 2005
3,300
4,843
136
Its really the other way around, millions of people are using GPT and other commercially/closed models. Where as most LLM questions/and even lots of image stuff will run just fine on a 16gb GPU from 2020 using distilled models. Meanwhile look at apple AI, they critically depend on GPT simply because they don't have any models or public facing tech of their own.

So in this world where the real OEMs are HP, Dell, Lenovo, ect. And their customers have a pressing need for high bandwidth matrix calculations, and this job can either be done by a giant GPU chip using special GDDR ram, or you can put an NPU into a SOC, save tons of power, use the same ram chips already on the mobo, and still get good enough performance. And if everyone has AI on their machine, it really makes it easy for developers to make AI based apps and programs they can sell to people. This is both microsoft and apple's angle.

Right now it's awful for me to try and run AI on my non-accelerated CPU and I fully have to commit my GPU and lots of power for good performance. Where as NPUs on CPUs really open up the doors to things like games using AI for npcs, or in digital media where the NPU can be used to free up compute for things like effects and transitions. For development it really blows the hinges because it means running one AI into another, locally, suddenly becomes feasible and thats where things really shine. I for one really enjoy using different system prompts on the same model to get more varied results. ie, one programmer bot thinks its a webdev, the other I tell its john carmack, now give me code for X. Would be super nice to get both results at the same time for comparison.

Also look at the GPU market lately, it's not like there's a crypto rush right now, those sales and scalping is legit pressure from the AI market. And if AMD can leverage even a tiny bit of that onto desktop CPU, thats becomes a win for potential volume and margins.
If you think millions of people are using LLMs and paying for the privilege to do it, I have a cheap bridge to sell you.

The highest paid adoption has been in the corporate sector, and that is not even close to “millions”.

Microsoft began forcing the bundling of their AI product with Office365 because they weren’t seeing the additional revenue they were hoping for. They also raised prices of Office365 because of this.
Don't companies like adobe and apple make a non-significant amount of their money on selling new software?
Adobe rents software, they don’t sell it. Apple only has a few paid products, they rent the rest.
 

eek2121

Diamond Member
Aug 2, 2005
3,300
4,843
136
They are paying for it by giving out priceless information, far more than search engines were getting
Not disagreeing with that. 🤣

That info doesn’t really translate easily to tons of revenue, however, and large LLMs are expensive to develop and run.
 

dr1337

Senior member
May 25, 2020
454
733
136
Adobe rents software, they don’t sell it. Apple only has a few paid products, they rent the rest.
I did mistype a bit there but I think my point still stands, its their interest to have good creative software. Just because they rent doesn't mean photoshop or final cut is free lol.
If you think millions of people are using LLMs and paying for the privilege to do it, I have a cheap bridge to sell you.

The highest paid adoption has been in the corporate sector, and that is not even close to “millions”.
Is that really what you thought I meant? Compared to the nearly 1bn people using openai alone as a huge indicator for the demand of AI? Here in 2025 I'd really think most people would know no product is truly free either.

Heck at this rate there's no way someone isn't training an LLM off these forums and crawling every major site right now, would explain the crazy lag we get sometimes too.
 
Last edited:
Reactions: igor_kavinski

Nothingness

Diamond Member
Jul 3, 2013
3,283
2,337
136
If you think millions of people are using LLMs and paying for the privilege to do it, I have a cheap bridge to sell you.
ChatGPT has added 4.5 million paying subscribers since the end of last year, according to The Information. With 20 million subscribers paying for the service, the company is earning “at least $415 million in revenue per month,” a rough estimate that doesn’t account for corporate plans or the new $200 a month Pro tier. The report adds that this could put OpenAI “well within reach” of its revenue projection of $12.7 billion this year.

Can I get my bridge?

Of course it doesn't mean they are profitable, but yes they have millions of paid subscribers on top of corporate licenses.
 
Reactions: dr1337

madtronik

Junior Member
Jul 22, 2019
9
20
81
Yeah they weren't.

The point is simple: ML slop farms do not have a business model to afford all that big iron running 24/7.
Simple enough for you?
Just a note. AFAIK YouTube has NEVER been a business because from its inception it has always lost money. Nowadays doesn't lose so much money as before due to Google making all they can do to break even. It was an objective to break even some years ago but they just gave up and didn't talk more about it. They just talk about YouTube revenues but never about profits.

Count that with IA could happen something similar. Burn tons of money from other profitable business in order to get as much market as possible and in 1-2 decades to try to break even.
 
Reactions: marees

fastandfurious6

Senior member
Jun 1, 2024
476
619
96
we are in the era of peak accelerationism

true AI is the holy grail of computing and whoever thinks the opposite is deluded

LLM is a brute-force way towards it but it works. it can already replace two-digit % of thought work / office workforce and its capabilities increase exponentially. entire industries job markets are getting wrecked just because of LLMs

copilot stuff is just marketing/ambitious nonsense, but real fat LLMs can already replace real workers. all the layoffs of last years are largely codriven by LLM impact, also accelerates new projects, new products, new architectures etc

ontopic, my speculation, without LLMs both TSMC N2 and Zen6 would be very delayed for more years
 
Reactions: HurleyBird

adroc_thurston

Diamond Member
Jul 2, 2023
5,437
7,608
96
AFAIK YouTube has NEVER been a business because from its inception it has always lost money.
It's an accretive play for GOOG's advertisement business plus it's far, far cheaper than anything ML.
It was an objective to break even some years ago but they just gave up and didn't talk more about it.
Pretty sure they did by making end user experience worse.
Count that with IA could happen something similar.
ML big iron is far, far more expensive to buy and operate.
 

OneEng2

Senior member
Sep 19, 2022
473
698
106
And those sweet, sweet margins selling highest for the lowest parts they can get away with. Mmm, margins.
... and that is what the board members all call a good investment. Intel needs to think on this a while IMO.... preferably before they go under.
Not sure if this needed any confirmation but Venice is on N2, officially https://ir.amd.com/news-events/pres...ieves-first-tsmc-n2-product-silicon-milestone
I saw that as well. Thanks.

I really never had any doubt that Zen 6 EPYC would be on N2, but it is nice to see it verified. At a minimum it means AMD has some good information saying that N2 is on track to be a high volume node with good performance. It would have to be as N3E did quite well for Turin D.
As far as Venice, I also wonder if the classic Zen 6 CCD will be 12 or 16 cores. Then, Dense likely at 32 cores.

Although, MLID said that on future roadmaps, AMD is not differentiating between classic and dense cores for servers.
I suspect that both Turin Classic will be 32 full Zen 6 cores (Turin D on N3E was 16. Doing a little math on the N2 shrink and it isn't hard to see a 32 core CCD). Perhaps Turin D might be 48 cores of Zen 6e? Oh well, speculation is fun at this point .... and about all we have until H2 2026 .
That is what I want to know most how many cores on a single CCD and if there will be an increase in 1:1 speed so higher speed DDR5 can be used.
I read this: https://wccftech.com/amd-zen-6-base...ure-two-imcs-changes-ddr5-memory-orientation/

I wonder if AMD might create some hybrid setup where one CCD was 16c Full Zen 6 while the other CCD was 32c Zen 6c? They may well need something like this to compete with Intel's 52 core Nova Lake. Feeding it with 4 channels of memory may have been a needed change with so many high performance cores with SMT. (48c 96t). That would be a monster .
None of that has anything remotely resembling a viable business model.
Seriously? Are you one of those guys that said digital photography would never pan out? Oh, Oh, and how about that crazy digital music thing? Pretty dumb eh?

Anyone betting against AI right now is seriously delusional IMO.
Can I get my bridge?

Of course it doesn't mean they are profitable, but yes they have millions of paid subscribers on top of corporate licenses.
I want a bridge too .
You mean AMD were not a business while they were leaking money? Waiting for your next stupid 3 words statement. Will that be some 40% silly claim?
Exactly. There you go. I one-upped him with a one-word-answer .
we are in the era of peak accelerationism

true AI is the holy grail of computing and whoever thinks the opposite is deluded
Agree.
 

eek2121

Diamond Member
Aug 2, 2005
3,300
4,843
136


Can I get my bridge?

Of course it doesn't mean they are profitable, but yes they have millions of paid subscribers on top of corporate licenses.
I intentionally left this one alone. First, corporate is included in those paid subscribers, but more importantly, until they become a public company, none of what they claim matters.

Even more importantly then that: it is costing them FAR MORE per subscriber than they are charging to the user.

This is actually my mistake however, when I shot down the millions number, I'm thinking like actual millions. like 20 million or something. At 20 million users with their current pricing and without severe limits, they'd already be out of business.

I've said AI is a bubble, I don't know if I said it here, however I've said it and since I know some folks misunderstand me, let me tell you why I think it is actually a bubble: when it comes to software, top developers are a dime a dozen. There are no patents either. Some companies MAY patent specific algorithms, but it rarely gets anywhere. That means that there is NO WAY to keep competition of a popular area from popping up. In certain cases this won't matter, but every big tech industry went in with billions, so it does matter here. It has already been shown that a properly optimized LLM can match/exceed models outed by OpenAI, Google, Facebook, Amazon, etc. They also don't require a subscription in many cases. It is a zero sum game, and there is no way, for example, for Open AI to justify a subscription to their service over a cheaper service or running a model locally. They aren't even the top in the field, and they missed the whole founder/dominance thing.

LLMs themselves are another topic entirely, the economics of these companies, however, isn't adding up. They all rely on subscriptions, and as Deepseek has pointed out (others did as well, they were just the first big backing of such a thing), a subscription isn't needed. Just a semi decent PC. There is also some other stuff coming out (hardware/software both) between now and next year that will be changing the game entirely...and it also won't require remote cloud computing. When you add in the fact that the US (which has the biggest tech industry in the world) is going into a recession, it really does not look good for AI startups, or even big tech.

I'm not hating on LLMs btw, they actually do have uses, especially if the tech gets better, however none of those uses will warrant a subscription fee to a billion/trillion dollar company when anyone can replicate your work without looking and do better.

Just my thoughts. Take from it what you will.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |