Chiplet confirmed, nice job everyone who made that prediction.
TL;DR explanation of why this is a big deal?
Chiplet confirmed, nice job everyone who made that prediction.
TL;DR explanation of why this is a big deal?
The next Ryzen chips almost certainly use the same 7nm chiplets, just with an IO die aimed for desktop (possibly with GPU).
Agreed. The extra die space to integrate all of the necessary IO for the Ryzen SKUs is worth it instead of making another smaller IO die just to have a chiplet approach with Ryzen.I doubt that.
While chiplets look like a good tradeoff for TR/Epyc that already have multi-chip latency issues, and are aimed at markets where those latencies don't really matter, Ryzen is a desktop part, where latency does matter and there are currently no multi-chip latency penalties.
A chiplet design for Ryzen is really not a great choice. AMD should, and hopefully feels they have enough market, and enough money to do a separate monolithic die for desktop.
Or maybe they desktop Part will only be the APU this time, but with 8 Cores.
That IO die looks fairly huge.
Might be from that fat L4 cache on there . . .
It's a dramatically different approach. There is a standard way how the system architecture is laid out that AMD has used since 2003 and Intel since 2008, and this is a major change from it.
This new way of doing things has both advantages and disadvantages. Some of those we can guess, others will only show themselves in use. It's very interesting to speculate on them.
- For server customers, this means no more 4 numa nodes per socket. This will really help some important loads. Interestingly, since Intel just released Cascade Lake-AP, this means a complete role reversal. Except that AMD still has more cores per socket.
- Since they use infinity fabric connecting the CPU to the IO system, it means they can iterate on them independently. Especially with the rise of lower cost-to-iterate processes that are better for IO (like GF 22nm and 12nm), this might mean a lot more products targeting different workloads, all using the same chiplets for compute, but with a different IO die. The next Ryzen chips almost certainly use the same 7nm chiplets, just with an IO die aimed for desktop (possibly with GPU).
- Since they can iterate independently, they can now bring products to market for different IO standards without having to ship a new CPU die. I think this means AMD will be first out with DDR5 support, since they can just design a new IO die for that, and use the same chiplets for both AM4 and AM5 socket processors. Same goes with next-gen PCI-E.
A lot of the speculation went well
Not sure if this a 64 core part.
But at least we are on the right track
That IO die looks fairly huge.
- 14nm IO Chiplet
- 7nm Chiplets
- MC on the IO die
- 8 cores per Zen 2 die (not sure if per CCX)
- 2x fp compared Zen1
- Improved front end (also according to Agner, Zen1 backend is bottlenecked by the front end)
- 33% better memory bandwidth (due to higher memory speed?)
I have to say a lot of those improvements they showed today left a patent trail if you have the patience to dig through.
I doubt that.
While chiplets look like a good tradeoff for TR/Epyc that already have multi-chip latency issues, and are aimed at markets where those latencies don't really matter, Ryzen is a desktop part, where latency does matter and there are currently no multi-chip latency penalties.
A chiplet design for Ryzen is really not a great choice.
8 Core CCX confirmed
Could you please explain the mechanics of why you think latency will get markedly worse?
Whats that?
Looks more like Naples than anything else...?
Not every solution makes sense everywhere. Chiplets are good for TR/EPYC, not so good for desktop.
Integrated MC as opposed to Memory controller on another chip. There is a latency penalty for going off chip.
For a small desktop chip, chiplets should have worse latency and be less cost effective.
Maybe it's just how the thing is packaged? Doubled Load/store could also mean 64B/cycle up from 32B/cycle.WOW so i was wrong....rofl.
did anyone else notice it looks like the chiplets are in pairs of two on a interposer???
edit: also we need conformation about the load store, "12:53PM EST - Improved pipeline, DOuble loading point and load store"
i hope this means 4 load and 2 store ports, but im gussing it just means 2/1 256bit load/store ports
What does this mean Doubles the BW/Channel for the same number of channels?
Increased frequency or just efficient usage?
Maybe it's just how the thing is packaged? Doubled Load/store could also mean 64B/cycle up from 32B/cycle.
EDIT: Are we both saying the same thing when you mention 4 load and 2 store ports and 64B/cycle?
Ah, I was thinking of cache b/w. I suppose it might have been increased as well, given the increased FP performance.see what i want is 4x128 load and 2x128 store a cycle
what i think it will be is 2x256 and 1x256 store a cycle
What Ian wrote in the live steam could be interpreted both ways (to me anyway).
What does this mean Doubles the BW/Channel for the same number of channels?
Increased frequency or just efficient usage?