Discussion Intel Meteor, Arrow, Lunar & Panther Lakes Discussion Threads

Page 656 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Tigerick

Senior member
Apr 1, 2022
702
632
106






As Hot Chips 34 starting this week, Intel will unveil technical information of upcoming Meteor Lake (MTL) and Arrow Lake (ARL), new generation platform after Raptor Lake. Both MTL and ARL represent new direction which Intel will move to multiple chiplets and combine as one SoC platform.

MTL also represents new compute tile that based on Intel 4 process which is based on EUV lithography, a first from Intel. Intel expects to ship MTL mobile SoC in 2023.

ARL will come after MTL so Intel should be shipping it in 2024, that is what Intel roadmap is telling us. ARL compute tile will be manufactured by Intel 20A process, a first from Intel to use GAA transistors called RibbonFET.



Comparison of upcoming Intel's U-series CPU: Core Ultra 100U, Lunar Lake and Panther Lake

ModelCode-NameDateTDPNodeTilesMain TileCPULP E-CoreLLCGPUXe-cores
Core Ultra 100UMeteor LakeQ4 202315 - 57 WIntel 4 + N5 + N64tCPU2P + 8E212 MBIntel Graphics4
?Lunar LakeQ4 202417 - 30 WN3B + N62CPU + GPU & IMC4P + 4E012 MBArc8
?Panther LakeQ1 2026 ??Intel 18A + N3E3CPU + MC4P + 8E4?Arc12



Comparison of die size of Each Tile of Meteor Lake, Arrow Lake, Lunar Lake and Panther Lake

Meteor LakeArrow Lake (N3B)Lunar LakePanther Lake
PlatformMobile H/U OnlyDesktop & Mobile H&HXMobile U OnlyMobile H
Process NodeIntel 4TSMC N3BTSMC N3BIntel 18A
DateQ4 2023Desktop-Q4-2024
H&HX-Q1-2025
Q4 2024Q1 2026 ?
Full Die6P + 8P8P + 16E4P + 4E4P + 8E
LLC24 MB36 MB ?12 MB?
tCPU66.48
tGPU44.45
SoC96.77
IOE44.45
Total252.15



Intel Core Ultra 100 - Meteor Lake



As mentioned by Tomshardware, TSMC will manufacture the I/O, SoC, and GPU tiles. That means Intel will manufacture only the CPU and Foveros tiles. (Notably, Intel calls the I/O tile an 'I/O Expander,' hence the IOE moniker.)



 

Attachments

  • PantherLake.png
    283.5 KB · Views: 24,014
  • LNL.png
    881.8 KB · Views: 25,501
Last edited:

Meteor Late

Member
Dec 15, 2023
116
98
61
Its on a better node so thats kind of given to be smaller. The performance improvements are pretty mundane as well.

The laptop version is always faster, so 8 Elite chip is going to be around 20% faster than X elite in SC if 8 elite is 10% faster, usually there is around a 10% increase in frequency from Apple smartphone chips to M chips.
 

SiliconFly

Golden Member
Mar 10, 2023
1,651
996
96
Lion Cove is a big problem for Intel, in how bloated it is.

After density improvement of N3 over N4 (up to 40% in logic), after removal of HT, after removal of AVX-512, it is still bigger than Zen 5 core.
P core team is ahead in only one thing

Wasting Silicon Area
My estimate puts Lion Cove P core around 25% larger than an equivalent Zen 5 P core (iso-node & without L2). And thats horrible considering there is no extra performance to justify the extra silicon. Lion Cove not just sucks... it sucks on a whole new level! It's fatter & slower than the other two.

Some napkin math (if all of them were on N3B):
  • Lion Cove - 3.4 mm2
  • Zen 5 P core - 2.74 mm2
  • Apple M4 P core - 2.8 mm2
Another interesting observation is, Zen 5 P core appears to be slightly smaller than Apple M4 P core (given the same node & without L2).
 

Meteor Late

Member
Dec 15, 2023
116
98
61
My estimate puts Lion Cove P core around 25% larger than an equivalent Zen 5 P core (iso-node & without L2). And thats horrible considering there is no extra performance to justify the extra silicon. Lion Cove not just sucks... it sucks on a whole new level! It's fatter & slower than the other two.

Some napkin math (if all of them were on N3B):
  • Lion Cove - 3.4 mm2
  • Zen 5 P core - 2.74 mm2
  • Apple M4 P core - 2.8 mm2
Another interesting observation is, Zen 5 P core appears to be slightly smaller than Apple M4 P core (given the same node & without L2).
Well yeah but M4 P core is much better than Zen 5 core and consumes much less power, so it makes sense.

An interesting comparison will be Oryon P core vs Zen 5 core, that core is barely above 2 mm2 IIRC.
 
Last edited:

511

Golden Member
Jul 12, 2024
1,034
894
106
Well yeah but M4 P core is much better than Zen 5 core and consumes much less power, so it makes sense.

An interesting comparison will be Oryon P core vs Zen 5 core, that core is barely above 2 mm2 IIRC.
Zen 5 dedicates lots of area towards AVX-512 so it's not fair for Zen 5 the Integer Performance might lack but not FP/SIMD
 

SiliconFly

Golden Member
Mar 10, 2023
1,651
996
96
Raichu is saying there is a change to FE for Darkmont and will result in 3-5% ipc improvements.
👇
Even if Darkmont has a mid single digit IPC increase and a slight bump in clocks, it should have ~80% the performance of a Lion Cove P core. Arctic Wolf should have higher IPC than a Lion Cove P core (but lower clocks). But even with lower clocks, it should have performance comparable to a Lion Cove P core.
 
Reactions: Tlh97

OneEng2

Senior member
Sep 19, 2022
259
356
106
Wow. You guys jumped all over my discussion vector. This is exactly where I was going with the original question. Awesome information in this thread, and very interesting consequences as well.

So the AMD is able to be very competitive across many markets with a design on a less dense, less expensive node. This makes Zen5 design very impressive in my book and Lion Cove ..... just sad guys. WTH? How is it even possible for Lion Cove to be this bad in comparison?
SoC area comparison (Measurements my own).

View attachment 110839

Notes
- Lunar Lake and M4 are on 3nm, whereas X Elite and Strix Point are on 4nm. So areas are not directly comparable between them.
- All numbers are in mm²
- Cores area with asterisks (*) include the private L2 cache
- Lunar Lake SoC area is the N3B Compute Tile
- Apple M4 NPU area is suspiciously small, but I have double checked with their iPhone SoCs, and they also have ~5 mm² NPUs

Sources
-
Lunar Lake and Strix Point die shot annotations by Nemez
- M4 die shot annotation by Frederic Orange
- X Elite dieshot annotation by Piglin
Great summary! This is what I was gearing up to do with my original question! Great chart.
True. That makes AMD's core super impressive and its on N4P. Zen6 on N3E/P is gonna be great to look forward too.
It definitely gives some much needed context to many of our design discussions in this and other threads.
My estimate puts Lion Cove P core around 25% larger than an equivalent Zen 5 P core (iso-node & without L2). And thats horrible considering there is no extra performance to justify the extra silicon. Lion Cove not just sucks... it sucks on a whole new level! It's fatter & slower than the other two.

Some napkin math (if all of them were on N3B):
  • Lion Cove - 3.4 mm2
  • Zen 5 P core - 2.74 mm2
  • Apple M4 P core - 2.8 mm2
Another interesting observation is, Zen 5 P core appears to be slightly smaller than Apple M4 P core (given the same node & without L2).
Indeed! Now the only thing missing from your observations are a cross-tab of major features for each core (SMT, AVX512, etc).

I keep hearing how impressive the M4 and M3 are; however, neither one has SMT and neither one supports AVX512 (or 256 for that matter I believe).

Considering what they do well though, it seems to come down to an argument about how specific you want your CPU design to be to a particular market segment.

Seems like the M3/4 design is uniquely qualified for thin-and-light laptops and tablets, but totally useless for a DC processor design.

Zen5 on the other hand seems to cover bases up and down the market chain, but is not as good as M3/4 where the M3/4 is strongest. This makes a great deal of sense to me since AMD is targeting the high margin (and rapidly growing) DC market where M3/4 are not (As far as I know anyway).
Well yeah but M4 P core is much better than Zen 5 core and consumes much less power, so it makes sense.

An interesting comparison will be Oryon P core vs Zen 5 core, that core is barely above 2 mm2 IIRC.
Better where? Better how?
 
Reactions: Tlh97 and Joe NYC

poke01

Platinum Member
Mar 8, 2022
2,581
3,409
106
Better where? Better how?
In apps where AVX512 isn’t used the M4 core is better. Essentially in non-SMID tasks its great.

it also delivers much higher single threaded performance while using much less power.
Like you said for the target market ie laptop they are great.
 

cannedlake240

Senior member
Jul 4, 2024
207
111
76
Raichu is saying there is a change to FE for Darkmont and will result in 3-5% ipc improvements.
Clearwater will be a decent jump over SRF and GNR-AP. Makes the rumor about Atom line Xeons/RRF being prematurely canned look more questionable... Unfortunately Pat hinted at this on earnings, apparently they don't like the 'complexity' of the dual track P/E Xeon lineup
 

Saylick

Diamond Member
Sep 10, 2012
3,644
8,222
136
If this is the big plan Intel came up with to catch up with V-Cache, it looks an inelegant approach / brute force.

Bloating the compute die, on most expensive node, with 144 MB of SRAM is not going to be very cost efficient.

JFC… here we go again. For a Western fab that claims to have packaging prowess, it sure is sad that they couldn’t do something better than to simply brute force more SRAM on the compute die.

I’m also not even sure how much diminishing returns 144 MB of SRAM on the same compute tile will have since there’s a latency trade off with larger caches, and this cache will be further away from the core since it’s all on the same die vs. 3D stacking with vias where the cache is physically closer.

 

poke01

Platinum Member
Mar 8, 2022
2,581
3,409
106
JFC… here we go again. For a Western fab that claims to have packaging prowess, it sure is sad that they couldn’t do something better than to simply brute force more SRAM on the compute die.

I’m also not even sure how much diminishing returns 144 MB of SRAM on the same compute tile will have since there’s a latency trade off with larger caches, and this cache will be further away from the core since it’s all on the same die vs. 3D stacking with vias where the cache is physically closer.

there is also this comment
 

511

Golden Member
Jul 12, 2024
1,034
894
106
Clearwater will be a decent jump over SRF and GNR-AP. Makes the rumor about Atom line Xeons/RRF being prematurely canned look more questionable... Unfortunately Pat hinted at this on earnings, apparently they don't like the 'complexity' of the dual track P/E Xeon lineup
He meant the complexity of multiple platforms for validations like 6700P/E and 6900P/E the would make a common P/E Platform
 

DavidC1

Golden Member
Dec 29, 2023
1,211
1,932
96
I thought they used 8T SRAM?
That's for the lower cache levels. 8T is 33% larger. The lower cache levels are larger even at the same "T" because it's optimized for speed, leakage, and latency. Then on top of that they use 8T.

I think it was during the Nehalem era they said they started using 8T and it got stuck. Kinda like how mass media thought Intel Gen graphics were PowerVR because they used Tile Rendering but PowerVR used deferred rendering while Intel used immediate mode.
 
Last edited:

naukkis

Senior member
Jun 5, 2002
962
829
136
So, that's a 16MB wadd for the core L3 and a 128 MB L4? Or is that just 12MB slices at 12 stations?

Probably 12MB at 12 stations. There's nothing preventing Intel doing so, L3 latency is mostly coming from ring not cache array so L3 latency is fine. Only reason they aren't doing large L3-caches is that doing it needs bigger silicon which drives costs up. For Intel internal process that's fine strategy - it's actually strange they haven't done it earlier.
 
Reactions: lightmanek

cannedlake240

Senior member
Jul 4, 2024
207
111
76
Probably 12MB at 12 stations. There's nothing preventing Intel doing so, L3 latency is mostly coming from ring not cache array so L3 latency is fine. Only reason they aren't doing large L3-caches is that doing it needs bigger silicon which drives costs up. For Intel internal process that's fine strategy - it's actually strange they haven't done it earlier.
Really so they can just increase the L3 by 4x without it incurring a massive latency penalty? Vcache is 4-5 cycle for instance
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |