Discussion Intel Meteor, Arrow, Lunar & Panther Lakes Discussion Threads

Page 645 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Tigerick

Senior member
Apr 1, 2022
702
632
106






As Hot Chips 34 starting this week, Intel will unveil technical information of upcoming Meteor Lake (MTL) and Arrow Lake (ARL), new generation platform after Raptor Lake. Both MTL and ARL represent new direction which Intel will move to multiple chiplets and combine as one SoC platform.

MTL also represents new compute tile that based on Intel 4 process which is based on EUV lithography, a first from Intel. Intel expects to ship MTL mobile SoC in 2023.

ARL will come after MTL so Intel should be shipping it in 2024, that is what Intel roadmap is telling us. ARL compute tile will be manufactured by Intel 20A process, a first from Intel to use GAA transistors called RibbonFET.



Comparison of upcoming Intel's U-series CPU: Core Ultra 100U, Lunar Lake and Panther Lake

ModelCode-NameDateTDPNodeTilesMain TileCPULP E-CoreLLCGPUXe-cores
Core Ultra 100UMeteor LakeQ4 202315 - 57 WIntel 4 + N5 + N64tCPU2P + 8E212 MBIntel Graphics4
?Lunar LakeQ4 202417 - 30 WN3B + N62CPU + GPU & IMC4P + 4E012 MBArc8
?Panther LakeQ1 2026 ??Intel 18A + N3E3CPU + MC4P + 8E4?Arc12



Comparison of die size of Each Tile of Meteor Lake, Arrow Lake, Lunar Lake and Panther Lake

Meteor LakeArrow Lake (N3B)Lunar LakePanther Lake
PlatformMobile H/U OnlyDesktop & Mobile H&HXMobile U OnlyMobile H
Process NodeIntel 4TSMC N3BTSMC N3BIntel 18A
DateQ4 2023Desktop-Q4-2024
H&HX-Q1-2025
Q4 2024Q1 2026 ?
Full Die6P + 8P8P + 16E4P + 4E4P + 8E
LLC24 MB36 MB ?12 MB?
tCPU66.48
tGPU44.45
SoC96.77
IOE44.45
Total252.15



Intel Core Ultra 100 - Meteor Lake



As mentioned by Tomshardware, TSMC will manufacture the I/O, SoC, and GPU tiles. That means Intel will manufacture only the CPU and Foveros tiles. (Notably, Intel calls the I/O tile an 'I/O Expander,' hence the IOE moniker.)



 

Attachments

  • PantherLake.png
    283.5 KB · Views: 24,014
  • LNL.png
    881.8 KB · Views: 25,501
Last edited:

AcrosTinus

Member
Jun 23, 2024
162
163
76
Ok, but still.

-The first excuse was that ARM could never catch x86 cores even in per clock performance. Well, they did.
-The second excuse is that performance is perf/clock x clocks. Well, the M4 beats x86 processors in absolute ST performance.

We're at the third point now, which is MT performance against 200W+ processors. And I'm conveniently ignoring the much, much low power required to achieve that performance.

Like, slap Conroe vs Netburst comparison silly. Apple would have done to "Conroe" what Conroe was doing to Netburst. I knew after I got out of the silly x86 duopoly bubble mindset, after about Apple A8 I realized that what Intel achieved with Conroe/Merom would be blown away very soon.

The real truth is none of those excuses mattered, and it's simply Apple(and ARM in general) doing lot better in execution.
I am waiting for the benchmarks, if the Cinebench score is at 14700K level, I will be buying it, it is a no brainer device if the performance is right.
 

AcrosTinus

Member
Jun 23, 2024
162
163
76
I see MC has them for $420 but where did you pick one up for $360?

The 14900K is future proofed with Intel's extended warranty. If it fails you get a new one or reimbursed. If it makes 5 years then good on you!
A friend bought a 14900K 3 Months ago and never finished building his system, the benefits of having hardware friends in whatsapp, one question into the group and you might even get hardware for free.

Yeah the prolonged warranty lets me ride out the 14900K into the future, hopefully all bugs are squashed by then.
 

511

Golden Member
Jul 12, 2024
1,034
894
106
Here is another kicker Apple beat Intel to include proper Thunderbolt 5 on desktop today and most likely tomorrow on a laptop. The full spec 120GB/s.

What IS Intel doing? They used to be at the fore front of USB/thunderbolt standards. You can tell Arrow lake was delayed by the fact it does not support TB5.
There is a Razer Balde that supports TB5 but thats only on 1 port @ 80GB/s. The Razer laptop just use a discrete chip, Barlow ridge and unlike TB4 its not baked in.
This shows how bad they have fallen anyway intel is the sole manufacturer for TB IC retimers have to be bought from Intel just released it to be first to ThunderBolt 5 in name
 

511

Golden Member
Jul 12, 2024
1,034
894
106
I am waiting for the benchmarks, if the Cinebench score is at 14700K level, I will be buying it, it is a no brainer device if the performance is right.
I think it should be with M4 Max the power consumption is increasing though if i had to guess around 80W
 
Last edited:

511

Golden Member
Jul 12, 2024
1,034
894
106
M4 ~ M3 Max level performance which is around 1600 r24 while consuming the same 52-55W iirc btw apple lists it as CPU Power not package power right?
 

DavidC1

Golden Member
Dec 29, 2023
1,211
1,932
96
Based on the benches below it seems as though the L2 is where ARL is falling behind RPL, yes?

View attachment 110559

View attachment 110560
Yea L2 has lot less bandwidth. So they enlarged the L2 cache but now has half the bandwidth in Read. Copy is Read + Write so naturally it'll be lower.

Don't forget the fabric affecting memory latency. L3 latency based on AIDA seems to be ok.
This may be the worst launch Intel has done since the Pentium 4. Granted Haswell to Kaby Lake weren't up to much, but at least it didn't freaking regress.
Here's a key difference. Pentium 4 had amazing engineering and ideas on it, just that the ideology behind it sucked. Trace Cache, Double pumped ALU. They demonstrated an ALU at 10GHz. They had undisputed process leadership. Idontcare and many other people were talking about how SEM shots of Intel's transistors were clean and neat, while others were more "messy". It was amazing.

They are losing that base. They used to lead in memory standard. It was Intel chips that supported the fastest memory speeds, it was Intel that supported the latest IO standards(PCI Express, USB). Some were flawed like BTX, but again computing was spearheaded by Intel.

I'm not happy at all saying this. They are losing the foundation they used to have. Yes, they still were led in a stupid direction, but the people under it made it work.

@igor_kavinski The common belief is that Swan was brought in to stabilize the company, so Gelsinger could fix it. They needed someone to manage after Kraznich was kicked out abruptly.
 
Last edited:

naukkis

Senior member
Jun 5, 2002
962
829
136
Yea L2 has lot less bandwidth. So they enlarged the L2 cache but now has half the bandwidth in Read. Copy is Read + Write so naturally it'll be lower.

That's not the whole case. LC L2 is actually L3, there's that intermediate cache level there. Adding cache levels have always drawbacks but increasing L2 lower level cache from 48KB to 192KB should reduce L2 bandwidth requirement for usual workload over factor 2, which makes it possible sacrifice L2 bandwidth for bigger size.
 
Reactions: Tlh97 and Elfear

DavidC1

Golden Member
Dec 29, 2023
1,211
1,932
96
That's not the whole case. LC L2 is actually L3, there's that intermediate cache level there. Adding cache levels have always drawbacks but increasing L2 lower level cache from 48KB to 192KB should reduce L2 bandwidth requirement for usual workload over factor 2, which makes it possible sacrifice L2 bandwidth for bigger size.
I forgot about the hierarchy change in Lion Cove.

So many changes for so little.
 

MS_AT

Senior member
Jul 15, 2024
365
798
96
This is why I keep saying, the mini-Netburst era needs to go away, both at AMD and Intel. Back down on the ridiculous clocks, get rid of the uop cache since now they have an alternative that's more efficient.
Efficient in which terms? Power efficient? Space efficient? We haven't seen three-way decode clusters in action in high performance oriented designs yet. And Skymont's goal is performance per area, I would guess, while uop cache is doing fine in Zen5.
Looks like RAM speed/bandwidth does play a role in CB2024.
Well, yes. That was conclusion from C&C review of CB2024 from October 2023. This together with little use of SIMD in instructions mix explains why Arrow Lake and Apple M chips are doing well there, though for M chips we would need somebody check the assembly, to verify the instruction mix is placing the same emphasis on scalar operations.
 

naukkis

Senior member
Jun 5, 2002
962
829
136
This together with little use of SIMD in instructions mix explains why Arrow Lake and Apple M chips are doing well there, though for M chips we would need somebody check the assembly, to verify the instruction mix is placing the same emphasis on scalar operations.

CB gives actually very realistic use-case of SIMD instructions use. Some 128-bits are used, about 6% of instructions in CB2024 which other way states that about 25% of it's code is vectorizable. Longer vectors aren't useful at all. And it's safe to assume that ray tracer engine is pretty well optimized. x86 vendors should also optimize their designs to those general-usage patterns instead of long-vector HPC use cases. If they don't gap to ARM designs will be widening generation after generation. Skymont is actually sane general use case-optimized x86 design - hopefully Intel won't ruin it too in next generations by widening SIMD-units.
 

511

Golden Member
Jul 12, 2024
1,034
894
106
CB gives actually very realistic use-case of SIMD instructions use. Some 128-bits are used, about 6% of instructions in CB2024 which other way states that about 25% of it's code is vectorizable. Longer vectors aren't useful at all. And it's safe to assume that ray tracer engine is pretty well optimized. x86 vendors should also optimize their designs to those general-usage patterns instead of long-vector HPC use cases. If they don't gap to ARM designs will be widening generation after generation. Skymont is actually sane general use case-optimized x86 design - hopefully Intel won't ruin it too in next generations by widening SIMD-units.
Widening of SIMD unit is undeniably going to happen to be able to handle 512bit Vectors from 256bit currently
AVX10 clears the fragmentation issues and removal of instructions whenever someone wants
 
Last edited:

MS_AT

Senior member
Jul 15, 2024
365
798
96
I think Arrow Lake shows why CB2024 should never ever be taken seriously as a benchmark ever again. And to an extent, Geekbench too. Scores of both failed to predict serious regression inducing issues/bugs in ARL's design.
I disagree. You just need to keep in mind what they are measuring when looking at the results as it is impossible to create one benchmark to rule them all.
CB gives actually very realistic use-case of SIMD instructions use. Some 128-bits are used, about 6% of instructions in CB2024 which other way states that about 25% of it's code is vectorizable. Longer vectors aren't useful at all. And it's safe to assume that ray tracer engine is pretty well optimized. x86 vendors should also optimize their designs to those general-usage patterns instead of long-vector HPC use cases. If they don't gap to ARM designs will be widening generation after generation. Skymont is actually sane general use case-optimized x86 design - hopefully Intel won't ruin it too in next generations by widening SIMD-units.
is the 25% statement from Maxon? As the whole benchmark seems to be written with idea of highest portability, first. The wider the SIMD the less instructions to decode, thus I guess why they opt to move in that direction
No, AVX10 just brings AVX512-instructions support to shorter registers without need to support 512 bits. So desktop-mobile could have AVX128 design.
The AVX10/128 was dropped. Desktop will stick with AVX10/256 [this make using 128b possible, but requires that 256 be supported], while server will use AVX10/512. This is a guess based on Intel's current policy. But this will anyway bring improvements, by porting AVX512 features without requiring 512b registers.
 

poke01

Platinum Member
Mar 8, 2022
2,581
3,409
106
I think Arrow Lake shows why CB2024 should never ever be taken seriously as a benchmark ever again. And to an extent, Geekbench too. Scores of both failed to predict serious regression inducing issues/bugs in ARL's design.
This doesn't make sense cause SPEC and R23 also showed that Arrow Lake made a generational increase over Raptor Lake.
Do we also disregard SPEC and R23 because these also failed to show regression in Arrow Lake?

IMO, these benchmarks should be taken as a guide and let real world productivity/gaming tests show the increases for your workflows.
 

511

Golden Member
Jul 12, 2024
1,034
894
106
No, AVX10 just brings AVX512-instructions support to shorter registers without need to support 512 bits. So desktop-mobile could have AVX128 design.
Yes that was the main point and the issue with different CUPID of instruction it will provide a better baseline than AVX-512 ever was in terms it will directly affect the fragmentation in AVX-512 it will actually provide a baseline unlike the mess AVX-512 is
 

AcrosTinus

Member
Jun 23, 2024
162
163
76
I think it should be with M4 Max the power consumption is increasing though if i had to guess around 80W
You might be right, the good thing is that Apple has the P-Core to E-Core ratio right, Intel should copy that. Even on the smartphone side with Snapdragon 8 Elite, people are slowly moving towards Apple's core configs. They might be onto something, a indication for a 4P-6E Lunar Lake successor in the future.
 
Last edited:

Kepler_L2

Senior member
Sep 6, 2020
581
2,401
136
Widening of SIMD unit is undeniably going to happen to 512Bits AVX10 clears the fragmentation issues
No
No, AVX10 just brings AVX512-instructions support to shorter registers without need to support 512 bits. So desktop-mobile could have AVX128 design.
Also no.

AVX10.2/128 is dead, Intel will continue with 512-bit implementation in servers and add 256-bit implementation in client.
 
Reactions: GTracing
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |