Question The "lens" of two NVMe(s) in RAID 0

preWarp

Junior Member
Dec 28, 2024
5
0
6
I have been trying to over come an illusion I am experiencing where I perceive little difference in running desktop applications on newer hardware.

If I am using a 16 lane graphics card and two NVMe SSD(s) in RAID 0 on a machine with only 20 pcie lanes I should notice significant gains in desktop software performance if I am using a graphics card and 2 NVMe SSD(s) in RAID 0 on a machine with 24 available pcie lanes.

This is the way the Intel core ultra 9 & core ultra 5 compare to Intel core i9-14900 and other desktop processors that came before it down to the core i9-11900k.

So, after installing a graphics card having only 4 remaining pcie lanes is the reason that, for example the Intel core i9-13900k is slower than a core ultra 9 285?
 

Soulkeeper

Diamond Member
Nov 23, 2001
6,731
155
106
It's doubtfull that the extra 4 pcie lanes would be making a big difference in most things.
Unless you are running a benchmark that specifically isolates that area.
 

zir_blazer

Golden Member
Jun 6, 2013
1,206
502
136
I believe that whatever your understanding about how "available pcie lanes" are supposed to work is completely wrong. Try reading this (For all practical purposes the Chipset is a giant multiplexer) and see if you can put a more specific example on an actual Motherboard.
 

preWarp

Junior Member
Dec 28, 2024
5
0
6
I believe that whatever your understanding about how "available pcie lanes" are supposed to work is completely wrong. Try reading this (For all practical purposes the Chipset is a giant multiplexer) and see if you can put a more specific example on an actual Motherboard.
Ok. So one can have a socket on the motherboard and using a specific motherboard chipset can have many more pcie lanes than exist from the processor package alone?

So going all the way from Sandy bridge (2011) to lion cove (2024) I can have 2 NVMe SSD(s) in RAID 0 and have the same disk I/O?

This then is the door that opens to the real power of new hardware at each stage from Sandy bridge onward?
 

zir_blazer

Golden Member
Jun 6, 2013
1,206
502
136
I still believe that your understanding is completely wrong because you began asking if available PCIe Lanes makes one Processor slower than other. PCIe Lanes only affect the possible amount of external data movement at the same time (Assuming you are using them), so if you're asking about if more PCIe lanes magically increases Processor processing performance, the answer is no.

For most practical purposes, assume that all data is going to go though the Processor (PCI Express supports P2P (Peer-to-Peer) so one PCI Device may theorically talk directly to another, but that could make the example too complicated...).
Processor has its own PCIe Controller that exposes PCIe Lanes that are wired directly to it, so these are the lowest possible latency connectivity for a PCI Device to get to the Processor, besides than the bandwidth is entirely dedicated because it is not multiplexed at any point with any other device (Well, except most likely internally on the Processor, but we don't care about that).
Some PCIe Lanes (Which Intel calls DMI, don't recall if AMD has any fancy name) also links Processor with Chipset. Chipset acts like a glorified multiplexer because it provides a lot of connectivity (Whenever extra PCIe Lanes, SATA, USB) but you have a limited upstream link bandwidth connecting the Processor and Chipset, so if you were to try to use ALL Chipset connectivity at the same time, you would be bottlenecked by this Processor-Chipset link. For example, most LGA 1700 Chipsets uses 8x PCIe 4.0 lanes between Processor and Chipset, but many Motherboards can be populated with three 4x PCIe 4.0 NVMe SSDs on Chipset lanes, yet you wouldn't get triple performance by putting them on RAID 0 because the uplink bottleneck (If they were PCIe 3.0, you could, because they can't saturate the upstream link). Also, EVEN if you were not bottlenecked, Chipset lanes are slighty worse because they incur extra latency because they're further away from the Processor, having to go though the internal Chipset PCIe Switch to get to the Processor.

Yet, for all practical purposes, I believe that you're most likely overstimating RAID 0 performance. Do you move data in bursts that could require a level of performance beyond what a single PCIe 5.0 SSD on Processor lanes gives?
 
Reactions: Super Spartan

preWarp

Junior Member
Dec 28, 2024
5
0
6
...so if you're asking about if more PCIe lanes magically increases Processor processing performance, the answer is no..

....if you were to try to use ALL Chipset connectivity at the same time, you would be bottlenecked by this Processor-Chipset link.

For example, most LGA 1700 Chipsets uses 8x PCIe 4.0 lanes between Processor and Chipset, but many Motherboards can be populated with three 4x PCIe 4.0 NVMe SSDs on Chipset lanes,

Also, EVEN if you were not bottlenecked, Chipset lanes are slightly worse because they incur extra latency because they're further away from the Processor, having to go though the internal Chipset PCIe Switch to get to the Processor.

Yet, for all practical purposes, I believe that you're most likely overestimating RAID 0 performance.

From reading your post I have a better understanding of the motherboard chipset role in assigning multiple addresses per each lane available from the processor.

I had a Skylake based Intel Core i3-6100. It had 16 PCIe lanes. I upgraded to another Skylake based processor that still had 16 PCIe lanes, the Core I7-10700. Then I upgraded to Rocket Lake, the Core i5-11600 and now I have 20 PCIe lanes. Rocket Lake, Alder Lake, and Raptor Lake all have 20 PCIe lanes. Onward to Arrow Lake, if I upgraded to a Core Ultra 245k I would now have 24 PCIe lanes.

The motherboard, can take the max available lanes from the processor and make more lanes for peripherals, but total bandwidth is limited by the sum of the max lanes available from the processor.

So my question is given the limitation to the max lanes available from the processor, in my case it would be 16 for Skylake and 20 for Rocket Lake, why am I seeing the same performance using desktop applications? Is it because I need 2x NVMe in RAID 0? If I benchmarked disk performance for 2x NVMe in RAID 0 on several Skylake machines from the 6th generation to the 10th Generation would the throughout be the same? If the same then upgrading from any generation based on Skylake would show no difference in use of desktop applications. Running CPU and Memory benchmarks, you could see speed jumps in your processor bandwidth and for the dynamic RAM but one could not perceive a change in performance of desktop applications. The loading time would make it seem like the computer is running at the same speed as the older machines.

If you upgraded to Arrow Lake, going from 16 lanes to 24 lanes there is a perceived improvement in loading times not because you have a much faster CPU and faster memory access times but because your two drives each have 4 whole lanes a piece instead of borrowing from some of the lanes going to the video card.
 
Last edited:

zir_blazer

Golden Member
Jun 6, 2013
1,206
502
136
The motherboard, can take the max available lanes from the processor and make more lanes for peripherals, but total bandwidth is limited by the sum of the max lanes available from the processor.
You got it wrong. Is not about total Processor lanes, is about the lanes that are connected between devices. You're failing to view PCIe lanes as what they are: Pipe of a certain size, where at any given time you may have a maximum flow of data going though it. If the devices that are connected to that pipe (Or better said, PCIe link) are actually capable of moving that amount of data is another matter.
If instead of a Video Card you used a 16x PCIe Slot wired to the Processor lanes with a 100G NIC and wanted to download a file to a NVMe SSD plugged to a M.2 slot wired to the Chipset PCIe Lanes, data has to go from NIC -> PCIe link -> Processor internal PCIe Switch -> PCIe link -> Chipset internal PCIe Switch -> PCIe link -> NVMe SSD. The links between the NIC and the Processor and the Chipset and the NVMe SSD are dedicated to usage for these devices, but Processor and Chipset link will be absolutely NOT be used only by the NIC, but data from multiple devices plugged to Chipset would be going though it at the same time. That is the multiplexing part that I mentione before.

I tried to look around if I can get a link that explains it better than me and I found this: https://www.highpoint-tech.com/post...-switch-chipsets-for-pcie-gen3-m-2-nvme-cards
The concept that I try to explain mostly works. You only need to consider than the Chipset is a PCIe Switch PLUS some built in devices (Mainly USB and SATA), but is limited by the upstream bandwidth from its usually 4 or 8 lanes to Processor.
You can also try looking at ThreadRipper Pro / WRX90 platform Block Diagrams if you want to see a 128 PCIe Lanes monster.


So my question is given the limitation to the max lanes available from the processor, in my case it would be 16 for Skylake and 20 for Rocket Lake, why am I seeing the same performance using desktop applications?
Running CPU and Memory benchmarks, you could see speed jumps in your processor bandwidth and for the dynamic RAM but one could not perceive a change in performance of desktop applications. The loading time would make it seem like the computer is running at the same speed as the older machines.
Because SSDs may be faster in throughput but latency has remained roughly the same since the introduction of the NVMe protocol, so you feel no difference in desktop applications, and you will not feel them either in a RAID 0 with PCIe 5 SSDs unless you daily routine involves moving terabytes of data from one place to another.
The only technology that was faster than standard Flash in a SSD was what Intel used in its Optane SSD and it was discontinued. This is what you want:

You don't want more bandwidth/throughput, you want lower latency.
 
Last edited:
Reactions: In2Photos

preWarp

Junior Member
Dec 28, 2024
5
0
6
I have thought about your explanation and it seems like no matter how many PCIe lanes the processor features and no matter how many NVMe drives you have in RAID 0 the storage performance will be woefully slower than the CPU.

I also started to wonder about features of a chipset specification by Intel and if it mentioned how many total PCIe lanes and how many lanes were for NVMe. The specifications for the z890 chipset show 24 maximum PCIe lanes and possible arrangements like 16x1 and 8x8. It seemed like a manufacturer could have 16 lanes for a x16 slot and the remaining 8 lanes for a x16 (limited to 8 lanes), or two NVMe slots.

Then I obtained the manual for the following Asus z890 motherboard:

Z890 AYW
GAMING
WIFI W

The manual shows 4 M.2 slots. Under a heading mentioning the processor (Intel core ultra series 2) showing is M.2_1 having 4 lanes PCIe 5.0 and M.2_2 having 4 lanes PCIe 4.0. Following is a heading for the chipset (the Intel z590) showing M.2_3 and M.2_4 as having 4 lanes PCIe 4.0 a piece.

So a graphics card occupying the 16 lane PCIe 5.0 slot would be connected to 16 of 24 lanes from the processor, thanks to the z890 chipset specification which allows (1) x16 + (1) x4 for PCIe 5.0. Leaving (1) x4 for PCIe 4.0 (24 total PCIe lanes). An NVMe SSD in the first M.2 slot would consume x4 PCIe 5.0 and a second disk in the second M.2 slot would consume x4 PCIe 4.0. 24 PCIe lanes are gone.

How much faster are desktop applications on this machine versus a z590 based motherboard with a Intel core i5-11600 and only one NVMe drive?

If the z890 chipset machine described only had one nVME drive would not this new PC seem the same slow speed of its predecessor because I only have one NVMe and not two?
 
Last edited:

zir_blazer

Golden Member
Jun 6, 2013
1,206
502
136
How much faster are desktop applications on this machine versus a z590 based motherboard with a Intel core i5-11600 and only one NVMe drive?

If the z890 chipset machine described only had one nVME drive would not this new PC seem the same slow speed of its predecessor because I only have one NVMe and not two?
I don't know how many times I have to state this. YOUR UNDERSTANDING OF HOW PCI EXPRESS IMPACTS PERFORMANCE IS COMPLETELY WRONG. A PCIe link is a highway with X amount of lanes and Y max car speed (Given by Generation, like PCIe 3/4/5) which gives a potential maximum of Z bandwidth if it was used to its full theorical capacity. If you are NOT actively using the bandwidth because there is no device connected, having more lanes does nothing, and even if you have a device connected, it doesn't mean that it will use the maximum bandwidth, so it doesn't matters how much lanes a platform has unless you plan to actively use them by having either more or wider, faster devices plugged in.
Having two SSDs doesn't even mean that you are accessing them simultaneously, because if you had a common non-RAID configuration with one drive for Windows and the other exclusively for games, the other SSD 4 lanes are sitting near idle unless you're actually loading a game.

Your dekstop application performance and how fast and responsive a system feels is 100% limited by latency, NOT throughput. Two SSDs in RAID 0 would have twice the maximum throughput, but latency would still remain the same OR worse, cause you need some processing power to strip what bits goes to which SSD, and that extra step will increase latency, even if it is minimum. As I said in my previous post, the only thing with lower latency than possible with a high end NAND Flash based SSD (Which is every SSD in existence that is NOT 3D XPoint/Optane) is an Optane SSD itself (If somehow you need more than that you're in fancy PCIe RAMDisk territory). All other performance differences are minimal or incremental unless you're actually moving or processing data. I don't know how else to explain this before you understand that PCIe lanes is the least of your problems.

Stop wasting your time with PCIe related questions because NOTHING related to that will make your system feel faster. Spend your time researching what makes Optane faster.
 

DaaQ

Golden Member
Dec 8, 2018
1,725
1,213
136
I don't know how many times I have to state this. YOUR UNDERSTANDING OF HOW PCI EXPRESS IMPACTS PERFORMANCE IS COMPLETELY WRONG. A PCIe link is a highway with X amount of lanes and Y max car speed (Given by Generation, like PCIe 3/4/5) which gives a potential maximum of Z bandwidth if it was used to its full theorical capacity. If you are NOT actively using the bandwidth because there is no device connected, having more lanes does nothing, and even if you have a device connected, it doesn't mean that it will use the maximum bandwidth, so it doesn't matters how much lanes a platform has unless you plan to actively use them by having either more or wider, faster devices plugged in.
Having two SSDs doesn't even mean that you are accessing them simultaneously, because if you had a common non-RAID configuration with one drive for Windows and the other exclusively for games, the other SSD 4 lanes are sitting near idle unless you're actually loading a game.

Your dekstop application performance and how fast and responsive a system feels is 100% limited by latency, NOT throughput. Two SSDs in RAID 0 would have twice the maximum throughput, but latency would still remain the same OR worse, cause you need some processing power to strip what bits goes to which SSD, and that extra step will increase latency, even if it is minimum. As I said in my previous post, the only thing with lower latency than possible with a high end NAND Flash based SSD (Which is every SSD in existence that is NOT 3D XPoint/Optane) is an Optane SSD itself (If somehow you need more than that you're in fancy PCIe RAMDisk territory). All other performance differences are minimal or incremental unless you're actually moving or processing data. I don't know how else to explain this before you understand that PCIe lanes is the least of your problems.

Stop wasting your time with PCIe related questions because NOTHING related to that will make your system feel faster. Spend your time researching what makes Optane faster.
Question, 2x4TB m.2 drives in raid to get around dual drive letters, would that be worth doing? 2x4TB was at the time cheaper than 1 8TB. I just don't want some that may access my PC forget to change install location from C to D.

Don't care about the speed increase from RAID0, or JBOD.
 

preWarp

Junior Member
Dec 28, 2024
5
0
6
I am sorry zir_blazer but I have metamorphosed into a new species since I first started doing fact checking for this post.

This being which has emerged on our planet feels like storage is the greatest latency in a PC of any time from the first to the last.

Doing anything in your budget to reduce the latency for storage is crucial to experiencing the full potential of an upgrade to a major milestone in technology such as, for example, rocket lake to arrow lake.

PCIe lanes on die, the chipset on the processor, is at a premium compared with the time when the chipset was on the motherboard. Now having 20 lanes instead of 16 or 24 instead of 20 matters more than ever

Anyway. Phone is out of juice. Bye.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |