To process these datasets efficiently on GPUs, the Blackwell architecture introduces a hardware decompression engine that can natively decompress compressed data at scale and speed up analytics pipelines end-to-end. The decompression engine natively supports decompressing data compressed using LZ4, Deflate, and Snappy compression formats.
Looks like the rumors were true: B100 is an MCM (not quite chiplet imo) comprised of two reticle limit TSMC N4P dies connected with a high-bandwidth interconnect, likely NVLink 5 C2C. 8 stacks of HBM3E memory total, or 4 stacks per die. It's an engineering marvel and an absolute beast but I'm kind of numb to all this AI stuff tbh, especially since it's the new buzzword so it's impossible not to get reminded of it on a daily basis. This stuff doesn't impact me directly so while it is impressive, it's largely forgettable for me, the average consumer casual gamer.
Fat chance anyone is working there who believes in improving old stuff for the good of mankind.Nvidia could perhaps improve their DirectStorage implementation from previous gen for consumer cards.
https://developer.nvidia.com/blog/n...rameter-llm-training-and-real-time-inference/
Nvidia could perhaps improve their DirectStorage implementation from previous gen for consumer cards.
I feel Nvidia trying to be better than others in some aspect of cutting edge graphics tech is as nvidia as nvidia ever gets.Fat chance anyone is working there who believes in improving old stuff for the good of mankind.
Cache in general is stupidly expensive on N3E.
Question: is consumer Blackwell also going to be a multi-chip solution? If so, that would make it very expensive since the wafer yield will essentially get halved.
We don't have any info. With what 5090 will compete N52 or N51Then if RDNA5 is really good or the AI market crashes, they can sell 5090 Ti later on.
> 512 bit GB202It's for real x2 SM compared to 5080?
Makes sense. The AI people will pay whatever it will cost. I do think that the 5090 will get a significant cut. They can sell a 176 SM 5090 for $2000 or so. Then if RDNA5 is really good or the AI market crashes, they can sell 5090 Ti later on.
Giant buses are not coming back except as the most desperate last ditch attempt to punch above your weight.
That ain't commercial, and honestly with GDDR7 having beefier chips, I'm not sure it makes sense either.
That ain't commercial, and honestly with GDDR7 having beefier chips, I'm not sure it makes sense either.
That's exactly what I said, yes. With GDDR7 having 3Go chips, the bus makes even less sense.
You do realise Nvidia is the stingiest company ever when it comes to VRAM?Going to 512-bit increases the memory capacity from 48 to 64... and 3 GB chips increases it to 96. That's the reason really.
Even the Titans didn't have 512 bit busses, Nvidia hasn't made one since GT200 15 years ago. Since then anything that's needed more than 384 bit could provide has been HBM.> 512 bit GB202
I wish we would stop with this meme. Every card that did this was an incredibly expensive waste of money. Look at the Titans.
Giant buses are not coming back except as the most desperate last ditch attempt to punch above your weight.
That's exactly what I said, yes. With GDDR7 having 3Go chips, the bus makes even less sense.
I wish we would stop with this meme. Every card that did this was an incredibly expensive waste of money. Look at the Titans.
Giant buses are not coming back except as the most desperate last ditch attempt to punch above your weight.
You seem to be judging it as a gaming chip. It's not. It's a professional & prosumer card. If gamers want to pay out of the nose for it, that's great, but it's not what it's designed for.
Even the Titans didn't have 512 bit busses, Nvidia hasn't made one since GT200 15 years ago. Since then anything that's needed more than 384 bit could provide has been HBM.