SiliconFly
Golden Member
- Mar 10, 2023
- 1,925
- 1,284
- 96
I know we've been over this before, but in the light of this release, and Blackwell being rather derivate, and arguably Ada and Ampere as well... What does NVidia have coming up over the next, say two generations? Do we just iterate? Is both node and architecture stagnating? Will that make it easier for others to catch up?
Yes, but they have an even bigger problem than Intel because the nodes are now bumping up against the laws of physics. Intel sat on the same lithography node for like a decade. nVidia or anyone else interested in shrinking down below 4nm is kinda in a no-win situation. Or at least there are no more easy performance gains to be had without extra cost being absorbed or passed on to the consumer.Now we’re talking! 🤣
Nvidia is the new Intel.
Seeing as how Blackwell already maxes out the die size and TDP on N4P, I suspect Rubin flagship will likely do the same, so whatever density improvements N3E gives over N4P, that'll be the upper limit on the raw performance improvement. They'll do their usual doubling of RT intersection performance and possibly 2x tensor units (not another reduction in precision), but that'll be it besides a whole dollop of software features.I know we've been over this before, but in the light of this release, and Blackwell being rather derivate, and arguably Ada and Ampere as well... What does NVidia have coming up over the next, say two generations? Do we just iterate? Is both node and architecture stagnating? Will that make it easier for others to catch up?
uh. nope.It's clear the push from AMD and NVidia will probably be more on AI and RT improvements.
The evidence doesn't really seem to bear that out though.
In lower end 50 series parts Die size is about the same as 40 series, for about the same Shader count as last generation, and 5090 expanded proportionally to the SM count increase, so NVidia does not seem to have any die size bloat from the technology update.
Plus this generation AMD is supposed to be adding real AI tensor cores, and beefing up it's RT capability. So AMD may actually have bloat while NVidia is showing none.
So again, no evidence that AMD has any advantage to exploit, so would need a die just as monstrous as 5090's to compete... which I still argue is a complete non starter.
Mind that it's binned. So it's 'only' like 50% moar SMs.OTOH the 4090 is 60% larger than the 4080 for ~30% additional gaming performance
Mind that it's binned. So it's 'only' like 50% moar SMs.
Zen is way different. Zen does not use a huge complex monolith for their consumer line and is cheap to build. AMD just been delivering at a consistence cadence for their CPU division while keeping power consumption relatively low which is what help it gain consumer confidence. It was not the threadrippers that did it.
Going Halo for Nvidia makes more sense since they have a professional visualization market which provides guaranteed income at high margins.
AMD doesn't have much for this market. As we can see with the possible performance flop of the RTX 5090, going big carries more risk as more things can go wrong. The GTX 480 and Fury X are examples of this. The RTX 5090 is clocked awfully low for the amount of power it uses. This might be a fermi like chip which misses targets in the gaming segments but recovers sales in the professional markets.
Also chip design is more expensive than that. Both companies R and D expenditures seem to line up with this chart.
Here is a newer chart.
I was looking into the software part and it's ridiculously expensive. Someone on reddit says the licenses for single software can be a million dollars per year and indeed use the software mentioned by XPEA mentioned(Cadence and synopsys) and they have hundreds of these licences which are only good for one year. A single big chip can use hundreds of licenses in their design.
So would AMD rather uses these resources on something like instinct or a Halo discrete graphics card? I think we all know the answer. AMD is just following the money.
Spending hundreds of millions on a single halo just doesn't make sense when that money can be spent on datacenter where sales are in billions, not hundreds of millions.
Don't have to. Just stack.so would need a die just as monstrous as 5090's to compete...
Maybe? GB100 was issues galore though.What's the odds that Nvidia will go with a MCM solution for gaming Rubin?
unlikelyWhat's the odds that Nvidia will go with a MCM solution for gaming Rubin? Just seems fishy that GB203 is half of GB202, as if they already have a split L2 solution working for GB202 and just need a silicon bridge to make MCM work next gen.
I also finished watching that. What an insane design. It's doing what I thought was possible only with HBM parts. Nvidia allowing an engineer to talk about one of the most interesting parts of Blackwell is appreciated.The FE cooler is in very impressive if it can dissipate that much heat in just 2 slots. But all that heat goes right into the case unlike an AIO.
Great engineering. Single slot high powered cards can be a thing again.The FE cooler is in very impressive if it can dissipate that much heat in just 2 slots. But all that heat goes right into the case unlike an AIO.
Yes, excellent job from the Nvidia engineers, I hope to see this rub off to AIBs too. One still needs to carefully plan what to do with the heat building up in the case, but ultimately that is a problem for all builds - one either fixes the heat flow or they get meh results.The FE cooler is in very impressive if it can dissipate that much heat in just 2 slots. But all that heat goes right into the case unlike an AIO.
128/76 = 68% SM difference between 4080 and 4090Mind that it's binned. So it's 'only' like 50% moar SMs.
Budget was reallocated to the RT core. Might even call it a level 4 RT core with the features.I get the feeling that dual issue isn't properly "fixed" as in all cores being FP32/INT now does not make them equal, perhaps they did not have transistor budget to fix other bottlenecks and left it for N3 work
If this is what future is going to look like from now on, not speaking strictly about Nvidia and GPUs, but broadly about industries/society, it pretty much means technological progress available only to "rich" people going forward.Just got this & it's a bit disturbing! Except for the top die, just look at the die size, transistor density & most importantly the transistor count. Kinda disillusioned & feeling a bit nauseated.
View attachment 115021
This ladies & gentlemen, is clear evidence of cheating. I think I'll be skipping 50 series.
Is this description based on what Imgtech introduced or does NV have its own reference?Budget was reallocated to the RT core. Might even call it a level 4 RT core with the features.
GPU market and the gaming industry should become more commodified for it to remain hot and interesting and price competitive.If this is what future is going to look like from now on, not speaking strictly about Nvidia and GPUs, but broadly about industries/society, it pretty much means technological progress available only to "rich" people going forward.