Question 'Ampere'/Next-gen gaming uarch speculation thread

Page 45 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Ottonomous

Senior member
May 15, 2014
559
292
136
How much is the Samsung 7nm EUV process expected to provide in terms of gains?
How will the RTX components be scaled/developed?
Any major architectural enhancements expected?
Will VRAM be bumped to 16/12/12 for the top three?
Will there be further fragmentation in the lineup? (Keeping turing at cheaper prices, while offering 'beefed up RTX' options at the top?)
Will the top card be capable of >4K60, at least 90?
Would Nvidia ever consider an HBM implementation in the gaming lineup?
Will Nvidia introduce new proprietary technologies again?

Sorry if imprudent/uncalled for, just interested in the forum member's thoughts.
 

Ajay

Lifer
Jan 8, 2001
16,094
8,106
136

And here, Kopite says that power draw will acttually be between 300 and 375W.

Considering the die configurations of 102 chip, I would say that he means that RTX 3080 is going to use 300W of power, and RTX 3090 will(or something with full die) will use 375W of power.
But that just nominal specs. 300W for 2x8 pin PCIe power connectors plus 75W PCIe slot power. So it's odd to have a 'leak' that just measures typical PCIe power specs.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
But that just nominal specs. 300W for 2x8 pin PCIe power connectors plus 75W PCIe slot power. So it's odd to have a 'leak' that just measures typical PCIe power specs.
He is talking about ACTUAL power draw. Not what is possible.
 

Konan

Senior member
Jul 28, 2017
360
291
106
There was a debate on chiphell forums regarding the attached (note: speculation). Now it looks like the thread got pulled strangely, or member locked. Anyways, I think the higher end Nvidia cards Titan / 3090 / 3080 TDP looks reasonable.
Also note the GDDR6 for 3080 and 3090/Titan with GDDR6x.

 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
Well, Kopite was the first to even come out with GDDR6X idea. And yet, people find it easy to believe this part of his info, and find it hard to believe 102 dies will draw past 300W of power.

Despite of what he was correct on in previous times when it goes to Nvidia products...
 

jpiniero

Lifer
Oct 1, 2010
14,834
5,450
136
Well, Kopite was the first to even come out with GDDR6X idea. And yet, people find it easy to believe this part of his info, and find it hard to believe 102 dies will draw past 300W of power.

Despite of what he was correct on in previous times when it goes to Nvidia products...

If it's 280, 290 W that might be too close for comfort.

Remember the 2080 Ti FE was also dual 8-pin.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
If it's 280, 290 W that might be too close for comfort.

Remember the 2080 Ti FE was also dual 8-pin.
For 3080, sure. That is what we might see. Something between 260 and 300W under load.

I find it baffling that people knowing three things: Ampere/GA100 TDP of 400W, Samsung Process being less efficient than TSMC's 7 nm, and more importantly - Nvidia being required to crank up the clock speeds way past 1.9 GHz, losing efficiency, reject the idea that next gen Nvidia GPUs could use up to 375W of power, for Founders Edition.

Guys, wake up. The paint is already on the wall touted time and time again by different people.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
Erm, no, it actually experiences decrease in GFLOP/watt, compared to Volta GPUs.

Why would we see anything different with Gaming cards, especially considering there are no physical design improvements on SS's process, because Nvidia hand't had the time to make them?
 

sontin

Diamond Member
Sep 12, 2011
3,273
149
106
So power consumption is over 900W for A100?!


3x more performance with FP16 and 11x more than T4 (70W).
 
Reactions: Mopetar and DXDiag

sontin

Diamond Member
Sep 12, 2011
3,273
149
106
Which gaming Ampere version will have huge TensorCores which cant be used for game rendering outside of DL? A100 delivers 2,5x more performance with TensorCores (FP16). Claiming that GFLOPs/Watt decreased is just wrong.
 

DXDiag

Member
Nov 12, 2017
165
121
116
Erm, no, it actually experiences decrease in GFLOP/watt, compared to Volta GPUs.
Nope, You are measuring traditional flops, this is an AI chip, most of it's transistors have gone to the tensor cores, for which it provides crazy 2.5X speeds against previous gen.

And the 400W thing is irrelevant, the A100 is SXM4 form factor, this form factor is 400W for unlocked performance during extended usage. NVIDIA could have made this chip 350W or 300W within the same clocks if they wanted to, V100 SXM2 was 300W, SXM3 was 350W and 450W respectively.

You need to follow the world of data centers to understand this.
 
Reactions: ozzy702

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
Nope, You are measuring traditional flops, this is an AI chip, most of it's transistors have gone to the tensor cores, for which it provides crazy 2.5X speeds against previous gen.

And the 400W thing is irrelevant, the A100 is SXM4 form factor, this form factor is 400W for unlocked performance during extended usage. NVIDIA could have made this chip 350W or 300W within the same clocks if they wanted to, V100 SXM2 was 300W, SXM3 was 350W and 450W respectively.

You need to follow the world of data centers to understand this.
I always find it funny, when people are unable to look the message that is BEYOND one particular post, and read the context.

That is not the point I was making.

Secondly. Nvidia is rumored to increase Tensor, and RT capabilities with next gen. gaming cards, compared to Turing. You guys genuinely believe that those capabilities come as a free lunch from perspective of power draw? The very reason why Ampere GA100 chip is less efficient per watt compared to Volta in FP32 and FP64 are those increased AI capabilities in GA100 chip!

What do you guys think will happen with gaming cards, considering that Samsungs process is LESS EFFICIENT than TSMC's 7 nm process, Nvidia will crank the clock speeds to the roof to compete with AMD, and they haven't had time to optimize the physical design of next gen gaming cards, like they had opportunity with Turing?

The end result is 300W power draw on RTX 3080 based on 102 chip, and RTX 3090 drawing way past 300W of power. Why is it so hard to see this?
 

sontin

Diamond Member
Sep 12, 2011
3,273
149
106
With FP16 A100 can deliver 2x the efficiency of V100 but with FP32 is would be worse?

And why cant you accept that Samsung has started mass production of 8nm chips since the summer of 2018 which would have given nVidia enough time to optimize the process like with the 7nm process.
 

Tup3x

Golden Member
Dec 31, 2016
1,010
998
136
Why do you assume that it's less efficient at FP32 and FP64? The TDP is when under maximum load. It is likely more efficient when running FP32 stuff than Turing. Die shrink alone should take care of that. Compute Ampere is different enough so that it probably doesn't make sense to judge gaming Ampere cards based on how it performs.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
Why do you assume that it's less efficient at FP32 and FP64? The TDP is when under maximum load. It is likely more efficient when running FP32 stuff than Turing. Die shrink alone should take care of that. Compute Ampere is different enough so that it probably doesn't make sense to judge gaming Ampere cards based on how it performs.
Because the numbers, yes theoretical, show that Ampere is actually less efficient in FP32/64 per watt.

Also, so FP32 Load is not "full" load, now. Got it.

Also, its actually on you to prove that that 400W TDP rating is not done on maximum load of FP32 numbers... And that it is not full load.
 

Stuka87

Diamond Member
Dec 10, 2010
6,240
2,559
136
So power consumption is over 900W for A100?!


3x more performance with FP16 and 11x more than T4 (70W).

You cannot compare that load to gaming load that we will see with 3x00 series chips. Those are very specific use cases, and part of the reason Ampere is faster there is because of improvements in those functions. Not from raw horse power alone. The benchmarks in that article are very much cherry picked by nVidia marketing. None of them translate to the gaming cards.
 

Ajay

Lifer
Jan 8, 2001
16,094
8,106
136
With FP16 A100 can deliver 2x the efficiency of V100 but with FP32 is would be worse?

And why cant you accept that Samsung has started mass production of 8nm chips since the summer of 2018 which would have given nVidia enough time to optimize the process like with the 7nm process.
Apparently SS 8nm DUV is just a bump of their 10nm process and is inferior to TSMC 7N. If Samsung 7N EUV delivered, then Nvidia would be in a different position.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
With FP16 A100 can deliver 2x the efficiency of V100 but with FP32 is would be worse?

And why cant you accept that Samsung has started mass production of 8nm chips since the summer of 2018 which would have given nVidia enough time to optimize the process like with the 7nm process.
Considering that CUT DOWN to 108 SM's chip draws 400W of power, why do you even believe Nvidia had time to optimize the physical design for N7 process, especially considering that is bog standard N7 process, and not like a version of 16 nm TSMC that Nvidia and TSMC called 12 nm FFN, for marketing purposes?

Remember guys, GA100 chip has 108 SM's out of whole design having 128 SM's. How much will it draw power? 500W? And remember, this thing has HBM2! Where each stack draws 4W's of power. So we are talking about 20-30W for memory subsystem.

GDDR6 and GDDR6X will draw the same amount of power, but per memory chip!

Why is it so hard to believe, that RTX 3080 could use 300W of power, and RTX 3090 375W's, considering even the person who has got legit, correct info touts that 102 based GPUs draw anywhere between 300-375W of power?

Deal with it.
 
Reactions: french toast

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
To be fair: one of my sources said to me that 2080 Ti replacement, that is currently being worked on was drawing anywhere between 260-300W of power, depending on workloads.

The thing that I do not know was what clock speeds it was on. Also, considering the leak of 3080 Shroud I believe this GPU is closer to launch, on which I base my assumption that it is that 300W GPU talked here.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |