Question 'Ampere'/Next-gen gaming uarch speculation thread

Page 141 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Ottonomous

Senior member
May 15, 2014
559
292
136
How much is the Samsung 7nm EUV process expected to provide in terms of gains?
How will the RTX components be scaled/developed?
Any major architectural enhancements expected?
Will VRAM be bumped to 16/12/12 for the top three?
Will there be further fragmentation in the lineup? (Keeping turing at cheaper prices, while offering 'beefed up RTX' options at the top?)
Will the top card be capable of >4K60, at least 90?
Would Nvidia ever consider an HBM implementation in the gaming lineup?
Will Nvidia introduce new proprietary technologies again?

Sorry if imprudent/uncalled for, just interested in the forum member's thoughts.
 

moonbogg

Lifer
Jan 8, 2011
10,637
3,095
136
How many 1080ti owners are going to see that buffer downgrade and take a pause on the 3080? I would, but I already sold my 1080 ti.

How many large vram cards with weak GPUs have been built for suckers in the past?

I would have committed to buying a 3080 by now, honestly, if it weren't for the downgrade in Vram. It bothers me. It's the reason I'm not clicking the buy button on day one. As for large buffers for suckers, I agree completely. We've all seen those cards. The argument doesn't apply to the 3080 IMO. People will be using the card for high resolutions, texture mods, ray tracing and future generation games. 8Gb has been breached in some cases, even at 1440p. The 3080 would have been a good fit for 16Gb. They could have got away with 12 even.
I don't care about quantities not matching the memory architecture or whatever. It's not my job to be a GPU engineer and base my purchasing decisions on that kind of knowledge. I base it on the simple fact that an upgrade should be an upgrade, especially one priced at $700.

I need to be fair though. Historically, GPU releases have often come with the same Vram as the last gen and it wasn't an issue. I feel this time is an issue.
 

uzzi38

Platinum Member
Oct 16, 2019
2,702
6,405
146
It was quite disappointing that it was not 7LPP. Ampere could have stretched its legs.
There could be something wrong with 7LPP for bigger dies.
Also this means there is no quick re-tapeout path to the 6LPP/5LPE for later refreshes.
Well you know, GA103 is still apparently around, maybe, just maybe we'll see something interesting
 

CP5670

Diamond Member
Jun 24, 2004
5,527
604
126
It will be interesting to see what the AIB versions of these cards look like. Usually the AIB cards have significantly better coolers (as well as stock clocks), but in this case the stock Nvidia coolers are already quite beefy. I wonder how much of an improvement the AIB cards will be.
 

JasonLD

Senior member
Aug 22, 2017
486
447
136
It was quite disappointing that it was not 7LPP. Ampere could have stretched its legs.
There could be something wrong with 7LPP for bigger dies.
Also this means there is no quick re-tapeout path to the 6LPP/5LPE for mid gen Super refreshes.

I do expect Ampere refresh on either 7nm or 6nm. I don't expect 5nm until Hopper.
Though looking at pricing, I think Nvidia did get some really good prices on those 8nm.

That said, 3080 is looking lot more impressive than I thought it would end up being.
 
Reactions: ozzy702

Ajay

Lifer
Jan 8, 2001
16,094
8,106
136
Nvidia are no doubt getting a good deal.

AIBs? I feel so bad for them. Margins must be awful having to deal with GDDR6X in all it's cursedness. Power draw per module is notably up, but worsed of all is the PAM4 signalling. The extra PCB costs alongside the cost per module is not fun.
I don't know why to you think PAM4 is all that bad - it's pretty common in other areas of electronics. The controller shouldn't be that difficult to design, toughest part I see is keeping the signals clean.

It was quite disappointing that it was not 7LPP. Ampere could have stretched its legs.
There could be something wrong with 7LPP for bigger dies.
Also this means there is no quick re-tapeout path to the 6LPP/5LPE for mid gen Super refreshes.
Or maybe 6LPP is not part of wafer shuttle arrangement?
Samsung has to come out publicly in their Foundry Forums. Too much rampant FUD from Taiwan.
I think the defect rate must has been pretty bad for large dice. If Samsung keeps improving their 8nm (NVIDIA special), then they may be able to turn the clock up in another year.
I suppose it depends on whether AMD plans on coming out with RNDA3 on 5nm in another 12-18 months after RDNA2.
 

MrTeal

Diamond Member
Dec 7, 2003
3,585
1,743
136
Has there been any information on whether RTX IO is interface agnostic? There's obvious performance differences between a top 4.0 NMVE drive and a SATA drive, but will a PCIe based drive be required to get good performance out of it?
 

uzzi38

Platinum Member
Oct 16, 2019
2,702
6,405
146
I don't know why to you think PAM4 is all that bad - it's pretty common in other areas of electronics. The controller shouldn't be that difficult to design, toughest part I see is keeping the signals clean.

12 layer PCBs.

Reduced distances between GPU die and memory.

Failure to reach 21Gbps as GDDR6X is supposed to allow at launch.

There's already several signs that GDDR6X has major limitations. Add onto the fact that between data travelling and module power you're looking at notably over 100W on the 3090 and yes, the stuff is absolutely cursed.

Just because Nvidia didn't want to have to take on the HBM tax on themselves. But they don't need to worry, as AIBs can deal with all the extra costs. No biggie for them
 
Reactions: USER8000 and Glo.

DiogoDX

Senior member
Oct 11, 2012
746
277
136
Its the slide they mentioned 1.9x perf/watt and according to official NVIDIA slide,

Turing 60fps / 240w = 0.25
vs
Ampere 105fps / 320W = 0.33

This is 32% higher perf/watt , again this is according to official NVIDIA slide.
2080S is like 20% better perf/watt that a 5700XT. So if AMD delivers the 50% better perf/watt with RDNA2 they will really catch Nvidia.
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,864
3,418
136
RTX IO seems interesting. Does that erase the big advantage of PS5/XBSX over pc? Will big navi have something similar?
Yes, but no and no.
So there is a lot more to File I/O like security etc. So it probably wont be as good as the Xbox system for equal parts. That said we can probably easily just brute force that on the PC side so xbox level or greater I/O I would expect to be quite achievable . PS5 on the other hand seems to be another level entirely that i dont know if brute force will be able to catch.
 

JasonLD

Senior member
Aug 22, 2017
486
447
136
Nvidia is calculating the perf/W improvement by using the max performance part of the Turing curve and comparing it to around the max efficiency portion of Ampere. Creative thinking by the marketing team & fooling a lot of people.

That is how the perf/w improvement is calculated on marketing slide, since it is easy to inflate that figure depends on how it is being compared. Reason I am not being so hot on 50% Perf/w claim on RDNA2 side.

Beside that, I think this is a pretty good win for Samsung foundry side since this might be the first time they scored big GPU dies.
 
Reactions: FatherMurphy

Thala

Golden Member
Nov 12, 2014
1,355
653
136
Nvidia is calculating the perf/W improvement by using the max performance part of the Turing curve and comparing it to around the max efficiency portion of Ampere. Creative thinking by the marketing team & fooling a lot of people.

Technically there is nothing wrong with specifying perf/W at iso performance levels - thats what they did. Of course they could have added the iso-power number, which naturally is lower and then peak-to-peak, which would be even lower again. But its all in the graph, its not that NVidia is hiding something.
 

moonbogg

Lifer
Jan 8, 2011
10,637
3,095
136
I'm having confused feelings about the 3080. I think I like it. I wish it had more Vram, but I'm guessing 10Gb might be fine for 3440x1440. Even I can admit when I'm being too damn picky. It looks like a great card at $700 and should lay waste to anything thrown at it for quite a long time. I do suspect an impressive refresh may be in the works, or a 3080Ti, but who cares. Waiting another year doesn't sound like fun, but playing around with a new card does sound like fun. I don't think I'll be able to resist taking a hammer to the "buy" button and smashing it into the ground on release day. Jacket man delivered.
 

DiogoDX

Senior member
Oct 11, 2012
746
277
136
That is how the perf/w improvement is calculated on marketing slide, since it is easy to inflate that figure depends on how it is being compared. Reason I am not being so hot on 50% Perf/w claim on RDNA2 side.

Beside that, I think this is a pretty good win for Samsung foundry side since this might be the first time they scored big GPU dies.
They said the same with RDNA vs VEGA and delivered (5700XT vs Vega64 or 5700 vs Vega 56). But we know that Vega was terrible in perf/watt so was "easy". If they deliver again, 50% over RDNA1 will put RDNA2 at the same perf/watt that Ampere.

 
Reactions: Mopetar

A///

Diamond Member
Feb 24, 2017
4,352
3,155
136
I'll be waiting it out since I don't plan to build any sooner than 5 or 6 months. I'll have a goo idea of where AMD is then but as I've said before, I expect RDNA2 to fall flat on its face. If nv wasn't bs'ing, these are some of the best cards they've put out in a long time. I missed the event and couldn't believe some of the prices I'd seen while wading through online articles that went up in near live time. If the 3060 ends up being around 275-350, nv will cover all their bases and there really won't be a point in team red's product, especially given the expected letdown with faulty software.
 

raghu78

Diamond Member
Aug 23, 2012
4,093
1,475
136
Its 2xFP32 THROUGHPUT.

Its the same marketing gibberish they used for A100 chips.

In reality, when tested actual, native FP32 performance, it was the same, as V100.

Nvidia have provided specs for RTX 3090, RTX 3080 and RTX 3070

RTX 3090 - 10496 cuda cores

RTX 3080 - 8704 cuda cores

RTX 3070 - 5888 cuda cores

But as always the devil is in the details and actual benchmarks in games should give us an idea. From my estimates based on Nvidia benchmarks and the digitalfoundry preview the RTX 3080 is 30% faster than RTX 2080 Ti. But the TDP is 23% higher. So the perf/watt improvement is just 5.6% . RTX 3070 is looking to be slightly faster than the RTX 2080 Ti for 40w less power. So the perf/watt increase is roughly 20-25%.
 

tomli747

Junior Member
Sep 1, 2020
1
2
51
Yeah, I agree with Glo here. They used the term "Shader-FLOPS" not "FP32 FLOPS". The INT cores actually do single-precision math (i.e. 32-bit) but just not floating point specifically. Going off of the SM diagram for A100, they don't list the INT cores as being capable of doing FP math either so my guess is either Nvidia tuned Ampere for graphics so that the the pipelines for an SM are (2) x 16-wide FP or 16-wide INT + 16-wide FP, or they are just listing Shader-FLOPS as a catch all term for all the concurrent single-precision math the entire GPU can do, INT and FP included. My money is on the latter.

My speculation is that the ALUs are running at doubled speed just like Fermi as Jensen said "Ampere does 2 shader calculations per clock". (i.e. 3070 has 2944 core @3.4GHz)

In this case, a CUDA core at doubled speed is advertised as 2 CUDA cores.
It probably also explains the high power consumption and the crazy numbers of core count for its die size and transistor density.
 

lobz

Platinum Member
Feb 10, 2017
2,057
2,856
136
Nvidia have provided specs for RTX 3090, RTX 3080 and RTX 3070

RTX 3090 - 10496 cuda cores

RTX 3080 - 8704 cuda cores

RTX 3070 - 5888 cuda cores

But as always the devil is in the details and actual benchmarks in games should give us an idea. From my estimates based on Nvidia benchmarks and the digitalfoundry preview the RTX 3080 is 30% faster than RTX 2080 Ti. But the TDP is 23% higher. So the perf/watt improvement is just 5.6% . RTX 3070 is looking to be slightly faster than the RTX 2080 Ti for 40w less power. So the perf/watt increase is roughly 20-25%.
Well, the phrase '2x throughput' was on an official NVIDIA slide during the presentation, so...

Don't forget, Zen2 has also 2x FP throughput compared to Zen1. That does NOT translate to 100% performance gain at all.
 

FatherMurphy

Senior member
Mar 27, 2014
229
18
81
I'm just a hobbyist and gamer, so no expert. But, I think it is safe to say that Nvidia didn't get rid of the separate INT unit and concurrent INT/FP performance. It was a big factor in the Turing uplift in "IPC". At the same time, Nvidia admitted at the time that for every 100 FP operations, there were ~ 35 INT operations.

It makes sense that Nvidia would create a GA102 with more flexible INT unit per SM so that it isn't sitting there idle. So, now you have FP + INT/FP, where you can do FP + INT or FP + FP, thus "twice" the throughput on FP32.

Given the GA100 was tailored more for HPC/AI with an emphasis on either higher precision (FP64) or lower precision (TF32 and lower), it is not a FP32 monster. GA102 is designed to be the FP32 monster to be used in workstations/servers where traditional FP32 is important. If I'm not mistaken, Nvidia has targeted its sub-Gx100 chips (e.g. TU102, 104, GP104) like this before. So, Nvidia took the transistors it saved by stripping out GA100s HPC specific parts and invested them in a more robust/hyrbid FP/INT unit. This helps in gaming (with somewhat limited gains because the rest of the graphics pipeline isn't doubled) but also gives Nvidia a FP32 (and raytracing) beast to sell to those industries who need it.

I think the CUDA core marketing could be clearer (if this is what has happened).

All of this is uneducated speculation.
 
Reactions: xpea and Gideon

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
2080S is like 20% better perf/watt that a 5700XT. So if AMD delivers the 50% better perf/watt with RDNA2 they will really catch Nvidia.

Well, if you take same tiers NAVI 10 has higher perf/w vs Turing

RX5600Xt higher perf/w vs RTX2060
RX5700 higher perf/w vs 2060S

RX5700XT is way too pushed in order to be faster vs RTX2070 but even then its very close

RTX2070 Super and RTX2080/S are using the much much bigger TU104




 

ddarko

Senior member
Jun 18, 2006
264
3
81
Since there will be no preorders for the Founders Edition cards this time, has the Nvidia store been the only place to get the cards on launch day in the past or have they also been available at retailers like Amazon, Best Buy, Microcenter etc.? Or do the FE cards only show up later at retailers?
 

Karnak

Senior member
Jan 5, 2017
399
767
136
You guys remember Jensen by the end of last year?
During GTC 2019 in China, NVIDIA founder and CEO Jensen Huang addressed questions from the tech press over who will be building out most ogf the new 7nm GPUs in 2020 and beyond. Huang said that most of the 7nm GPU production will be done by TSMC while Samsung will only handling a small portion of 7nm GPU production for NVIDIA.

Still possible though if GA106 and smaller ones will be on TSMC N7. But for now that looks weird to me since GA102 and GA104 are on Samsung's 8nm...
 

linkgoron

Platinum Member
Mar 9, 2005
2,334
857
136
Given Nvidia's recent GPU history, I wonder if GA104 failed to deliver the expected performance, and the 3080/3090 cards were released earlier because of this. This is the first time since 2013 that the 80s card is not a X04 - 980 was GM204 in 2014, 1080 was GP104 in 2016 and 2080 was TU104 in 2018. The 780 was GK110 in 2013 (Technically, the 680 was also GK104). In addition, everything seems over-the-top with the cards (cooling and TDP).

This is obviously pending reviews, but I wonder if the 3070 just failed to deliver the required performance for an 80 class card (i.e. faster than the previous gen's ti). According to Nvidia it's just on-par with the 2080ti, while even the 2080 was around 10% faster than the 1080ti (and it was also released alongside the 2080ti).


You guys remember Jensen by the end of last year?

Still possible though if GA106 and smaller ones will be on TSMC N7. But for now that looks weird to me since GA102 and GA104 are on Samsung's 8nm...

Samsung is 8nm, they're technically not making any 7nm cards.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |