Question Speculation: RDNA2 + CDNA Architectures thread

Page 81 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

uzzi38

Platinum Member
Oct 16, 2019
2,703
6,405
146
All die sizes are within 5mm^2. The poster here has been right on some things in the past afaik, and to his credit was the first to saying 505mm^2 for Navi21, which other people have backed up. Even still though, take the following with a pich of salt.

Navi21 - 505mm^2

Navi22 - 340mm^2

Navi23 - 240mm^2

Source is the following post: https://www.ptt.cc/bbs/PC_Shopping/M.1588075782.A.C1E.html
 

Glo.

Diamond Member
Apr 25, 2015
5,763
4,667
136
I love how things are in discussion twisted.

I was talking about 40 CU GPU being 10% above RTX 2080 Super or 35% above RX 5700 XT.

I've just checked the performance comparison with RTX 2080 Super in TPU suite. Well, if we look at TechPowerUp charts, RTX 2080 Super in 4K is 25% above RX 5700 XT.


Secondly. Thats the performance level that RTX 3070 will achieve, even according to Galax:

So lets get back to the discussion. How come 40 CU GPU chip cannot compete with RTX 3070, while clocked at 2.3 GHz and having 10% higher IPC than RDNA1 GPUs?

Can anyone explain this to me? From the start, Im trying to tell people that RTX 3070 WILL NOT ACHIEVE RTX 2080 Ti performance levels. I don't know why people believe in this, and spin the discussion that in order to compete with RTX 3070, 40 CU GPU has to beat RTX 2080 Ti.

Its absolutely ridiculous how overestimated Nvidia is, and their Ampere GPUs.
 
Last edited:

dzoni2k2

Member
Sep 30, 2009
153
198
116
Can anyone explain this to me? From the start, Im trying to tell people that RTX 3070 WILL NOT ACHIEVE RTX 2080 Ti performance levels. I don't know why people believe in this, and spin the discussion that in order to compete with RTX 3070, 40 CU GPU has to beat RTX 2080 Ti.

Nvidias marketing is probably just behind Apples in brainwashing power. Some people actually believe Nvidia more than independent benchmarks. They are still quoting ridiculous performance and perf/W claims from marketing slides that were proven to be complete horse*. It's quite mind boggling.
 

Glo.

Diamond Member
Apr 25, 2015
5,763
4,667
136
Nvidias marketing is probably just behind Apples in brainwashing power. Some people actually believe Nvidia more than independent benchmarks. They are still quoting ridiculous performance and perf/W claims from marketing slides that were proven to be complete horse*. It's quite mind boggling.
Simplest possible calculations.

RTX 3080 is 25-30% faster than RTX 2080 Ti.

RTX 3080 has 68 SMs, massive bandwidth.

So how come suddenly 44 CU GPU, will achieve RTX 2080 Ti performance, considering that RTX 3080 has 54%(!) more SMs? And does not use GDDR6X, but only GDDR6?

How will it mitigate the undeniable lack of hardware? Nvidia's magic?

So maybe that 40CU has way smaller hill to climb, despite what people want to believe?
 

HurleyBird

Platinum Member
Apr 22, 2003
2,726
1,342
136
And it never will achieve its potential.

Nvidia simply just GCN'd their gaming architecture. It behaves EXACTLY like GCN in games, and EXACTLY like GCN in compute. Mosterous in compute, mediocre in games, with insane inefficiency. The performance increase in compute is not reflected in gaming. For the same exact reasons why GCN never reflected its performance in games.

I'm not sure. GA100 is only designed for compute, but doesn't have the doubling of fp32 resources. fp32 doubling in Ampere seems to be aimed at gaming specifically.

But to play devil's advocate, it's obvious that Nvidia's implementation of fp32 doubling could be better for gaming too.

The 50% ratio between fp32 and int/fp32 hybrid resources seems arbitrary. How many games are going to utilize 50% int? One or two edge cases? Zero? And you can ask the same question about how many games will use 0% int, because if you use any int some of those extra fp32 transistors will go to waste.

No, what you want, and I wouldn't be surprised to see with Hopper, is some percentage of pure fp32 cores, some percentage of hybrid cores, and some percentage of pure int cores. Maybe something like 60%, 25%, and 15% respectively.
 

Saylick

Diamond Member
Sep 10, 2012
3,386
7,151
136
I'm not sure. GA100 is only designed for compute, but doesn't have the doubling of fp32 resources. fp32 doubling in Ampere seems to be aimed at gaming specifically.

But to play devil's advocate, it's obvious that Nvidia's implementation of fp32 doubling could be better for gaming too.

The 50% ratio between fp32 and int/fp32 hybrid resources seems arbitrary. How many games are going to utilize 50% int? One or two edge cases? Zero? And you can ask the same question about how many games will use 0% int, because any time you use any int some of those extra fp32 transistors are going to waste.

No, what you want, and I wouldn't be surprised to see with Hopper, is some percentage of pure fp32 cores, some percentage of hybrid cores, and some percentage of pure int cores. Maybe somewhere around 60%, 25%, and 15% respectively.
You've hit the nail on the head. The tricky part is that on a given clock cycle, it's hard to pinpoint how much INT and FP are used. Nvidia says that for every 100 FP operations, there's 36 INT operations, but that's clearly an average over some timeframe of the average workload. Some cycles the GPU doesn't need as much INT and on others it needs more INT. I'm not sure what the ideal balance of purely dedicated FP, INT, and FP/INT cores would be, but a logical approach seems to be one where you profile the typical workload and for each clock cycle, you determine the INT/FP ratio. Then you plot a histogram to see what the distribution looks like, and then you create a GPU where the hybrid cores covers like 1 or 2 standard deviations out from the mean INT/FP ratio is so that you can swing towards more INT or more FP in the vast majority of cases. A parametric study can be conducted to determine the optimal balance of perf to area of using more hybrid cores vs dedicated FP and INT cores.
 

Heartbreaker

Diamond Member
Apr 3, 2006
4,262
5,259
136
I'm not sure. GA100 is only designed for compute, but doesn't have the doubling of fp32 resources. fp32 doubling in Ampere seems to be aimed at gaming specifically.

It's really over-provisioned in FP32 for common gaming resolutions, since they didn't really double ROPs/memory BW and other functional units to support all those extra FP32 cores.

A couple of reviewers have noted that this is the reason why it does so much better at 4K (even when not CPU limited), FP32 needs increase disproportionately at higher resolutions. Probably why NVidia is marketing 8K with 3090.

While it is FP32 overkill, it is still efficient design. They took the INT32 unit and added FP32 capability, so there will be significant reuse of die area, like register and cache space. It's cheaper than doing a complete separate FP32 unit.
 
Reactions: FatherMurphy

Zstream

Diamond Member
Oct 24, 2005
3,396
277
136
It's odd but some people seem to want benchmarks done at:

4K for CPU tests.
Under 4K for GPU tests.

Which is precisely backwards to where the differences emerge.

Heck if I know, I frankly don't care anymore about lusting over hardware. I'm sure AMD will have plenty of SKU's to choose from, and compete just fine with Nvidia.
 

dzoni2k2

Member
Sep 30, 2009
153
198
116
Heck if I know, I frankly don't care anymore about lusting over hardware. I'm sure AMD will have plenty of SKU's to choose from, and compete just fine with Nvidia.

They sure set the bar low cheaping out on 7nm. Everyone at AMD is probably grinning from ear to ear right now. 1st Intel and the 10nm debacle now Nvidia with Fermi v2 on 8nm. Mobile is going to be an absolute massacre for Nvidia.
 

Qwertilot

Golden Member
Nov 28, 2013
1,604
257
126
Mobile is certainly going to be rather interesting, and much more market relevant than the very top cards.

I would be very surprised if it was a massacre. NV started genuinely significantly ahead & they did get the die shrink and do know how important the mobile market is.

Fairly confident they'll get a decent performance improvement out for mobile.

AMD might well get some market share back though.
 

TESKATLIPOKA

Platinum Member
May 1, 2020
2,429
2,914
136
I love how things are in discussion twisted.

I was talking about 40 CU GPU being 10% above RTX 2080 Super or 35% above RX 5700 XT.

I've just checked the performance comparison with RTX 2080 Super in TPU suite. Well, if we look at TechPowerUp charts, RTX 2080 Super in 4K is 25% above RX 5700 XT.


Secondly. Thats the performance level that RTX 3070 will achieve, even according to Galax:

So lets get back to the discussion. How come 40 CU GPU chip cannot compete with RTX 3070, while clocked at 2.3 GHz and having 10% higher IPC than RDNA1 GPUs?

Can anyone explain this to me? From the start, Im trying to tell people that RTX 3070 WILL NOT ACHIEVE RTX 2080 Ti performance levels. I don't know why people believe in this, and spin the discussion that in order to compete with RTX 3070, 40 CU GPU has to beat RTX 2080 Ti.

Its absolutely ridiculous how overestimated Nvidia is, and their Ampere GPUs.
RTX 2080 vs RTX 2080 Ti -> Ti is 26% faster
RTX 2080 vs RTX 3070 -> pretty much the same specs(SMs, TMUs, ROPs, Bandwidth) except the number of FP32 cores.
RTX 3070 is then only 14.4% faster than 2080, 8% faster than Super and 2080Ti is only 10% faster. Ok, this is possible, but Bandwidth is a bottleneck in my opinion, 16Ghz would bring extra performance.
Back to Navi2x with 40CU.
RX 5700XT has 1887Mhz clockspeed on average.
2.3GHz is actually 21.9% higher and with 10% IPC It's actually 100*1.1*1.219=134%, let's say 35% faster as a combination of clocks and IPC. This is not impossible, but the bigger problem for me is only 150W TBP and bandwidth If It's only 192bit. 192bit bus with 14-16GHz GDDR6 provides 336-384GB/s which is 14-25% less compared to 448GB/s of RX5700 XT while feeding a 35% faster chip. I would expect 256bit bus with 14-16GHz memory. TBP I won't comment, I did It more than once in the past.
 
Last edited:

TESKATLIPOKA

Platinum Member
May 1, 2020
2,429
2,914
136
Simplest possible calculations.

RTX 3080 is 25-30% faster than RTX 2080 Ti.

RTX 3080 has 68 SMs, massive bandwidth.

So how come suddenly 44 CU GPU, will achieve RTX 2080 Ti performance, considering that RTX 3080 has 54%(!) more SMs? And does not use GDDR6X, but only GDDR6?

How will it mitigate the undeniable lack of hardware? Nvidia's magic?

So maybe that 40CU has way smaller hill to climb, despite what people want to believe?
RTX 3070 does not have 44CU but 46SM.
 
Reactions: Konan

Timorous

Golden Member
Oct 27, 2008
1,727
3,152
136
RTX 2080 vs RTX 2080 Ti -> Ti is 26% faster
RTX 2080 vs RTX 3070 -> pretty much the same specs(SMs, TMUs, ROPs, Bandwidth) except the number of FP32 cores.
RTX 3070 is then only 14.4% faster than 2080, 8% faster than Super and 2080Ti is only 10% faster. Ok, this is possible, but Bandwidth is a bottleneck in my opinion, 16Ghz would bring extra performance.
Back to Navi2x with 40CU.
RX 5700XT has 1887Mhz clockspeed on average.
2.3GHz is actually 21.9% higher and with 10% IPC It's actually 100*1.1*1.219=134%, let's say 35% faster as a combination of clocks and IPC. This is not impossible, but the bigger problem for me is only 150W TBP and bandwidth If It's only 192bit. 192bit bus with 14-16GHz GDDR6 provides 336-384GB/s which is 14-25% less compared to 448GB/s of RX5700 XT while feeding a 35% faster chip. I would expect 256bit bus with 14-16GHz memory. TBP I won't comment, I did It more than once in the past.

I think 3070 vs 2080Ti will be like 5700XT vs R7. At 1080p and 1440p they might be very close but at 4k the 3070 will fall behind slightly in the comparison like 5700XT does.

PS5 is 2.23Ghz peak on the GPU. I can see a 40 CU RDNA2 gpu matching the 2080S. 12 Tflops of Series X is already getting a baseline of 2080 level performance in a really quick port.

The 5700XT has 225W tbp. When AMD compared navi to GCN they compared products not just the dies so on the basis they are doing that again it suggests they can hit 2080Ti performance at 225w tbp. It also suggests they could hit 5700XT perf with a 150W tbp. Given this it would mean that to hit 5700XT +32% the 40CU part would need to have a 200W tbp.

Looking at PS5 vs Series X power supplies it seems that it could go either way because the difference could be down the the PS5 SSD IO arrangement being more power hungry or the audio engine rather than SoC power. These power supplies are also undersized if the GPU in the SoC + Ram is consuming 200W on their own meaning if the performance is there AMD could have beaten their stated 50% perf/watt goal (I bet they had a higher internal target).
 

Antey

Member
Jul 4, 2019
105
153
116
just imagine a big vega (power consumption would be beyond insane though)... 80 Compute Units, 5120 stream processors, 2,3GHz... 5120*2*2,3= 23,55 TFLOPs. if radeon 7 is 13,44 TFLOPs then it would be 75% faster... that would be 26% faster than the 2080ti (based on that techpowerup relative perf data)... hey! not bad.
 
Reactions: spursindonesia

TESKATLIPOKA

Platinum Member
May 1, 2020
2,429
2,914
136
just imagine a big vega (power consumption would be beyond insane though)... 80 Compute Units, 5120 stream processors, 2,3GHz... 5120*2*2,3= 23,55 TFLOPs. if radeon 7 is 13,44 TFLOPs then it would be 75% faster... that would be 26% faster than the 2080ti (based on that techpowerup relative perf data)... hey! not bad.
Wrong. Vega TFLOPs don''t provide the same performance as RDNA1 or RDNA2 TFlops. Just compare Radeon VII vs RX 5700 XT.
 
Reactions: spursindonesia

TESKATLIPOKA

Platinum Member
May 1, 2020
2,429
2,914
136
What? I'm not talking about RDNA or RDNA2... Radeon 7 is GCN5 Vega.
I see, my bad, didn't notice you mentioning big Vega at the beginning, but even so It wouldn't be 75% faster. BTW I don't see a reason for bringing up a hypothetical Big Vega when we have RDNA2 Big Navi.
 
Last edited:

Antey

Member
Jul 4, 2019
105
153
116
Yes, RDNA has much better use of its resources and can provide a better gaming performance. you are saying that vega tflops and gaming performance doesnt scale linearly? (with pixel fillrate and bandwidth increased accordingly)
 

TESKATLIPOKA

Platinum Member
May 1, 2020
2,429
2,914
136
First of all, the question is If you also doubled the ROPs and significantly increased the bandwidth or just doubled the CU number in your calculation. Even If you doubled everything, then the scaling is never linear or at least I didn't see It with any GPU.

Edit: actually it is almost linear in 4K at least with RDNA1, If there is enough ROPs and bandwidth.
You can compare RX5500XT 8GB vs RX5700XT 8GB(+87% in TFLOPs and ROPs, bandwidth is doubled)
difference:
FullHD: +79%
WQHD: +83%
4K: +86%
Review
 
Last edited:
Reactions: Mopetar

Antey

Member
Jul 4, 2019
105
153
116
mmmh ok, i will try it.

5500XT has 22CU / 1408 SM at 1,845 MHz (boost max) or 1717 MHz for gaming freq; thats 5195 / 4835 GFLOPS
5700XT has 40CU / 2560 SM at 1905 MHz (boost max) or 1755 MHz for gaming freq; thats 9753 / 8986 GFLOPS

5500XT boost 100% -> 5700XT boost 188%
5500XT gaming 100% -> 5700XT gaming 186%


i'm going to use 4k data because is close to not CPU bottlenecked (isn't it?)

5500XT 4K gaming perf 100% -> 5700XT 4K gaming perf 186%

i dont know friend, that very linear to me.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |