NVIDIA Pascal Thread

Page 40 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Genx87

Lifer
Apr 8, 2002
41,091
513
126
Emm,you mean Kepler?? The GK210 was used in those markets not the GM200.



Intel can afford to throw billions at the market - you only have to look at how they have managed to sneak into the top 10 list of the TOP500.

I thought they had a maxwell based Tesla and it wasnt that popular due to FP64 performance?
 

Det0x

Golden Member
Sep 11, 2014
1,264
4,020
136
Hmm. The new FirePro delivers a maximum single-precision performance of 13.9 teraflops, while Nvidia nextgen p100 delivers 10.6 teraflops ?

Are i am missing something ?
 

n0x1ous

Platinum Member
Sep 9, 2010
2,572
248
106
Hmm. The new FirePro delivers a maximum single-precision performance of 13.9 teraflops, while Nvidia nextgen p100 delivers 10.6 teraflops ?

Are i am missing something ?

Yes you Are. That is a dual Fiji part so 596mm2 X 2 GPUs that are single precision focused with minimal double precision hardware
 

Glo.

Diamond Member
Apr 25, 2015
5,802
4,776
136
Hmm. The new FirePro delivers a maximum single-precision performance of 13.9 teraflops, while Nvidia nextgen p100 delivers 10.6 teraflops ?

Are i am missing something ?

At the same power and thermal envelope. Nothing you are missing. Apart from the fact that S9300x2 has 1/16 ratio of Double Precision and P100 has 1/2 ratio of DP.

But in single precision AMD has better solution, unless you run CUDA software.
 

ThatBuzzkiller

Golden Member
Nov 14, 2014
1,120
260
136
GP100 will need much more parallelism compared to GM200 ...

49152 warps vs 114688 warps

In that respect Pascal is more closer to GCN the ever in terms of on chip memory resources and occupancy and it even has 64K of shared mem per SM (64 wide) just like GCN CU (64 wide) having 64K of shared mem! (The same ALU/shared mem ratio!)
 
Last edited:

nvgpu

Senior member
Sep 12, 2014
629
202
81
Nvidia is probably waiting for 8 stack HBM2 and better yields before enabling all 3840 CUDA cores on GP100 for a 32GB Tesla P100/P110 if a customer asks them for 32GB of HBM2 memory.
 

Erenhardt

Diamond Member
Dec 1, 2012
3,251
105
101
Yes you Are. That is a dual Fiji part so 596mm2 X 2 GPUs that are single precision focused with minimal double precision hardware

If by minimal you mean 4 times more than TitanX, then I need to study my English dictionary again.

Interesting, Titan lineup was introduced by a Kepler generation as a professional GPU for cheap (hell expensive for a geforce, third the price of pro card).
Maxwell made Titan lineup a glorified regular Geforce at an insane premium, tesla seems to continue that trend.
 

xpea

Senior member
Feb 14, 2014
451
153
116
Hmm. The new FirePro delivers a maximum single-precision performance of 13.9 teraflops, while Nvidia nextgen p100 delivers 10.6 teraflops ?

Are i am missing something ?
yes, these "good on paper" AMD FLOPS are most of the time "useless" because AMD has no resource to provide adequate support.
For example, if you want to do deep learning (the biggest trendy HPC market now), Nvidia provide lot of libraries and SDK like CuDNN when... you have nothing on AMD...
a good read :
http://timdettmers.com/2015/03/09/deep-learning-hardware-guide/
So what kind of GPU should I get? NVIDIA or AMD?
NVIDIA’s standard libraries made it very easy to establish the first deep learning libraries in CUDA, while there were no such powerful standard libraries for AMD’s OpenCL. Right now, there are just no good deep learning libraries for AMD cards – so NVIDIA it is. Even if some OpenCL libraries would be available in the future I would stick with NVIDIA: The thing is that the GPU computing or GPGPU community is very large for CUDA and rather small for OpenCL. Thus in the CUDA community good open source solutions and solid advice for your programming is readily available.

some pages on nvidia website about deep learning:
http://www.nvidia.com/object/machine-learning.html
https://developer.nvidia.com/deep-learning
https://developer.nvidia.com/deep-learning-resources
and thousands of pages/studies/courses/research papers/code source available on google

and on AMD :
nothing or close to nothing with some hobby projects that are too rough to be seriously considered...
 

n0x1ous

Platinum Member
Sep 9, 2010
2,572
248
106
If by minimal you mean 4 times more than TitanX, then I need to study my English dictionary again.

Interesting, Titan lineup was introduced by a Kepler generation as a professional GPU for cheap (hell expensive for a geforce, third the price of pro card).
Maxwell made Titan lineup a glorified regular Geforce at an insane premium, tesla seems to continue that trend.

minimal compared to Hawaii. The point was that comparing a dual GPU SP focused product vs a single GPU HPC/DP focused product and then saying the latter is not impressive because the former has more SP TFlops is a bad comparison to make.
 
Last edited:

Azix

Golden Member
Apr 18, 2014
1,438
67
91
yes, these "good on paper" AMD FLOPS are most of the time "useless" because AMD has no resource to provide adequate support.
For example, if you want to do deep learning (the biggest trendy HPC market now), Nvidia provide lot of libraries and SDK like CuDNN when... you have nothing on AMD...
a good read :
http://timdettmers.com/2015/03/09/deep-learning-hardware-guide/


some pages on nvidia website about deep learning:
http://www.nvidia.com/object/machine-learning.html
https://developer.nvidia.com/deep-learning
https://developer.nvidia.com/deep-learning-resources
and thousands of pages/studies/courses/research papers/code source available on google

and on AMD :
nothing or close to nothing with some hobby projects that are too rough to be seriously considered...

opencl is not "AMD's". Should basically be what everybody else uses. And clearly its more developed than you suggest since serious projects are using it.
 

NTMBK

Lifer
Nov 14, 2011
10,322
5,352
136
If by minimal you mean 4 times more than TitanX, then I need to study my English dictionary again.

Interesting, Titan lineup was introduced by a Kepler generation as a professional GPU for cheap (hell expensive for a geforce, third the price of pro card).
Maxwell made Titan lineup a glorified regular Geforce at an insane premium, tesla seems to continue that trend.

Not all professional workloads need double precision. Nvidia offered GK210 for DP heavy workloads, GM200 was for SP heavy workloads.
 

Det0x

Golden Member
Sep 11, 2014
1,264
4,020
136
yes, these "good on paper" AMD FLOPS are most of the time "useless" because AMD has no resource to provide adequate support.
For example, if you want to do deep learning (the biggest trendy HPC market now), Nvidia provide lot of libraries and SDK like CuDNN when... you have nothing on AMD...
a good read :
http://timdettmers.com/2015/03/09/deep-learning-hardware-guide/


some pages on nvidia website about deep learning:
http://www.nvidia.com/object/machine-learning.html
https://developer.nvidia.com/deep-learning
https://developer.nvidia.com/deep-learning-resources
and thousands of pages/studies/courses/research papers/code source available on google

and on AMD :
nothing or close to nothing with some hobby projects that are too rough to be seriously considered...

Hmm so lets me see if i understand you right then

The only advantage Nvidia possibly have , is on the software-side ?

It just seems really strange Nvidia next-gen 16nm finfet (HPC gpu only), cant match AMD's "old" 28nm tech (mainstream Fiji), in regards to single precision compute ?

I mean nevermind if they have to throw 2x the die-space to be able to compete (or even crush), as long as they have to operate within the same power and thermal envelope. (even at much lower cost)

@ n0x1ous

Yeah your right about the double precision, no comparison there :thumbsup:
 
Last edited:

Erenhardt

Diamond Member
Dec 1, 2012
3,251
105
101
opencl is not "AMD's". Should basically be what everybody else uses. And clearly its more developed than you suggest since serious projects are using it.

Not to mention those linked by him are in C and amd have tools to run C and even CUDA code.

Not all professional workloads need double precision. Nvidia offered GK210 for DP heavy workloads, GM200 was for SP heavy workloads.

Does it mean DP performance increase is solely production process related and we should expect to see a single generation od DP products for each xtor generation?
 

Samwell

Senior member
May 10, 2015
225
47
101
People here speculating about new Titan should better think, whether we will see GP100 as a consumer chip anyway. My bet is no, GP100 will never go to consumer. This chip is a clear pure HPC Chip. Seems like the big gamer chip will be GP102, which was seen in drivers along with the other pascal chips. Just forget GP100 for gaming. It won't ever come.
 

Kris194

Member
Mar 16, 2016
112
0
0
People here speculating about new Titan should better think, whether we will see GP100 as a consumer chip anyway. My bet is no, GP100 will never go to consumer. This chip is a clear pure HPC Chip. Seems like the big gamer chip will be GP102, which was seen in drivers along with the other pascal chips. Just forget GP100 for gaming. It won't ever come.

It won't because gamers don't need features such as NvLink
 

ThatBuzzkiller

Golden Member
Nov 14, 2014
1,120
260
136
Does the GP100 even feature many of the previous fixed function units ?

If not then it's absolutely worthless for gaming and in that case there will be a GP102 which will be much less impressive gaming successor to the GM200 ...
 

Qwertilot

Golden Member
Nov 28, 2013
1,604
257
126
Why less impressive? Surely a good thing if they're doing a gaming focused big die?

Means you'll get most/all of the (non trivial) benefits that maxwell got vs Kepler from being purely gaming/SP focused. Might not get HBM2 so early and things, but whatever.

Scary amount of engineering if they have split their product stack, but I guess the main writers were speculating about it with Maxwell.
 

ThatBuzzkiller

Golden Member
Nov 14, 2014
1,120
260
136
Why less impressive? Surely a good thing if they're doing a gaming focused big die?

Means you'll get most/all of the (non trivial) benefits that maxwell got vs Kepler from being purely gaming/SP focused. Might not get HBM2 so early and things, but whatever.

Scary amount of engineering if they have split their product stack, but I guess the main writers were speculating about it with Maxwell.

Means Nvidia will have to trade the SMs in order to get back the fixed function units which means lower theoretical compute performance in the end supposing that the GP100 didn't have those fixed function units to begin with ...
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |