OK, I'm lost. What is a 3840CC Maxwell?A 2560 CC Pascal is keeping up with a 3840 CC Maxwell, looks like clock per clock Pascal's overall throughput is much better than Maxwell.
OK, I'm lost. What is a 3840CC Maxwell?A 2560 CC Pascal is keeping up with a 3840 CC Maxwell, looks like clock per clock Pascal's overall throughput is much better than Maxwell.
OK, I'm lost. What is a 3840CC Maxwell?
How reliable is the 2560CC in GP104? An error there throws everything off.Lol you're right, I had gloss over it.
980Ti is 2816 CC.
Titan X is 3072.
55% higher clocks vs 980Ti.
25% higher perf.
~10% less cores.
Er, there's actually no IPC gains.
What the heck.
There's even a regression based on 3dMark. !!!
Edit: Hmm, perhaps adding fine-grained preemption to the uarch along with graphics <-> compute instant context switch features may hurt IPC in game engines that don't benefit from it, such as 3dMark. Becoming GCN-like for better compute + graphics workloads has a cost associated, and it looks to be increased TDP and poorer IPC in older games. Thoughts?
I think 1070 will end up being the star of the show here. NV will probably charge a hefty premium for the 1080 with GDDR5X being an excuse to bump up the price more. Anyone on a 980 Ti already that wants bang for their buck should probably wait for the true high end Pascal cards.
980 was a bad value relative to the 970, and ended up being completely obliterated by the 980 Ti for not much more taking launch prices into account, expect it'll be the same deal here.
As a GTX 980 owner looking to add a bit more GPU kick for running games in 4k, is the 1080 going to be worth the upgrade do you think? Or would it be worth waiting for 2nd gen FinFET GPUs with stacked mem?
I've only recently come back to the GPU discussion so I'm behind on a lot of this and 113 pages of thread is too much info to scan through.
A 2560 CC Pascal is keeping up with a 3840 CC Maxwell, looks like clock per clock Pascal's overall throughput is much better than Maxwell.
Hmm, perhaps adding fine-grained preemption to the uarch along with graphics <-> compute instant context switch features may hurt IPC in game engines that don't benefit from it, such as 3dMark. Becoming GCN-like for better compute + graphics workloads has a cost associated, and it looks to be increased TDP and poorer IPC in older games. Thoughts?
What games do you think we should expect huge gains vs 980 Ti in with the better compute power of 1080? Quantum Break and Hitman instead of Crysis 3 and Witcher 3?
Any game that has a higher % of the rendering using compute shaders.
Quantum Break definitely, expect a massive performance leap for the 1080 vs 980Ti.
It's not better compute power.
It's the graphics <-> compute switch that the GPU has to go through currently. If you're rendering graphics and the game calls for a compute workload, it has to flush everything and wait for full idle before starting the compute work. This is a slow process which leads to a performance penalty.
GCN can do this instantly and according to NV's paper, Pascal can too.
Moreover if giving up 5% IPC allows you to modify the architecture and increase clock by 20% then you take it..
That is a very risky design strategy. On the CPU front, both Intel (Netburst) and AMD (Bulldozer) tried that, and both efforts were miserable failures - the clock gains weren't nearly as high as hoped, and the IPC losses were worse than expected.
Maybe it works better on GPUs, but it has the potential to fail hard if anything is out of place even a little.
That is a very risky design strategy. On the CPU front, both Intel (Netburst) and AMD (Bulldozer) tried that, and both efforts were miserable failures - the clock gains weren't nearly as high as hoped, and the IPC losses were worse than expected.
Maybe it works better on GPUs, but it has the potential to fail hard if anything is out of place even a little.
At the expense of perf/watt.Decreasing the core count and combine them with a higher clock helps for a better utilizing of the architecture.
At the expense of perf/watt.
Depends on the workload.Maxwell has a better perf/watt than Kepler and GCN.
Are you really trying to convince us that increasing frequency (linear increase in perf, exponential increase in power) is the way to increase efficiency?Maxwell has a better perf/watt than Kepler and GCN.
GPUs are not designed for latency. They are designed for throughput. There is no "IPC" in the same context.
Higher clocks have huge benefits: Better geometry performance, better pixel performance (ROPs), better L2 cache bandwidth etc.
Decreasing the core count and combine them with a higher clock helps for a better utilizing of the architecture.
Every process has a different sweet spot in efficiency.Are you really trying to convince us that increasing frequency (linear increase in perf, exponential increase in power) is the way to increase efficiency?
Hey everybody, Guru3D has landed in Texas for two days of briefings with Nvidia.
http://vrworld.com/2016/04/25/16nm-nvidia-geforce-gtx-1080-leak-ahead-computex-taipei-2016/Lol you're right, I had gloss over it.
980Ti is 2816 CC.
Titan X is 3072.
55% higher clocks vs 980Ti.
25% higher perf.
~10% less cores.
Er, there's actually no IPC gains.
What the heck.
There's even a regression based on 3dMark. !!!
Edit: Hmm, perhaps adding fine-grained preemption to the uarch along with graphics <-> compute instant context switch features may hurt IPC in game engines that don't benefit from it, such as 3dMark. Becoming GCN-like for better compute + graphics workloads has a cost associated, and it looks to be increased TDP and poorer IPC in older games. Thoughts?
How come people did not seen this even if it was on previous pages, already?VRZone said:While the specs of GP104 are still hidden, we can now say that the chip packs 1920 CUDA Cores and a 256-bit controller supports both GDDR5 and GDDR5X (MSI uses GDDR5X) memory.
Maxwell has a better perf/watt than Kepler and GCN.
http://vrworld.com/2016/04/25/16nm-nvidia-geforce-gtx-1080-leak-ahead-computex-taipei-2016/How come people did not seen this even if it was on previous pages, already?
1920 CUDA core GPU with 1.86 GHz core clock, 10 GHz GDDR5X memory is 25% faster than reference GTX 980 Ti, and 4% faster than GTX 980 Ti Waterforce Gaming.
It also has around 1 TFLOPs of compute power more than reference GTX 980 Ti.
TDP between 165 and 200W. Because of extremely high core clocks it will be closer to 200W.