There are rumored die shots of the GP106, but no hard data of any kind yet. In the Polaris 10 threads, we've seen this upcoming chip discussed as Nvidia's likeliest direct competitor. So, what can we expect from GP106?
The basic building block of Pascal chips appears to be the GPC cluster. This consists of 10 shader modules with 64 CUDA cores each. That means each GPC has 640 CUDA cores, and we can expect all Pascal chips to have an even multiple of this. Although this was never officially specified anywhere, it's likely there are also 16 ROPs per GPC, which would match what we know about the GP104 (64 ROPs total with four GPCs).
GTX 1070 uses a GP104 chip that has 25% of its shaders disabled (1920 active CUDA cores). Thus, we can assume that GP106 will be even smaller than that. The only logical assumption is that GP106 will have two GPCs, for 1280 CUDA cores and 32 ROPs. This would match what was done with Maxwell, where GM206 was almost exactly half of a GM204.
What does this mean in terms of performance? The GM206-based GTX 960 ran at about the same clock rate as the GM204-based GTX 980, and provides about 53% as much performance at 1080p. (TPU's latest ranking puts the 960 at 34% of a GTX 1080, and the 980 at 64%. 34 / 64 ~= 0.53.) This is very close to what we'd expect, given that GM206 has almost exactly half the resources of GM204. If we go with the assumptions outlined above, we could place a hypothetical GTX 1060 at 53% (with GTX 1080 at 100%). That's a bit lower than the GTX 970 (55%) and R9 390 (56%). But it should still be just powerful enough to qualify as "VR Ready" based on Steam's benchmarks, if Nvidia thinks that is a useful marketing point.
How about performance per watt? The GM206 actually had the worst perf/watt of any Maxwell chip. Peak gaming power was 118W for the GTX 960, compared to 184W for the GTX 980 (and GTX 1080). In other words, GTX 960 used 64% as much power as GTX 980 to get only 53% the performance. That is a roughly 17% reduction in perf/watt. Will we see the same thing for GP106? It depends. Some rumors say that GP106 has a 192-bit bus, and if that's the case, it will hurt perf/watt a bit. (They would get similar bandwidth and better perf/watt with a 128-bit GDDR5X bus, but that's likely not yet cost-effective.) Another question is whether Nvidia (and their customers) will be satisfied with only 970/390-level performance. Assuming the memory bus isn't a bottleneck, they could do better by upping the core clocks to 2000 MHz. That would probably outcompete an overclocked GTX 980, but at the cost of a substantial perf/watt hit. It might be difficult to keep it from throttling at a 150W power limit.
Overall, I expect GP106 to fall slightly behind Polaris in terms of both raw performance and perf/watt. It won't be a total blowout, but I think that it is going to be hard to justify the purchase of a GTX 1060 over a RX 480 (or whatever other SKUs that AMD comes out with). Of course, this won't stop Nvidia from selling a ton of them anyway. Sales of the GTX 960 were huge, even though for much of its life it was nearly as expensive as the far more powerful R9 290. (And surely not all of these sales were to people with crappy PSUs, 4K TVs with HDMI 2.0 only, or people who needed HEVC for whatever reason.)
Thoughts?
The basic building block of Pascal chips appears to be the GPC cluster. This consists of 10 shader modules with 64 CUDA cores each. That means each GPC has 640 CUDA cores, and we can expect all Pascal chips to have an even multiple of this. Although this was never officially specified anywhere, it's likely there are also 16 ROPs per GPC, which would match what we know about the GP104 (64 ROPs total with four GPCs).
GTX 1070 uses a GP104 chip that has 25% of its shaders disabled (1920 active CUDA cores). Thus, we can assume that GP106 will be even smaller than that. The only logical assumption is that GP106 will have two GPCs, for 1280 CUDA cores and 32 ROPs. This would match what was done with Maxwell, where GM206 was almost exactly half of a GM204.
What does this mean in terms of performance? The GM206-based GTX 960 ran at about the same clock rate as the GM204-based GTX 980, and provides about 53% as much performance at 1080p. (TPU's latest ranking puts the 960 at 34% of a GTX 1080, and the 980 at 64%. 34 / 64 ~= 0.53.) This is very close to what we'd expect, given that GM206 has almost exactly half the resources of GM204. If we go with the assumptions outlined above, we could place a hypothetical GTX 1060 at 53% (with GTX 1080 at 100%). That's a bit lower than the GTX 970 (55%) and R9 390 (56%). But it should still be just powerful enough to qualify as "VR Ready" based on Steam's benchmarks, if Nvidia thinks that is a useful marketing point.
How about performance per watt? The GM206 actually had the worst perf/watt of any Maxwell chip. Peak gaming power was 118W for the GTX 960, compared to 184W for the GTX 980 (and GTX 1080). In other words, GTX 960 used 64% as much power as GTX 980 to get only 53% the performance. That is a roughly 17% reduction in perf/watt. Will we see the same thing for GP106? It depends. Some rumors say that GP106 has a 192-bit bus, and if that's the case, it will hurt perf/watt a bit. (They would get similar bandwidth and better perf/watt with a 128-bit GDDR5X bus, but that's likely not yet cost-effective.) Another question is whether Nvidia (and their customers) will be satisfied with only 970/390-level performance. Assuming the memory bus isn't a bottleneck, they could do better by upping the core clocks to 2000 MHz. That would probably outcompete an overclocked GTX 980, but at the cost of a substantial perf/watt hit. It might be difficult to keep it from throttling at a 150W power limit.
Overall, I expect GP106 to fall slightly behind Polaris in terms of both raw performance and perf/watt. It won't be a total blowout, but I think that it is going to be hard to justify the purchase of a GTX 1060 over a RX 480 (or whatever other SKUs that AMD comes out with). Of course, this won't stop Nvidia from selling a ton of them anyway. Sales of the GTX 960 were huge, even though for much of its life it was nearly as expensive as the far more powerful R9 290. (And surely not all of these sales were to people with crappy PSUs, 4K TVs with HDMI 2.0 only, or people who needed HEVC for whatever reason.)
Thoughts?