Zen APUs made by GloFo, 14nm FinFET node, and packaged by Amkor

Page 5 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Glo.

Diamond Member
Apr 25, 2015
5,802
4,773
136
Polaris is the name of architecture. So... which one Polaris you are talkin' 'bout, Shintai?

I think there is very real chance that bigger Polaris will have HBM1.
 
May 11, 2008
20,145
1,149
126
Polaris is the name of architecture. So... which one Polaris you are talkin' 'bout, Shintai?

I think there is very real chance that bigger Polaris will have HBM1.

Fury has 4 HBM stacks of 1GB each. Perhaps Polaris 11 (?) will have 6 of those HBM stacks. Eight HBM stacks may be a little to much, thinking of all those traces that needs to be layout on the interposer.

Which will arrive first ? Polaris 10 (little polaris) or polaris 11 (Big polaris).
I seem to recall that AMD already had polaris 10 silicon working end of 2015.
Polaris 10 could be outfitted with GDDR5 for cost reasons.
 
Last edited:

ShintaiDK

Lifer
Apr 22, 2012
20,378
145
106
Polaris is the name of architecture. So... which one Polaris you are talkin' 'bout, Shintai?

I think there is very real chance that bigger Polaris will have HBM1.

The name of the architecture is GCN 4, also called GCN 1.3 in layman terms.
 

Yuriman

Diamond Member
Jun 25, 2004
5,530
141
106
The name of the architecture is GCN 4, also called GCN 1.3 in layman terms.

I wasn't aware that AMD actually designated different GCN versions internally - this was something review sites have tacked on to differentiate different feature sets in newer chips.

Please correct me if I'm wrong.
 

tenks

Senior member
Apr 26, 2007
287
0
0
Fury has 4 HBM stacks of 1GB each. Perhaps Polaris 11 (?) will have 6 of those HBM stacks. Eight HBM stacks may be a little to much, thinking of all those traces that needs to be layout on the interposer.

Which will arrive first ? Polaris 10 (little polaris) or polaris 11 (Big polaris).
I seem to recall that AMD already had polaris 10 silicon working end of 2015.
Polaris 10 could be outfitted with GDDR5 for cost reasons.

Polaris 10 is the bigger gpu meant for desktops
Polaris 11 is the low power/mobile gpu
 

ShintaiDK

Lifer
Apr 22, 2012
20,378
145
106
I wasn't aware that AMD actually designated different GCN versions internally - this was something review sites have tacked on to differentiate different feature sets in newer chips.

Please correct me if I'm wrong.

Review sites uses the term GCN 1.x. AMD documentation and developer tools uses GCN 1 to 4.

GCN 1.3/GCN4 is equal to NVidia Pascal. While Polaris 10/11 is equal to something like GP106/104.
 

tenks

Senior member
Apr 26, 2007
287
0
0
Review sites uses the term GCN 1.x. AMD documentation and developer tools uses GCN 1 to 4.

GCN 1.3/GCN4 is equal to NVidia Pascal. While Polaris 10/11 is equal to something like GP106/104.

perfect comparison, money.
 

JDG1980

Golden Member
Jul 18, 2013
1,663
570
136
You need volume for it to be meaningful. If AMD can only sell a handful og HBM APUs, it doesn't matter how much they can charge.

An Apple design win alone could provide adequate volume. The first-generation Intel Iris Pro was basically a specialty product made for the 15" rMBP.

There must be a demand for 250-300$ APUs before it is possible for AMD to actually release one. Before AMD can ask such prices, the APUs need a complete overhaul. After that they are no longer targeted for customers in 3rd world / developing countries, but the same customers as Intel Iris Pro.

Before any of that can happen, Zen must first succeed.

Yes, if Zen were to flop, then none of this matters. But I don't think it will. I expect roughly Sandy Bridge levels of IPC with clock speeds of 3.0-4.0 GHz on desktop parts (depending on core count and other factors) and lower clocks on laptop and server parts as usual. It's one thing to be a bit cautious about what AMD can do, it's another to not think they can match an Intel design from 2011 with a 2016 clean-sheet release. The fact that matching a 2011 Intel design will be reasonably competitive shows how much the market has stagnated in the last 5 years due to the lack of meaningful competition.

It would be a new chip design. And not only would Zen have to deliver. Polaris would have to as well. And dont quote PR numbers, they are meaningless.

There's no reason to think Polaris won't deliver. AMD's current GPUs are much more competitive than their CPUs are. They don't have to deliver a huge change with Polaris to have a solid, successful product. Just the gains from going to 14LPP will be huge.

Mind you all we already have APU with 7870 IGP since late 2013. Don't act like it is something new and beyond reach

And that's on a 28nm process. 14LPP allows roughly 2.2x the transistor density. The Playstation 4 APU has eight Jaguar cores and a GPU only slightly below full Pitcairn (18 CUs compared to Pitcairn's 20, with the same 32 ROPs and 256-bit GDDR5 memory bus). Go down to 14LPP and replace the big GDDR5 controller with the smaller HBM2 controller, and you could almost certainly fit in eight Zen cores plus 40 GCN CUs (2560 shaders, the same count as R9 290/390) and 64 ROPs. With execution resources this wide, you wouldn't have to crank up the CPU or GPU clock speeds too high to get competitive performance, thus saving power. (CPU clocks would of course have a turbo feature for lightly-threaded apps.).
 

jpiniero

Lifer
Oct 1, 2010
15,108
5,665
136
A 200W APU? Come on, be realistic.

An Apple design win alone could provide adequate volume. The first-generation Intel Iris Pro was basically a specialty product made for the 15" rMBP.

Apple is probably the only OEM that would be interested in a premium priced APU, but I doubt they want anything to do with GloFo now after they screwed up a second time.
 

Yuriman

Diamond Member
Jun 25, 2004
5,530
141
106
Yes, if Zen were to flop, then none of this matters. But I don't think it will. I expect roughly Sandy Bridge levels of IPC with clock speeds of 3.0-4.0 GHz on desktop parts (depending on core count and other factors) and lower clocks on laptop and server parts as usual. It's one thing to be a bit cautious about what AMD can do, it's another to not think they can match an Intel design from 2011 with a 2016 clean-sheet release. The fact that matching a 2011 Intel design will be reasonably competitive shows how much the market has stagnated in the last 5 years due to the lack of meaningful competition.

I'd argue that matching 2011 Intel IPC is only relevant if they also match 2016 Intel performance per watt. They can be behind on one metric, but not on most/all.
 

el etro

Golden Member
Jul 21, 2013
1,581
14
81
3D stacking mastering(good yields, good prices) will change everything on the RAM world.
 

Xpage

Senior member
Jun 22, 2005
459
15
81
www.riseofkingdoms.com
i would hope for a HBM2 APu with memory built in, no system memory needed. Saves board space, simplifies for OEMs, lets AMD control how much RAM is present to prevent gimping by single channel solutions,etc.

Cheap OEM products can have more disabled or bad HBM stacks. Price premium would be worth it for laptop or NUC like devices. I also hope they get an apple win to drive down costs so they can offer OEMs something like this.

This all depends on Zen and HBM2 availability and cost.
 

Shehriazad

Senior member
Nov 3, 2014
555
2
46
i would hope for a HBM2 APu with memory built in, no system memory needed. Saves board space, simplifies for OEMs, lets AMD control how much RAM is present to prevent gimping by single channel solutions,etc.

Cheap OEM products can have more disabled or bad HBM stacks. Price premium would be worth it for laptop or NUC like devices. I also hope they get an apple win to drive down costs so they can offer OEMs something like this.

This all depends on Zen and HBM2 availability and cost.

And you wouldn't be able to fix bad ram and would have to replace the entire thing whenever a single "component" breaks down!

Golden! This would be a win for OEMs AND AMD....let's do this boys! Fk the consumer!!!!
 

cbn

Lifer
Mar 27, 2009
12,968
221
106
Perhaps if having two smaller dies (CPU SoC die and GPU die) vs. one large "one size fits all die" allows better yields or advantages in mix and match binning it could help justify the cost of the interposer for consumer APUs?

No, the Interposer solution would never work due to cost. Even at say 5$. Remember all the lower parts that would need it. Selling a 40-50$ APU with a 5$ interposer hurts.

If "mix and match" binning (ie, using the best quad core Zen die and the best GPU die) allows AMD to make better APUs then maybe that extra $5 could be worth it?

This in contrast to having dedicated APU dies where maybe some have highest bin CPU cores, but lower bin iGPU while others have lower bin CPU with highest bin GPU.

P.S. As far as all the lower bin CPU SoC and GPU dies go .... those can just become desktop parts (ie, AM4 CPU or desktop video card). So the $5 interposer charge would not apply to those.

Also be careful with all your extra chip designs, they each cost a lot of money on their own in masks and design. if you listen to the Intel/Asus podcast you will also hear why multiple chip designs doesn't make economic sense.

I think having an interposer could actually reduce the number of chip designs needed.

Example: Instead of having a specific APU die there would just be the desktop CPU SoC die + one of the dGPU dies.
 
Last edited:

Maxima1

Diamond Member
Jan 15, 2013
3,538
759
146
And you wouldn't be able to fix bad ram and would have to replace the entire thing whenever a single "component" breaks down!

Golden! This would be a win for OEMs AND AMD....let's do this boys! Fk the consumer!!!!

LOL They're already making it so all parts (battery, RAM, storage) aren't removable.
 
Mar 11, 2004
23,250
5,694
146
Actually what I think would make the most sense is if they make a large interposer and then slap a mid-range GPU, even just 2GB of HBM (although the more the better), and then a Zen CPU. There shouldn't be much of a speed penalty, and it would actually help cooling as then they could spread out the heat density while still being able to shrink the final product and they wouldn't be gambling with big dies for niche markets (and it would enable them to be nimble and offer some improvements without having to do a whole new tapeout). Plus they could put SoC stuff (isn't AMD lagging behind Intel on integrating stuff like the Northbridge?) on the interposer so that that could be adjusted as necessary.

Let's face it, AMD isn't going to be competing with Intel to slot into premium thin laptops and tablets that use x86. It just isn't going to happen. But they could definitely carve out niches in budget friendly devices that have real gaming power. If AMD was smart they would be partnering up with Valve to come up with a good Steambox design that would be console esque. I think AMD could even do a double whammy by fitting an ARM SoC in as well and then offering Android compatibility through that (and Microsoft would be smart to try and hedge that into growth for Windows Mobile by pushing for it to be included for free). Imagine a NUC sized device that offered 1080p rendering, Steam integration, and Android (which could handle stuff like Twitch streaming and other things leaving the main Zen APU to just process game related things).

Another benefit to separating the CPU and GPU is that consoles could be improved more incrementally. Say every 2-3 years they offer a new version that has the same CPU but a more powerful GPU to better handle graphics improvements that have happened. This is something I've thought was likely to happen, and recent rumors suggest both Microsoft and Sony are looking towards that even this gen.

It would also enable them to be more nimble as necessary (either if say there's big changes in memory quickly, or if a certain customer needs some special specification, instead of having to come up with a specific die, they can just adjust things on the interposer).

Yes I know I'm basically describing modern laptop and tablet motherboards, but that would be a step up for AMD really. And I think it would enable them to do what they want and adapt more quickly than they have (and especially with how the CPU and GPU designs have not matched up well, they're not stuck holding one for the other).
 

DrMrLordX

Lifer
Apr 27, 2000
22,000
11,560
136
Another benefit to separating the CPU and GPU is that consoles could be improved more incrementally. Say every 2-3 years they offer a new version that has the same CPU but a more powerful GPU to better handle graphics improvements that have happened. This is something I've thought was likely to happen, and recent rumors suggest both Microsoft and Sony are looking towards that even this gen.

GPUs are dependent on CPUs to make them work. We should not expect the current low-clockspeed octocore APUs in the PS4 and Xbone to power GPUs that are more powerful than what they have already. Also, much of what makes stuff like the PS4 and Xbone functional is the integration of the CPU, GPU, and RAM. It would be difficult to just slap on a dGPU that is faster than the iGPU of the PS4 (for example) and expect it to work properly with the low-level development apis already in use for the PS4. It would change a lot of things at the hardware level.
 

ShintaiDK

Lifer
Apr 22, 2012
20,378
145
106
If "mix and match" binning (ie, using the best quad core Zen die and the best GPU die) allows AMD to make better APUs then maybe that extra $5 could be worth it?

This in contrast to having dedicated APU dies where maybe some have highest bin CPU cores, but lower bin iGPU while others have lower bin CPU with highest bin GPU.

P.S. As far as all the lower bin CPU SoC and GPU dies go .... those can just become desktop parts (ie, AM4 CPU or desktop video card). So the $5 interposer charge would not apply to those.



I think having an interposer could actually reduce the number of chip designs needed.

Example: Instead of having a specific APU die there would just be the desktop CPU SoC die + one of the dGPU dies.

If you first design something for an interposer, you are going to need it. Both for the CPU or GPU alone. You will need more chip designs. Plus you need interconnects on all chips you wish to use and such. This will also add cost and size.

And yes, 5$ would completely break the lower end. Specially for a company where 50 million $ can mean the difference between loss or a profit.

Remember people buy CPU performance before GPU performance. If the CPU doesn't deliver, the IGP is meaningless as shown by current sales.
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,911
3,523
136
If you first design something for an interposer, you are going to need it. Both for the CPU or GPU alone. You will need more chip designs. Plus you need interconnects on all chips you wish to use and such. This will also add cost and size.

you mean you would need a physical interface on the GPU and CPU that uses PCI-E x.x based physicals , keying and encoding (like most custom interconnects these days). I can see how that would be a very big problem
 

NTMBK

Lifer
Nov 14, 2011
10,301
5,302
136
you mean you would need a physical interface on the GPU and CPU that uses PCI-E x.x based physicals , keying and encoding (like most custom interconnects these days). I can see how that would be a very big problem

The problem is more fundamental than that. Interposer-die connections use tiny microbumps, which are far more dense than traditional PCB-die bumps. (This is why the Fury X can fit such a massive memory bus- a normal die of that size could not physically fit enough pins.) The entire physical interface has to change.
 

R0H1T

Platinum Member
Jan 12, 2013
2,582
162
106
And that's on a 28nm process. 14LPP allows roughly 2.2x the transistor density. The Playstation 4 APU has eight Jaguar cores and a GPU only slightly below full Pitcairn (18 CUs compared to Pitcairn's 20, with the same 32 ROPs and 256-bit GDDR5 memory bus). Go down to 14LPP and replace the big GDDR5 controller with the smaller HBM2 controller, and you could almost certainly fit in eight Zen cores plus 40 GCN CUs (2560 shaders, the same count as R9 290/390) and 64 ROPs. With execution resources this wide, you wouldn't have to crank up the CPU or GPU clock speeds too high to get competitive performance, thus saving power. (CPU clocks would of course have a turbo feature for lightly-threaded apps.).
Just to add to this raging debate i.e. if it hasn't been posted already
Sources: Sony Is Working On A ‘PlayStation 4.5’
http://kotaku.com/sources-sony-is-working-on-a-ps4-5-1765723053
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |