AMD Raven Ridge 'Zen APU' Thread

Page 6 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
I'm not sure AMD has any immediate plans to produce clusters of Zen cores smaller than 4, but otherwise . . . gosh I hope so! It'd better be faster than Carrizo/Bristol Ridge.

Im not suggesting a different die, im saying a cut down die with only Dual CPU Cores and cut down iGPU to 8x CUs could be more than 30% faster in Graphics than BristolRidge if paired with DDR-4 3200MHz memory.
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
You can't cut down a CCX, without re-designing it. You can disable cores and cache within the CCX, but that's pretty pointless. Raven is most likely going to have a single CCX, i.e 4C/8T.
 

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
Ehm I mean to disable 2x CPU cores and left with only 2x Cores + HT(4x Threads). Same with the iGPU, disable 3x CUs and left with 8x CUs (if rummors of 11CUs is correct).
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
Ehm I mean to disable 2x CPU cores and left with only 2x Cores + HT(4x Threads). Same with the iGPU, disable 3x CUs and left with 8x CUs (if rummors of 11CUs is correct).

Why in the hell would they do such thing?
 

DrMrLordX

Lifer
Apr 27, 2000
21,797
11,143
136
Why in the hell would they do such thing?

Maybe if one or two of the cores and/or some of the CUs are unstable/faulty. I'm not sure if that'll be a big issue though, or if having faults in one core would render the entire CCX unusable.
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
In that case there would be harvested models with 2C/4T. Including a CCX with four cores in it to a design and then only selling parts with half of the cores enabled would be plain madness.
 

DrMrLordX

Lifer
Apr 27, 2000
21,797
11,143
136
Yeah I'm thinking he was referring to a hypothetical harvested model. It might take a bit for them to get enough "bad" CPUs to sell that way, delaying its release (sort of like the 860k).
 

cbn

Lifer
Mar 27, 2009
12,968
221
106
HBM is out of the question at least for Raven Ridge and it's possible successor in the consumer segment. AMD needs either to implement eDRAM or GDDR5X, or alternatively forget about the APUs completely. HBM is way too complex (expensive) and DDR4 frequencies won't reach high enough to provide sufficient bandwith on a 128-bit bus.

Sad but true.

I think AMD will probably include an HBM2 controller for the successor of Raven Ridge (ie, the APU with the Vega iGPU). The question is will they enable the HBM2 controller or will they disable it like they did with GDDR5 controller in Kaveri/Godavari?

I hope they enable it. Besides the obvious performance boost from HBM2, the PCB size reduction (from using HBM2 as system RAM) means the laptop can either be smaller or have a bigger fan and/or battery than it normally would.

P.S. At one time (back in 2010) AMD had a program called the Fusion fund. This, in part, was support from AMD to develop new factors that would enhance the benefit of the APUs.

The AMD Fusion Fund enables investment in companies that are developing innovative solutions and applications that leverage the features of the AMD Fusion family of APUs. The program is intended to accelerate time to market for solutions such as:

– Applications that are newly enabled or greatly enhanced by the synergy between CPU and GPU
– Tools that enable developers to easily develop, scale, optimize or test applications
– Technology or intellectual properties that can be a foundational building block of applications
– Form factors or devices that expand the scope for deployment of applications

Too bad they discontinued that program because today I think such investment is even more important than it was in the past. (re: it very well might be future APUs become extremely integrated to the point of even having the NAND much closer to the processor than we expected)
 
Last edited:

cbn

Lifer
Mar 27, 2009
12,968
221
106
Some interesting additional info on HBM2:

http://www.linleygroup.com/newsletters/newsletter_detail.php?num=5514

HBM2 Doubles Bandwidth, Capacity
May 3, 2016

Author: David Kanter

High Bandwidth Memory 2 (HBM2) has several incremental but critical improvements over the previous-generation technology: it increases the memory capacity and bandwidth, adds ECC support, and improves memory-controller efficiency. These enhancements are essential for high-performance computing (HPC) and workstation tasks and will lead to wider industry adoption. Samsung’s HBM2 chips are already in production, and the memory will appear in new graphics cards later this year.

Over the last decade, GDDR has been the high-performance graphics memory of choice, balancing bandwidth and cost for low-capacity systems. GDDR5 is the latest incarnation, but it can’t scale much past 8Gbps; in practice, no GPU has achieved more than 320GB/s using GDDR5. The industry is currently developing GDDR5X, which will roughly double the bandwidth, but this memory is unlikely to reduce power consumption at higher data rates.

HBM uses 2.5D or 3D packaging to increase bandwidth and power efficiency by implementing a wide slow interface, but the interposer raises the cost. Since modern computing systems are almost always limited by power consumption, HBM gives architects an effective tool to improve system performance. For example, the first GPU to use HBM was AMD’s Fiji, which delivered an incredible 512GB/s while using 3x less power than an equivalent GDDR interface. But HBM primarily serves in consumer graphics because of limitations in the first-generation technology—limitations that are alleviated by HBM2.
 

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
Why in the hell would they do such thing?

Every dual core (Single Module) APU today is coming from a Dual Module Quad Core die. There are Dual cores with 256 iGPU Shaders APUs, exactly half the Kaveri die. A6-7400K is one of them, this is a harvested Kaveri exactly in Half, but keeping the 128bit memory.
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
Every dual core (Single Module) APU today is coming from a Dual Module Quad Core die. There are Dual cores with 256 iGPU Shaders APUs, exactly half the Kaveri die. A6-7400K is one of them, this is a harvested Kaveri exactly in Half, but keeping the 128bit memory.

Not exactly true, since AMD just created Stoney Ridge to avoid this.
 

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
Not exactly true, since AMD just created Stoney Ridge to avoid this.

Ok true forgot about StoneRidge, they could do the same with ZEN APUs, half the big die.

Dual ZEN Cores + HT and 384 shaders (6x CUs) with Dual channel DDR-4 memory. Something like that ??
 

Insert_Nickname

Diamond Member
May 6, 2012
4,971
1,692
136
Dual ZEN Cores + HT and 384 shaders (6x CUs) with Dual channel DDR-4 memory. Something like that ??

Doesn't seem half bad for an entry-level solution, if the frequency is solid. It would essentially be a modern version of Trinity/Richland, but with both a better CPU and GPU architecture.
 

cbn

Lifer
Mar 27, 2009
12,968
221
106
Regarding the 35W situation OEMs have been struggling with, I am assuming with the reported ECC support for Raven Ridge AMD is going to bring back FirePro.

So that is one factor that should help establish the needed 35W APU only chassis (which can also be shared with non FirePro 35W Raven Ridge consumer APU laptops).

With that mentioned, I wonder how many Mobile workstations are actually used on battery power. I would think the predominant usage is AC power (ie, plugged into the wall).

Maybe to make the chassis more compact and/or to provide extra airflow to the VRMs and NAND (which I assume will only get closer and closer to the processor over time) there could be a "transformer" type "35W Notebook (without battery)" that can turn into a 15W Workstation with the addition of a battery in place of a supplementary cooler/fan system.

Notebook without battery (and supplementary cooler in place) = 35W cTDP

Notebook with battery (and supplementary cooler removed) = 15W cTDP (possibly with lower memory clocks and tighter timings on the DDR4 if this would boost CPU performance at this lower TDP)

So, in a nutshell, having the replaceable battery/cooler allows two types of laptops:

1. More compact than usual 35W for folks that mostly use or purely use AC power (but still have the option of 15W battery power for mobile usage if it becomes necessary)

2. More quiet than usual 35W for folks that mostly use or purely use AC power (but still have the option of 15W battery power for mobile usage if it becomes necessary).

Of the two options above, I am thinking number 2 would be more popular.

P.S. The only down side I can think of would be lack of built-in UPS for those using the laptop in 35W mode (ie, without battery). Maybe if necessary there could be a compromise solution where there is a type supplementary cooler that has a very small battery in it (Maybe even just two rechargeable AA batteries).
 

cbn

Lifer
Mar 27, 2009
12,968
221
106
A third transforming option would be 35W SFF slim tower (with the notebook lid closed and the machine positioned vertically on some type of stand with feet).

So these three options in total:

1. 35W SFF slim tower with battery removed and supplemental cooler in place (mounted in some type vertical docking stand).

2. 35W notebook with battery removed and supplemental cooler in place.

3. 15W notebook with battery installed and supplemental cooler removed.
 

NTMBK

Lifer
Nov 14, 2011
10,269
5,134
136
Adding removable this, that and the other will increase size, increase cost, and reduce robustness and build quality.
 
Reactions: TurtleCrusher

Shehriazad

Senior member
Nov 3, 2014
555
2
46
All I want is an APU that has at least GTX 750 Ti performance with at least last gen i3 CPU perf levels at less than 100 watts of total system power draw.

Can we get there soon please? xD
 

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
All I want is an APU that has at least GTX 750 Ti performance with at least last gen i3 CPU perf levels at less than 100 watts of total system power draw.

Can we get there soon please? xD

Raven Ridge with 11x CUs at 1000MHz and 3200MHz memory could come close to GTX 750 (non Ti) at 95W TDP. It will have lower memory bandwidth but second generation color compression.
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
1000MHz on a 11 GCU GPU would be a major waste of power. The relative memory bandwidth even at the maximum theoretical 3200MHz memory frequency would be ~13% lower than on a FX-8800P Carrizo (2133MHz). And even the FX-8800P is completely clogged up in terms of the available GPU bandwidth.

Anything higher than < 800MHz (at DDR-3200) would be waste of power. Since the maximum theoretical frequency supported by the controller is most likely not even supported / enabled on Raven, the highest "power efficient" frequency for the GPU would be even lower.
 

AtenRa

Lifer
Feb 2, 2009
14,003
3,361
136
1000MHz on a 11 GCU GPU would be a major waste of power. The relative memory bandwidth even at the maximum theoretical 3200MHz memory frequency would be ~13% lower than on a FX-8800P Carrizo (2133MHz). And even the FX-8800P is completely clogged up in terms of the available GPU bandwidth.

Anything higher than < 800MHz (at DDR-3200) would be waste of power. Since the maximum theoretical frequency supported by the controller is most likely not even supported / enabled on Raven, the highest "power efficient" frequency for the GPU would be even lower.

Games and applications are becoming more Compute heavy (DX-12/Vulkan, OpenCl etc). If we have 1000MHz on BristolRidge (8x CUs) at 28nm, there is no reason not have 1000MHz on Raven Ridge (11x CUs) at 14nm. This is for the Desktop high-End 95W TDP APUs, lower SKUs could have lower clocks. Example, the 8x CU 35W TDP Raven Ridge could only have 800MHz iGPU.
 

Sweepr

Diamond Member
May 12, 2006
5,148
1,142
131
Games and applications are becoming more Compute heavy (DX-12/Vulkan, OpenCl etc). If we have 1000MHz on BristolRidge (8x CUs) at 28nm, there is no reason not have 1000MHz on Raven Ridge (11x CUs) at 14nm. This is for the Desktop high-End 95W TDP APUs, lower SKUs could have lower clocks. Example, the 8x CU 35W TDP Raven Ridge could only have 800MHz iGPU.

Both Fudzilla and SemiAccurate say it tops out at 65W, so not sure where are you seeing 95W models.
 

Glo.

Diamond Member
Apr 25, 2015
5,761
4,666
136
Both Fudzilla and SemiAccurate say it tops out at 65W, so not sure where are you seeing 95W models.

APUs are rumoured to have 65W TDP.

HEDT chips 95W. But they don't have integrated GPU.
 

jpiniero

Lifer
Oct 1, 2010
14,831
5,444
136
Yeah I think going forward I wouldn't expect any APU to be more than 65W. That's what Intel is doing, and they're obviously doing it because the OEMs want it that way.
 

The Stilt

Golden Member
Dec 5, 2015
1,709
3,057
106
Games and applications are becoming more Compute heavy (DX-12/Vulkan, OpenCl etc). If we have 1000MHz on BristolRidge (8x CUs) at 28nm, there is no reason not have 1000MHz on Raven Ridge (11x CUs) at 14nm. This is for the Desktop high-End 95W TDP APUs, lower SKUs could have lower clocks. Example, the 8x CU 35W TDP Raven Ridge could only have 800MHz iGPU.

The thing is that the top Bristol Ridge SKU is probably the most retarded product AMD has ever created. In the very best case (i.e 2400MHz memory) it has < 33.85MB/s per GFlop of shared bandwidth available. Ideally a GCN dGPU should have > 50MB/s to not to be held back by the memory bandwidth. For a iGPU this figure is even higher (due the bandwidth being shared), but the performance scales nearly linearily up to 90MB/s per GFlop.

For Carrizo / Bristol Ridge anything higher than 625MHz for a 8 GCU is mostly wasted power, given that 2400MHz dual channel memory is used. Even if the color compression on Raven would be 20% more efficient than the one used in Carrizo / Bristol Ridge, anything higher than 606MHz - 727MHz (at 2666MHz - 3200MHz memory) for a 11 GCU configuration would be waste of power.

Since AMD seems to be completely incapable of learning their past mistakes, I wouldn't be surprised if they still end up clocking the GPU to unnecessary high levels. Also regarding the computing argument, OpenCL surely uses the shaders, however I'm not that sure that the ACEs even operate at core clock directly.
 
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |