Discussion RDNA4 + CDNA3 Architectures Thread

Page 165 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

DisEnchantment

Golden Member
Mar 3, 2017
1,747
6,598
136





With the GFX940 patches in full swing since first week of March, it is looking like MI300 is not far in the distant future!
Usually AMD takes around 3Qs to get the support in LLVM and amdgpu. Lately, since RDNA2 the window they push to add support for new devices is much reduced to prevent leaks.
But looking at the flurry of code in LLVM, it is a lot of commits. Maybe because US Govt is starting to prepare the SW environment for El Capitan (Maybe to avoid slow bring up situation like Frontier for example)

See here for the GFX940 specific commits
Or Phoronix

There is a lot more if you know whom to follow in LLVM review chains (before getting merged to github), but I am not going to link AMD employees.

I am starting to think MI300 will launch around the same time like Hopper probably only a couple of months later!
Although I believe Hopper had problems not having a host CPU capable of doing PCIe 5 in the very near future therefore it might have gotten pushed back a bit until SPR and Genoa arrives later in 2022.
If PVC slips again I believe MI300 could launch before it

This is nuts, MI100/200/300 cadence is impressive.



Previous thread on CDNA2 and RDNA3 here

 
Last edited:

blackangus

Member
Aug 5, 2022
141
189
86
There isn't. I'm not going to repeat the whole argument if you didn't get it the first time.

In my picture the Moon is just a white disc.
In an AI enhanced one, the Moon would look like the Moon, as it did when I took the picture.
There's literally no other component, it's a lesser fake vs a better fake. You can argue how to make an even better fake I.E an even closer to the original Moon, but in the end it's all fake.

I find it insane that you actually have people here arguing that they don't live in a world full of fakeness and approximations.
Also as gdansk said that's enough off topic for a position that was entirely silly in the first place.

I dont think anyone is arguing that we don't have alot of fake in the world.
There is a difference in creating something that was not originally there (Fake) vs a failure to reproduce the actual source content. (Not Perfect)
 
Mar 11, 2004
23,250
5,693
146
Well its nice to see that the constant moaning about AMD's GPUs will just be replaced with AI discussions.

I'm personally gung ho and all on board with AMD abandoning high end GPUs. I think they should abandon dGPU add-in cards entirely as its a pointless waste for them. They should have pivoted to eGPUs, and have been moving towards large dedicated gaming APU boxes. This would let them focus laptops and tablets on efficiency, and for gamers it'd have benefits as well (CPUs close by lowers latency, shared memory and even cache so something like 3D V-Cache could benefit GPU more directly than it currently does in the X3D CPU). It would have utilized their embedded/semi-custom teams, but also they could use the entire market chain used for making dGPUs for those so it wouldn't even need a huge shakeup in how they do things since dGPUs are already large enough, would be almost trivial to just make it a fully enclosed box. They also could use it as an opportunity to change the form factor to benefit things (a Gamecube sized box which is basically just a shroud for a modern tower cooler would improve cooling and noise compared to the 3 fan shrouds of modern dGPU). They could do some other further things, like putting an x16 PCIe port (no longer needed for GPU) for ultrafast SSDs where the bandwidth would be more like DRAM memory bandwidth but with improved latency. Uncompress entire game assets and stream them in at high quality, reducing load times and also pop-in. Maybe they could even pre-process ray data for path tracing or something that would make that more feasible.

It would also let Microsoft and Sony (rumors are that despite its success - Playstation was basically singlehandedly keeping Sony overall in business in the 2010s as they were floundering in a lot of their other markets like movies and TV and other consumer electronics and PCs and smartphones - that they struggle with accepting the development cost of Playstation hardware) get out of the costly hardware development business, but keep the benefits of all the software development (run the Xbox software environment on these boxes, and of course Steam could be a viable option as well - Sony really could do themselves favors by partnering with Valve there and it would benefit gamers and developers as well).

That's step 1. Step 2 they replace your beach with a nicer beach, and Step 3 they replace your friends with hotter friends.

You're assuming those are his friends...

And the moon in that pic is perfectly fine for what that pic is. The problem is the person is taking issue with the wrong thing (the poor light/light source handling of smartphone cameras that is a result of the necessary - due to it being a smartphone camera - lens and sensor size situation). Frankly their skills at using even a smartphone camera might be the biggest issue with the quality of that pic. They certainly either are trying to pass off a still capture from a vertical video or did a crop (of a highly zoomed pic) or don't know how to use the resolution of their camera sensor properly. Heck I think you can buy lenses (or filters) that could be used on a smartphone that would have improved that pic for them. Using AI to replace the moon (which, uh, the inherent problem is calling this stuff "AI" when its not at all the AI that people think of as AI - which none of this "AI" can even do and they're just straight up lying about that fact; machine learning image modification is neither new nor really needs these overwrought generative AI models to do) definitely doesn't make it any truer or whatever nonsene argument is being used to justify that.

On the flipside, smartphone cameras inherently are manipulating every image they take (unless you're specifically not doing that which I don't know of basically anyone, including pro photographers, that even use them like that outside of when making something specifically for that reason, and most pro photographers talk about how they use their smartphones specifically to not have to do a lot of work to get an acceptable quality picture), so complaining about using it to help overcome major deficiencies that are inherent because of technical limitations is well, already too late. I would like if it keeps the native image info and you could do a quick switch to see the original vs the modified version. I wouldn't even mind a pro mode where it could do better tweaks of like the color space, making the pro features more useful. Maybe even make an app that teaches people what different settings do to impact image quality (which could help them take better pictures), and then make an automated process (think almost like an eye exam "do you prefer image 1 or 2 and it cycles through a bunch building a profile) so people can create their own algorithm to adjust images to what suits them (and could then be used across apps vs say Snapchat filters and the like).

And just watch, this "AI" replacement or removal will cause all sorts of drama, just like the iMessage stuff does. Actually it already is because people are insane and have gone full creepy, making nudes of people and then sharing that for gross or mean reasons. Which, wonder how many people will use it to add in celebrities they like to make it seem like they were taking pics with them. Bet scammers will use that. "I'm totally rich see me hanging out with these other rich and famous people!"
 
Last edited:

Thunder 57

Platinum Member
Aug 19, 2007
2,954
4,481
136
Well its nice to see that the constant moaning about AMD's GPUs will just be replaced with AI discussions.

I'm personally gung ho and all on board with AMD abandoning high end GPUs. I think they should abandon dGPU add-in cards entirely as its a pointless waste for them. They should have pivoted to eGPUs, and have been moving towards large dedicated gaming APU boxes. This would let them focus laptops and tablets on efficiency, and for gamers it'd have benefits as well (CPUs close by lowers latency, shared memory and even cache so something like 3D V-Cache could benefit GPU more directly than it currently does in the X3D CPU). It would have utilized their embedded/semi-custom teams, but also they could use the entire market chain used for making dGPUs for those so it wouldn't even need a huge shakeup in how they do things since dGPUs are already large enough, would be almost trivial to just make it a fully enclosed box. They also could use it as an opportunity to change the form factor to benefit things (a Gamecube sized box which is basically just a shroud for a modern tower cooler would improve cooling and noise compared to the 3 fan shrouds of modern dGPU). They could do some other further things, like putting an x16 PCIe port (no longer needed for GPU) for ultrafast SSDs where the bandwidth would be more like DRAM memory bandwidth but with improved latency. Uncompress entire game assets and stream them in at high quality, reducing load times and also pop-in. Maybe they could even pre-process ray data for path tracing or something that would make that more feasible.

It would also let Microsoft and Sony (rumors are that despite its success - Playstation was basically singlehandedly keeping Sony overall in business in the 2010s as they were floundering in a lot of their other markets like movies and TV and other consumer electronics and PCs and smartphones - that they struggle with accepting the development cost of Playstation hardware) get out of the costly hardware development business, but keep the benefits of all the software development (run the Xbox software environment on these boxes, and of course Steam could be a viable option as well - Sony really could do themselves favors by partnering with Valve there and it would benefit gamers and developers as well).



You're assuming those are his friends...

And the moon in that pic is perfectly fine for what that pic is. The problem is the person is taking issue with the wrong thing (the poor light/light source handling of smartphone cameras that is a result of the necessary - due to it being a smartphone camera - lens and sensor size situation). Frankly their skills at using even a smartphone camera might be the biggest issue with the quality of that pic. They certainly either are trying to pass off a still capture from a vertical video or did a crop (of a highly zoomed pic) or don't know how to use the resolution of their camera sensor properly. Heck I think you can buy lenses (or filters) that could be used on a smartphone that would have improved that pic for them. Using AI to replace the moon (which, uh, the inherent problem is calling this stuff "AI" when its not at all the AI that people think of as AI - which none of this "AI" can even do and they're just straight up lying about that fact; machine learning image modification is neither new nor is ) definitely doesn't make it any truer or whatever nonsene argument is being used to justify that.

On the flipside, smartphone cameras inherently are manipulating every image they take (unless you're specifically not doing that which I don't know of basically anyone, including pro photographers, that even use them like that outside of when making something specifically for that reason, and most pro photographers talk about how they use their smartphones specifically to not have to do a lot of work to get an acceptable quality picture), so complaining about using it to help overcome major deficiencies that are inherent because of technical limitations is well, already too late. I would like if it keeps the native image info and you could do a quick switch to see the original vs the modified version. I wouldn't even mind a pro mode where it could do better tweaks of like the color space, making the pro features more useful. Maybe even make an app that teaches people what different settings do to impact image quality (which could help them take better pictures), and then make an automated process (think almost like an eye exam "do you prefer image 1 or 2 and it cycles through a bunch building a profile) so people can create their own algorithm to adjust images to what suits them (and could then be used across apps vs say Snapchat filters and the like).

And just watch, this "AI" replacement or removal will cause all sorts of drama, just like the iMessage stuff does. Actually it already is because people are insane and have gone full creepy, making nudes of people and then sharing that for gross or mean reasons. Which, wonder how many people will use it to add in celebrities they like to make it seem like they were taking pics with them. Bet scammers will use that. "I'm totally rich see me hanging out with these other rich and famous people!"

Paragrahps are your friend. Please use them more.
 

marees

Senior member
Apr 28, 2024
371
427
96
Well its nice to see that the constant moaning about AMD's GPUs will just be replaced with AI discussions.

I'm personally gung ho and all on board with AMD abandoning high end GPUs. I think they should abandon dGPU add-in cards entirely as its a pointless waste for them. They should have pivoted to eGPUs, and have been moving towards large dedicated gaming APU boxes. This would let them focus laptops and tablets on efficiency, and for gamers it'd have benefits as well (CPUs close by lowers latency, shared memory and even cache so something like 3D V-Cache could benefit GPU more directly than it currently does in the X3D CPU). It would have utilized their embedded/semi-custom teams, but also they could use the entire market chain used for making dGPUs for those so it wouldn't even need a huge shakeup in how they do things since dGPUs are already large enough, would be almost trivial to just make it a fully enclosed box. They also could use it as an opportunity to change the form factor to benefit things (a Gamecube sized box which is basically just a shroud for a modern tower cooler would improve cooling and noise compared to the 3 fan shrouds of modern dGPU). They could do some other further things, like putting an x16 PCIe port (no longer needed for GPU) for ultrafast SSDs where the bandwidth would be more like DRAM memory bandwidth but with improved latency. Uncompress entire game assets and stream them in at high quality, reducing load times and also pop-in. Maybe they could even pre-process ray data for path tracing or something that would make that more feasible.

It would also let Microsoft and Sony (rumors are that despite its success - Playstation was basically singlehandedly keeping Sony overall in business in the 2010s as they were floundering in a lot of their other markets like movies and TV and other consumer electronics and PCs and smartphones - that they struggle with accepting the development cost of Playstation hardware) get out of the costly hardware development business, but keep the benefits of all the software development (run the Xbox software environment on these boxes, and of course Steam could be a viable option as well - Sony really could do themselves favors by partnering with Valve there and it would benefit gamers and developers as well).



You're assuming those are his friends...

And the moon in that pic is perfectly fine for what that pic is. The problem is the person is taking issue with the wrong thing (the poor light/light source handling of smartphone cameras that is a result of the necessary - due to it being a smartphone camera - lens and sensor size situation). Frankly their skills at using even a smartphone camera might be the biggest issue with the quality of that pic. They certainly either are trying to pass off a still capture from a vertical video or did a crop (of a highly zoomed pic) or don't know how to use the resolution of their camera sensor properly. Heck I think you can buy lenses (or filters) that could be used on a smartphone that would have improved that pic for them. Using AI to replace the moon (which, uh, the inherent problem is calling this stuff "AI" when its not at all the AI that people think of as AI - which none of this "AI" can even do and they're just straight up lying about that fact; machine learning image modification is neither new nor is ) definitely doesn't make it any truer or whatever nonsene argument is being used to justify that.

On the flipside, smartphone cameras inherently are manipulating every image they take (unless you're specifically not doing that which I don't know of basically anyone, including pro photographers, that even use them like that outside of when making something specifically for that reason, and most pro photographers talk about how they use their smartphones specifically to not have to do a lot of work to get an acceptable quality picture), so complaining about using it to help overcome major deficiencies that are inherent because of technical limitations is well, already too late. I would like if it keeps the native image info and you could do a quick switch to see the original vs the modified version. I wouldn't even mind a pro mode where it could do better tweaks of like the color space, making the pro features more useful. Maybe even make an app that teaches people what different settings do to impact image quality (which could help them take better pictures), and then make an automated process (think almost like an eye exam "do you prefer image 1 or 2 and it cycles through a bunch building a profile) so people can create their own algorithm to adjust images to what suits them (and could then be used across apps vs say Snapchat filters and the like).

And just watch, this "AI" replacement or removal will cause all sorts of drama, just like the iMessage stuff does. Actually it already is because people are insane and have gone full creepy, making nudes of people and then sharing that for gross or mean reasons. Which, wonder how many people will use it to add in celebrities they like to make it seem like they were taking pics with them. Bet scammers will use that. "I'm totally rich see me hanging out with these other rich and famous people!"
AMD will be back with full stack for UDNA

& of course with all the AI fakery & trickery. RDNA 4 is stop gap. It wouldn't have tensor cores or matrix processors

AI fakery is the future. Atleast in games. This is because, as mahboiiii says, all game visuals are as it is fake. So AI suits gaming. (He has a philosophic point that even all camera images are fake, but I am strictly limiting to games here)

AMD has the example of sony & nvidia to look up to
PS5 pro — $800 to $1000
4070 super (the pro eq. GPU) = $600
 

branch_suggestion

Senior member
Aug 4, 2023
372
829
96
The split happened as AMD had to abandon GCN to move forward, but for the parts destined for HPC abandoning GCN would've been too risky and expensive to justify (and would break GCN software backcompatibility), so RDNA served as the next generation GFX platform with CDNA being GCN pushed to its very extremes.
MI300's weaknesses vs NV are all the inherent weaknesses of GCN, namely the garbage area eff and high latencies at low level execution along with outdated memory hierarchy.
MI350 is supposedly GFX11 so RDNA3 derived, which is a good foundation for sure regarding the wave model.
UDNA is simply having both client and server back on the same GFX version, just like the GCN days.
This will probably be GFX12.5 or whatever RDNA5 is using which should be the same for MI400, but both will be branded UDNA1 if I were to guess.
It really is a win-win-win situation, hopefully. I wonder how much longer AMD will use Navi for the gaming die names, I suppose for so long the WGP remains the core unit?

Ironically NV did the opposite, their DC uArchs have been more innovative with the gaming ones being more derivative.
 
Reactions: Tlh97 and marees

Mahboi

Golden Member
Apr 4, 2024
1,001
1,803
96
The split happened as AMD had to abandon GCN to move forward, but for the parts destined for HPC abandoning GCN would've been too risky and expensive to justify (and would break GCN software backcompatibility), so RDNA served as the next generation GFX platform with CDNA being GCN pushed to its very extremes.
MI300's weaknesses vs NV are all the inherent weaknesses of GCN, namely the garbage area eff and high latencies at low level execution along with outdated memory hierarchy.
Funny, I thought AMD was all about that area efficiency and NV was all about ignoring area and just getting the product to work no matter the incurred costs.
MI350 is supposedly GFX11 so RDNA3 derived, which is a good foundation for sure regarding the wave model.
Huh. CDNA 3 gets melted down into RDNA 5 rather than a reunification then...
UDNA is simply having both client and server back on the same GFX version, just like the GCN days.
This will probably be GFX12.5 or whatever RDNA5 is using which should be the same for MI400, but both will be branded UDNA1 if I were to guess.
WYM 12.5? Why not 13?
 
Reactions: Tlh97 and marees

branch_suggestion

Senior member
Aug 4, 2023
372
829
96
Funny, I thought AMD was all about that area efficiency and NV was all about ignoring area and just getting the product to work no matter the incurred costs.
They are, that is why they created RDNA in the first place. Instinct does not have a strict area budget, but it had to work in tight deadlines. But now for the most popular accelerated workloads, GCN has run its course.
Huh. CDNA 3 gets melted down into RDNA 5 rather than a reunification then...
CDNA4 is MI350, MI400 is UDNA. Getting the modern ISA from RDNA tailored for accelerated workloads in DC is an upgrade.
WYM 12.5? Why not 13?
Could be 13, but signs are that RDNA5 is also GFX12, and that is the time things reunify.
 

Aapje

Golden Member
Mar 21, 2022
1,506
2,060
106
Attraction isn't even necessary. Men like women who do as they are told. Universal truth. If some guy likes to be controlled and dicked around by a woman's whims, he's not right in the head.

Not sure what world you live in, but I've heard many a man say that they wouldn't be allowed to do some of the things I do. If it was the other way around I'm sure that the feministas would call it systemic abuse.

Anyway, kids are often lonely and I foresee them having AI friend as their best friends and how they learn to interact with women. And then they get older and then they learn that real people suck and they will prefer the AI friends/partners. Just a matter of time, the way I see it.

Anyway, UDNA makes a lot of sense to me, to save them a lot of development work for two separate platforms, and to allow them to support ROCM on regular GPUs way more easily. The downside is that it probably worsens the upward pressure on prices, as gamers will have to pay for the compute hardware, just like we have to on the Nvidia side.
 
Reactions: marees
Jul 27, 2020
19,613
13,477
146
The downside is that it probably worsens the upward pressure on prices, as gamers will have to pay for the compute hardware, just like we have to on the Nvidia side.
Nvidia doesn't keep its prices high for that reason. Consumer cards will always have limited compute capacity so they will need to be priced accordingly. Nvidia only charges high because of AMD's bad reputation and bad software support (like they currently have the fastest NPU at 55 TOPS in premium HP laptops but you can't benchmark it using Geekbench AI). Once UDNA materializes, AMD may even find it easier supporting both consumer and enterprise GPU products due to them being very similar so that may help keep prices reasonable for consumers.
 

Aapje

Golden Member
Mar 21, 2022
1,506
2,060
106
All the more I don't understand why RDNA 4 was guttered then. Keep getting different opinions.

The way I assume it went was that they discovered an unsolvable bug with the chiplet-design of RDNA3 that kept the clocks low, and this same bug was present in the RDNA4-chiplet designs, so they scrapped those (N41, N42 and N43). The design for RDNA 4 already got started way before RDNA3 was released, so this meant that work they did was wasted.

So at that point, they only kept the N44-design, which after all, was always designed as a monolith.

Then I think that they did some work on replacement chips for N41/N42/N43, and those were of course the N45/N46/N47. But then they looked at how much work that was, and concluded that they would have to take resources away from the CNDA-team. Or perhaps they concluded that they wanted to put more resources on UDNA. Or perhaps they thought that these chips would not be competitive. In any case, they then scrapped N45/N46/N47 as well, in favor of one new chip aimed to serve the middle of the market (8800/8700 tiers) and this became the N48.

The actual reasons are hard to know, but I think that the chip names tell a pretty clear story of there being 6 scrapped chip designs.
 

Mahboi

Golden Member
Apr 4, 2024
1,001
1,803
96
The way I assume it went was that they discovered an unsolvable bug with the chiplet-design of RDNA3 that kept the clocks low, and this same bug was present in the RDNA4-chiplet designs, so they scrapped those (N41, N42 and N43). The design for RDNA 4 already got started way before RDNA3 was released, so this meant that work they did was wasted.
We've said about a zillion times that RDNA 3 chiplets have no bugs whatsoever. The performance problem was with the ROPs. Or some last stage rasterization element. That's why RDNA 3 Compute is actually working perfectly while raster/gaming suffers.
Here the XTX beats even the 4090, except in Blender where AMD just is incredibly behind for a reason I don't understand.
Also the packaging technology was meant to be CoWoS-L on RDNA 4, which isn't the one RDNA 3 uses.
 

soresu

Diamond Member
Dec 19, 2014
3,190
2,463
136
except in Blender where AMD just is incredibly behind for a reason I don't understand
2 reasons:

#1. RTX 4xxx HW RT is significantly better than the RDNA3 equivalent.

#2. The OptiX backend for nVidia RT offline rendering is also better than AMD's HIP-RT, and the Blender devs have been working on the former for longer.

Most offline RT renderers only have an OptiX backend for GPU rendering at this point, so it's not like there is much of a talent base out there to discuss problems and solutions with beyond AMD themselves.
 

SolidQ

Senior member
Jul 13, 2023
490
553
96
On the plus side AMD did open source HIP-RT, so hopefully if Blender devs see possible improvements they can make the changes themselves and submit the code to AMD for merging rather than just submitting feature requests.
Just open Blender 4.2

Interesting how is oneAPI is work

I'm planning to take RDNA4 next year and hope it's gonna works fine in Blender(not for rendering)

Seems RDNA4 RT can take more effects
 
Last edited:

dr1337

Senior member
May 25, 2020
410
680
136
Hybrid bonding is a slow process.
Not suitable for volume parts.
AMD is literally running at least 3 if not 4 separate chains of consumer facing 3D stacked products right now. Everything is slow in the beginnings but thats how all production is? Its not like new nodes are super high volume at first ramp either.

Yeah sure some things aren't in Sony's timescale but that isn't a reflection of the entire market let alone what is and isn't possible. AMD probably could scale to their fullest order but it would be too expensive for Sony, and AMD gets better margins out of other 3D products than they would any OEM contract. (and that's better for consumers too, could you imagine if v-cache were console exclusive?)
 
Reactions: Tlh97

adroc_thurston

Diamond Member
Jul 2, 2023
3,492
5,054
96
AMD is literally running at least 3 if not 4 separate chains of consumer facing 3D stacked products right now.
These are low volume.
Consoles are gigaunits relative to all of them.
Everything is slow in the beginnings but thats how all production is?
No, hybrid bonding is inherently a slow process which is why bumped 3d options like SoIC-P exist for mainstream.
AMD probably could scale to their fullest order
There's literally not enough SoIC capacity for that.
 

Josh128

Senior member
Oct 14, 2022
290
403
96
Hybrid bonding is a slow process.
Not suitable for volume parts.
RIght, but I never said anything about using X3D on Playstation APU, that was Igor. I specifically said "its got to be all monolithic". Obviously, adding large cache to monolithic chips greatly increases cost and/or decreases yields, otherwise X3D wouldnt be a thing and they would have included infinity cache from their console APUs.
 
Last edited:
sale-70-410-exam    | Exam-200-125-pdf    | we-sale-70-410-exam    | hot-sale-70-410-exam    | Latest-exam-700-603-Dumps    | Dumps-98-363-exams-date    | Certs-200-125-date    | Dumps-300-075-exams-date    | hot-sale-book-C8010-726-book    | Hot-Sale-200-310-Exam    | Exam-Description-200-310-dumps?    | hot-sale-book-200-125-book    | Latest-Updated-300-209-Exam    | Dumps-210-260-exams-date    | Download-200-125-Exam-PDF    | Exam-Description-300-101-dumps    | Certs-300-101-date    | Hot-Sale-300-075-Exam    | Latest-exam-200-125-Dumps    | Exam-Description-200-125-dumps    | Latest-Updated-300-075-Exam    | hot-sale-book-210-260-book    | Dumps-200-901-exams-date    | Certs-200-901-date    | Latest-exam-1Z0-062-Dumps    | Hot-Sale-1Z0-062-Exam    | Certs-CSSLP-date    | 100%-Pass-70-383-Exams    | Latest-JN0-360-real-exam-questions    | 100%-Pass-4A0-100-Real-Exam-Questions    | Dumps-300-135-exams-date    | Passed-200-105-Tech-Exams    | Latest-Updated-200-310-Exam    | Download-300-070-Exam-PDF    | Hot-Sale-JN0-360-Exam    | 100%-Pass-JN0-360-Exams    | 100%-Pass-JN0-360-Real-Exam-Questions    | Dumps-JN0-360-exams-date    | Exam-Description-1Z0-876-dumps    | Latest-exam-1Z0-876-Dumps    | Dumps-HPE0-Y53-exams-date    | 2017-Latest-HPE0-Y53-Exam    | 100%-Pass-HPE0-Y53-Real-Exam-Questions    | Pass-4A0-100-Exam    | Latest-4A0-100-Questions    | Dumps-98-365-exams-date    | 2017-Latest-98-365-Exam    | 100%-Pass-VCS-254-Exams    | 2017-Latest-VCS-273-Exam    | Dumps-200-355-exams-date    | 2017-Latest-300-320-Exam    | Pass-300-101-Exam    | 100%-Pass-300-115-Exams    |
http://www.portvapes.co.uk/    | http://www.portvapes.co.uk/    |