- Mar 3, 2017
- 1,747
- 6,598
- 136
matrix math isn't all that useful.I am a straight up bubble huffer. I say that AI functions will be the main reason for sales of computing devices by 2040. I still find it bizarre there’s so many people on various text site forums don’t see that AI is THE future of computing in society.
xir we're launching TLAMs at Taiwan in a few years.I believe the next 30 years will yield a bigger technological and social transformation than than the period from the dawn of civilization to 2024.
MoP means SKU spam.I can’t think of any for STX halo.
this shows a complete lack of understanding about the hardware your algorithms are running on and the evolution of hardware and what the future looks like.I am a straight up bubble huffer. I say that AI functions will be the main reason for sales of computing devices by 2040. I still find it bizarre there’s so many people on various text site forums don’t see that AI is THE future of computing in society.
To be sure, new algorithms are needed. The current ones are just stopgaps, as impressive as they may be at these small simple tasks. But, just as current computers are many orders of magnitude more complex and capable than a Commodore 64, neural networks are going to follow the same trajectory, at about triple the speed.
Research is already underway to boost reasoning capacity by 100 to 1000 fold. Those algorithms will be here in a few short years, ones that will make large language models feel like stone age technology. Critically, those won’t be the last ones.
I believe the next 30 years will yield a bigger technological and social transformation than than the period from the dawn of civilization to 2024.
Yeah, I’m huffing big time.
matrix math isn't all that useful.
Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.xir we're launching TLAMs at Taiwan in a few years.
You sure?
Well obviously, I think the lack of understanding is fully in your camp.this shows a complete lack of understanding about the hardware your algorithms are running on and the evolution of hardware and what the future looks like.
Your Computations are already running on hyper targeted and optimised hardware that has the biggest BOM's out side of mainframes we have ever seen. Also who exactly is making any money from AI outside Nvida? Im watching this daily as i want to move before the A** falls out of the market.
First time?and yet is the main focus of every major chipmaker on the planet, as well as nearly all of the minor ones. From CEOs, CTOs, PhD fellows all the way down to interns, it is the most important aspect of every architecture and chip at every stage from concept through design, production and validation. Hmmmmmmmm.
That's not how it works.Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.
Mod DAPUNISHERAs to why the equivalent delay is 1 and not 0.5, this is a major problem I'm having at the moment.The current version of the microcode seems that a single thread cannot see two decoders no matter what, that is, after the op$ is released or turned off, the front end directly becomes 4-wide and can only take 1 per cycle (regardless of whether there is a branch jump or not). This is obviously inconsistent with AMD's propaganda that a single thread can use two decoders, and more investigation is needed.
Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.>RDNA double-issue flashbacks.
At 2000 it's a goner. For it to be successful it needs to be at 1200-1500€. There is already a product of Minisforum which combines a 7945hx and a 7600 XT which they are selling now for 1200€ (1000 right now). The price can't be more then cpu+gpu of same perf, or it will flop.If AMD wants to keep it around $2000
>+30% ipc is back on the menu boys 🤡>RDNA double-issue flashbacks.
^we are coping.Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.
As it turns out, the software is lagging the hardware and holding it back, as is tradition for AMD.
Whether there is a new AGESA in the next week or so that magically enables the core to use dual decoders for compatible 1t workloads, well that would make any review delay justified.
Yes I am coping.
So one model that cant do anything useful is more efficient then another model that cant do anything useful....and yet is the main focus of every major chipmaker on the planet, as well as nearly all of the minor ones. From CEOs, CTOs, PhD fellows all the way down to interns, it is the most important aspect of every architecture and chip at every stage from concept through design, production and validation. Hmmmmmmmm.
Surely, you’ve noticed that multiple countries have begun spending hundreds of billions of dollars to ensure that they have angstrom scale fabs inside their borders. Each of those countries will massively increase spending and several other nations will join them.
Well obviously, I think the lack of understanding is fully in your camp.
Llama 3 405B has for the last week been used and tested by thousands of people who have been heavily using LLMs for two to three years now. So far it consistently demonstrates effectively equal capabilities as ChatGPT 4o. And it will run on 2 Macbook Pros connected with a single Thunderbolt cable.
More importantly, LLMs are not the AI revolution. They are the precursor.
int PRF is barely a bump over Z4.Didn't think they could get any more out of that.
I am afraid it's about bad marketing message and miscommunication. The materials were mentioning that decoders are statically partitioned in SMT mode. Now traditionally when you wanted to turn off SMT, you went to BIOS and disabled it. Now the question is, is the SMT mode static when enabled [If SMT is on in the BIOS is the core always in SMT mode] or is it dynamic like the interviews are leading us to believe.Well yeah, AMD is doing that funny PPA trick of stripping out hardware, adding double pump logic cause it is cheap and then having to rely on compilers to actually utilise the hardware.
As it turns out, the software is lagging the hardware and holding it back, as is tradition for AMD.
Whether there is a new AGESA in the next week or so that magically enables the core to use dual decoders for compatible 1t workloads, well that would make any review delay justified.
Yes I am coping.
🤔 It seems that the AMD employee who submitted the patch knew it was 4-wide all along.I guess it's the reason why GCC still treats Zen 5 as 4-wide decode uarch.
Someone needs to run GNR with SMT off once it's released to see what's really going on.🤔 It seems that the AMD employee who submitted the patch knew it was 4-wide all along.
We were bamboozled by Mike Clark yet again.
Zen 5.5* 🤣
It might be other things too. These were early patches, they might have wanted just to get znver5 option added that would not dramatically break the situation for people using --march=native rather than to give accurate representation of the core. They might have wanted not to share everything or they might have forgot to update. Don't forget that CPUs are designed to handle less than ideal code [OoO, branch prediction etc.] so this won't have terrible effect. It would be much worse if somebody forgot to turn on all available instructions sets as that would hamper the generated code more.🤔 It seems that the AMD employee who submitted the patch knew it was 4-wide all along.
We were bamboozled by Mike Clark yet again.
The real AI has never been tried, just wait 2-3 weeks (insert 💲💲 here). Well, seriously, what kind of use the current transformers have apart from replacing the politicians (as they can lie and get even more delusional than the most flamboyant political figures around the world). And also, why do we need it integrated into general-purpose CPUs and GPUs at all, it could be just a separate addon board or card and we won't need to sacrifice 16 mb of cache and employ a weird dual ccd setup for this deadweight siliconTo be sure, new algorithms are needed.