And while this isn't really appropriate for the article itself, I really like the new documentation website they've launched for ROCm: https://rocm.docs.amd.com/en/latest/
While the bulk of this documentation has always existed, the new website makes it easier to find and easier to follow.
Yeah like why only the 7900 XTX and not the 7900 XT? Makes not a lot of sense. I doubt AMD actual gets the problem? people don't develop on instinct cards unless you are HPC pro. but that it a small niche.
The 7900 XTX and W7900 have the exact same CU configurations. So it cuts down on their support matrix, as any code generated for one card will execute identically on the other.
Intel is plagued with execution problems. Until management addresses the problems, challenging AMD in data center compute is a pipe dream.
For example, Intel this month halted all shipments of its highest volume Sapphire Rapids SKUs due to a bug. No date has been provided when the shipments will resume, which is quite the bad look for Intel given that Sapphire Rapids experienced multiple production delays and missed launch dates. Not surprisingly, Oracle announced this month that it is no longer buying Intel processors for its data center. Instead, Oracle will buy from AMD for its x86 needs. You can expect to see more of these kinds of announcements going forward.
I agree with your comments as they pertain to Intel's horrifying server-space CPU execution problems, but my comment is focused on discrete GPUs and their compute abilities. People are actually engaging with Intel's OneAPI (I think that's the branding? idk) in a meaningful way already, which is already a faster uptake than AMD has ever had.
Yes, Intel's inroads could stagnate due to hardware execution on the GPU side, but given that Intel can afford to slap excessive amounts of VRAM on consumer-priced parts with full compute access, I think Intel actually will surpass AMD in data center compute on the GPU side.
Intel's x86 data center design was completely arrogant and had a low chance of success. It's a shame that Xeon Phi designs aren't around to let Intel juice some fake "market leading performance" numbers while their weird tiled design flounders. A Xeon Phi socketed CPU running OneAPI combined with their current ARC platform would have been a pretty fun platform, even if optimizing for it would be on par in difficulty with optimizing for the modern mainframe platforms.
they really need to step up the software game ... its not going to help jsut hardware if no one can use it efficiently enough.
Also, big part of why CUDA was a success, coz its accessible(apart from having almost 0 competition ..OpenCL doesnt count :P ) on lower end hardware and Windows (which is where almost all studends and many devs start on)
Actually, AMD did in fact improve ROCm recently for their Instinct GPU's... and now they are getting up to 80% of performance of NV's equivalents such as A100.
Granted, that's the Instinct line, and that's their main priority it seems, but I do agree they need to step up their game in the consumer space too, since most people would be in fact using consumer grade GPU's as they are cheaper than the PRO ones.
Unbelievable that people have to wait months for computer support for RDNA3. I assumed that it was compatible on day one like new Nvidia GPUs are compatible with CUDA on day one.
And on top of that, not all RDNA3 cards are supported.
My last Kaveri retired without ever running the least bit of HSA and I've really given up on AMD as a compute platform.
The MI300s seems a little out of range for average developers, educational institutes or hobbyists but it's those people who are needed to create an eco system.
Perhaps the wisest ever decision made by Nvidia was to not limit AI or medium/low precision CUDA like they did with FP64 before on consumer hardware. With that the entry level to Pytorch is as low as a Jetson Nano and you can tax deduct a system that you can also use for gaming during after hours.
By not starting at the APU level for ROCm AMD is wasting such a lot of potential for engagement by students and enthusiasts, who will go to team green instead and will be very difficult to win back.
This, exactly. Frankly, the IGP on Ryzen 7000 should be a priority to receive compute support just for the sake of increasing the number of users able to test against the ROCm stack.
We’ve updated our terms. By continuing to use the site and/or by logging into your account, you agree to the Site’s updated Terms of Use and Privacy Policy.
15 Comments
Back to Article
Ryan Smith - Thursday, June 29, 2023 - link
And while this isn't really appropriate for the article itself, I really like the new documentation website they've launched for ROCm: https://rocm.docs.amd.com/en/latest/While the bulk of this documentation has always existed, the new website makes it easier to find and easier to follow.
nandnandnand - Thursday, June 29, 2023 - link
Take your time, it's not like you're years behind Nvidia.beginner99 - Friday, June 30, 2023 - link
Yeah like why only the 7900 XTX and not the 7900 XT? Makes not a lot of sense. I doubt AMD actual gets the problem? people don't develop on instinct cards unless you are HPC pro. but that it a small niche.Ryan Smith - Friday, June 30, 2023 - link
The 7900 XTX and W7900 have the exact same CU configurations. So it cuts down on their support matrix, as any code generated for one card will execute identically on the other.lmcd - Friday, June 30, 2023 - link
At this point they'll be behind Intel within a year or two in the compute space. It's disturbing to watch the incompetence.State of Affairs - Friday, June 30, 2023 - link
Intel is plagued with execution problems. Until management addresses the problems, challenging AMD in data center compute is a pipe dream.For example, Intel this month halted all shipments of its highest volume Sapphire Rapids SKUs due to a bug. No date has been provided when the shipments will resume, which is quite the bad look for Intel given that Sapphire Rapids experienced multiple production delays and missed launch dates. Not surprisingly, Oracle announced this month that it is no longer buying Intel processors for its data center. Instead, Oracle will buy from AMD for its x86 needs. You can expect to see more of these kinds of announcements going forward.
lmcd - Tuesday, July 4, 2023 - link
I agree with your comments as they pertain to Intel's horrifying server-space CPU execution problems, but my comment is focused on discrete GPUs and their compute abilities. People are actually engaging with Intel's OneAPI (I think that's the branding? idk) in a meaningful way already, which is already a faster uptake than AMD has ever had.Yes, Intel's inroads could stagnate due to hardware execution on the GPU side, but given that Intel can afford to slap excessive amounts of VRAM on consumer-priced parts with full compute access, I think Intel actually will surpass AMD in data center compute on the GPU side.
Intel's x86 data center design was completely arrogant and had a low chance of success. It's a shame that Xeon Phi designs aren't around to let Intel juice some fake "market leading performance" numbers while their weird tiled design flounders. A Xeon Phi socketed CPU running OneAPI combined with their current ARC platform would have been a pretty fun platform, even if optimizing for it would be on par in difficulty with optimizing for the modern mainframe platforms.
andrewaggb - Tuesday, July 4, 2023 - link
well who uses ai anyways 🤦let's mark a card with 24GB of ram and nothing to do with it...
wolfesteinabhi - Friday, June 30, 2023 - link
they really need to step up the software game ... its not going to help jsut hardware if no one can use it efficiently enough.Also, big part of why CUDA was a success, coz its accessible(apart from having almost 0 competition ..OpenCL doesnt count :P ) on lower end hardware and Windows (which is where almost all studends and many devs start on)
deksman2 - Sunday, July 2, 2023 - link
Actually, AMD did in fact improve ROCm recently for their Instinct GPU's... and now they are getting up to 80% of performance of NV's equivalents such as A100.Granted, that's the Instinct line, and that's their main priority it seems, but I do agree they need to step up their game in the consumer space too, since most people would be in fact using consumer grade GPU's as they are cheaper than the PRO ones.
lemurbutton - Monday, July 3, 2023 - link
Unbelievable that people have to wait months for computer support for RDNA3. I assumed that it was compatible on day one like new Nvidia GPUs are compatible with CUDA on day one.And on top of that, not all RDNA3 cards are supported.
nandnandnand - Monday, July 3, 2023 - link
They've earned their tiny sliver of market share relative to Nvidia.abufrejoval - Tuesday, July 4, 2023 - link
Been waiting for this since the launch of Kaveri.My last Kaveri retired without ever running the least bit of HSA and I've really given up on AMD as a compute platform.
The MI300s seems a little out of range for average developers, educational institutes or hobbyists but it's those people who are needed to create an eco system.
Perhaps the wisest ever decision made by Nvidia was to not limit AI or medium/low precision CUDA like they did with FP64 before on consumer hardware. With that the entry level to Pytorch is as low as a Jetson Nano and you can tax deduct a system that you can also use for gaming during after hours.
By not starting at the APU level for ROCm AMD is wasting such a lot of potential for engagement by students and enthusiasts, who will go to team green instead and will be very difficult to win back.
lmcd - Wednesday, July 5, 2023 - link
This, exactly. Frankly, the IGP on Ryzen 7000 should be a priority to receive compute support just for the sake of increasing the number of users able to test against the ROCm stack.lanwatch - Friday, July 14, 2023 - link
I am happily using my Radeon 680M iGPU to tinker with HIP, just compile for gfx1030 and remember toexport HSA_OVERRIDE_GFX_VERSION=10.3.0
before running.