Ask HN: Is anyone using AMD GPUs for their AI workloads?

4 technoabsurdist 2 6/25/2025, 1:55:07 AM
^ title. I've been renting MI300Xs coz they are cheaper than H100s and my experience has been generally OK (smoother than i expected based on people shitting on AMD so much online). ROCm 6.x seems decent out of the box now, and I'll happily spend 30 more minutes setting up my GPU if it means 20% cheaper. that being said, it's still annoying to run inference for LLMs on AMD's hardware (e.g. You have to install vLLM from source). And there are some other small details which still suck. As a small example, nvidia-smi gives you a nice clear interface while rocm-smi dumps 3 pages of context that's hard to navigate.

would be curious to hear experiences from other folks experimenting with AI workloads.

Comments (2)

dlcarrier · 9h ago
I'm using an MI25, flashed as a PRO WX 9100, which requires an older version of ROCm to work. That's expectedm, because my GPU is depricated in future versions of ROCm, but what irks me is that everything neural network related barely works. You need the exact version of every interpreter and library, which ends up working on some distributions but not others. I've noticed that when people program in compiled languages, they seem to make a concerted effort to do some kind of bounds testing, but anything in Python or Node.js seems to be released as soon as it kind-of-sort-of works, some of the time.
technoabsurdist · 7h ago
oh yeah, in my experience anything below ROCm6.x really sucks.

I tried to run qwen2.5-32B on ROCm5.x and it was running at <15tok/s lol.

Have you tried running any sort of LLM inference on your MI25, or what NN workloads are you running?