Ask HN: Is anyone using AMD GPUs for their AI workloads?
4 technoabsurdist 2 6/25/2025, 1:55:07 AM
^ title. I've been renting MI300Xs coz they are cheaper than H100s and my experience has been generally OK (smoother than i expected based on people shitting on AMD so much online). ROCm 6.x seems decent out of the box now, and I'll happily spend 30 more minutes setting up my GPU if it means 20% cheaper. that being said, it's still annoying to run inference for LLMs on AMD's hardware (e.g. You have to install vLLM from source). And there are some other small details which still suck. As a small example, nvidia-smi gives you a nice clear interface while rocm-smi dumps 3 pages of context that's hard to navigate.
would be curious to hear experiences from other folks experimenting with AI workloads.
I tried to run qwen2.5-32B on ROCm5.x and it was running at <15tok/s lol.
Have you tried running any sort of LLM inference on your MI25, or what NN workloads are you running?