A GPU Calculator That Helps Calculate What GPU to Use

37 chlobunnee 11 7/24/2025, 10:14:20 PM calculator.inference.ai ↗

Comments (11)

zargon · 2h ago
The best VRAM calculator I have found is https://apxml.com/tools/vram-calculator. It is much more thorough than this one. For example, it understands different models' attention schemes for correct KV cache size calculation, and supports quantization of both the model and the KV cache. Also, fine-tuning. It has its own limitations, such as only supporting specific models. In practice though, the generic calculators are not very useful because model architectures vary (mainly the KV cache) and end up being way off. (Not sure whether or not it would be better to discuss it separately, but I submitted it at https://news.ycombinator.com/item?id=44677409)
zeroq · 10m ago
This one is indeed much better and it instantly answers my immediate feedback I wanted to leave for the one originally posted, which is - instead of calculating an artificial scenario I would like to state what can I run on the hardware I actually have at hand. Thanks!
funfunfunction · 3h ago
This is a cheap marketing ploy for a GPU reseller with billboards on highway 101 into SF.
LorenDB · 3h ago
Where's AMD support? I have a 9070 XT and would love to see it listed on here.
amanzi · 2h ago
I would have liked to see the RTX 5060 Ti with 16GB mentioned. I can't tell if it's omitted because it won't work, or if it's excluded for some other reason?
amatecha · 1h ago
Yeah, weird miss, but maybe just because it came out more recently. It can be used for ~anything a 5070 could be used for, no? Maybe slower, but still.
chlobunnee · 3h ago
I built a calculator to help researchers and engineers pick the right GPUs for training and inference workloads!

It helps compare GPU options by taking in simple parameters (# of transformer layers, token size, etc) and letting users know which GPUs are compatible + their efficiency for training vs inferencing.

The idea came from talking with ML researchers frustrated by slow cluster queues or wasting money on overkill GPUs.

I'd love feedback on what you feel is missing/confusing!

Some things I'm thinking about incorporating next are >Allowing users to directly compare 2 GPUs and their specs >Allowing users to see whether a fraction of the GPU can complete their workload

I would really appreciate your thoughts/feedback! Thanks!

timothyduong · 2h ago
Where's 3090? Or should that fall in the 4090 (24GB VRAM) category?
snvzz · 2h ago
Rather than GPU calculator, this is an NVIDIA calculator.
nodesocket · 2h ago
In case you’ve been living in a cave, Nvidia is the defacto standard for LLM compute.
quotemstr · 3h ago
No sharding? At all?