Show HN: Selfhostllm.org – Plan GPU capacity for self-hosting LLMs

4 erans 1 8/8/2025, 5:49:49 PM selfhostllm.org ↗
A simple calculator that estimates how many concurrent requests your GPU can handle for a given LLM, with shareable results.

Comments (1)

erans · 3h ago
I also added a Mac version: https://selfhostllm.org/mac/ so you can know which models you can run on your Mac and get an estimated tokens/sec.