Ask HN: How expensive are LLMs to query, really?
5 teach 3 5/9/2025, 7:58:32 PM
I'm starting to see things pop-up from well-meaning people worried about the environmental cost of large language models. Just yesterday I saw a meme on social media that suggested that "ChatGPT uses 1-3 bottles of water for cooling for every query you put into it."
This seems unlikely to me, but what is the truth?
I understand that _training_ an LLM is very very expensive. (Although so is spinning up a fab for a new CPU.) But it seems to me the incremental costs to query a model should be relatively low.
I'd love to see your back-of-the-envelope calculations for how much water and especially how much electricity it takes to "answer a single query" from, say, ChatGPT, Claude-3.7-Sonnet or Gemini Flash. Bonus points if you compare it to watching five minutes of a YouTube video or doing a Google search.
Links to sources would also be appreciated.
https://www.sustainabilitybynumbers.com/p/carbon-footprint-c...
https://andymasley.substack.com/p/a-cheat-sheet-for-conversa...
(discussion on lobste.rs - https://lobste.rs/s/bxixuu/cheat_sheet_for_why_using_chatgpt...)
(discussion on HN, 320 comments: https://news.ycombinator.com/item?id=42745847)
So.. picking some numbers for calculation. 4 answers per minute @ 120 watts is about .5 watt-hours per answer. ~200 responses would be enough to drain the (normally quite long lasting battery).
How does that compare to the more common nvidia GPUs? I don't know.