We're training LLMs to hallucinate by rewarding them for guessing

2 WASDAai 2 9/7/2025, 10:46:34 PM lightcapai.medium.com ↗

Comments (2)

jqpabc123 · 8h ago
"large language models sometimes guess when uncertain"

I'd argue that most of what they do is guess. Everything they do is some function of weighted probability.

Nothing they do is ever really determinate --- unlike traditional computing. Ask the exact same question from different devices or at different locations and times and you are unlikely to get the same exact response. The wording will vary but sometimes so will the meaning.

WASDAai · 8h ago
A new paper from OpenAI (Sept 2025) makes a compelling argument that the stubborn problem of LLM hallucination isn't a mysterious glitch or something that can be solved with more scale alone. I wrote a deeper analysis of this idea and what it means for the future of AI evaluation .