A.I. Is Getting More Powerful, but Its Hallucinations Are Getting Worse

22 dewarrn1 10 5/5/2025, 7:33:53 PM nytimes.com ↗

Comments (10)

silisili · 29m ago
I was playing with a toy program trying to hyperoptimize it and asked for suggestions. ChatGPT confidently gave me a few, with reasoning for each.

Great. Implement it, benchmark, slower. In some cases much slower. I tell ChatGPT it's slower, and it confidently tells me of course it's slower, here's why.

The duality of LLMs, I guess.

thechao · 24m ago
Me: What is the tallest tree in Texas?

CGT: The tallest tree in Texas is a 44 foot tall tree in ...

Me: No it's not! The tallest tree is a pine in East Texas!

CGT: You're right! The tallest tree in Texas is probably a Loblolly Pine in East Texas; they grow to a height of 100–150', but some have been recorded to be 180' or more.

Me: That's not right! In 1890 a group of Californians moved to Houston and planted a Sequoia, it's been growing there since then, and is nearly 300 feet tall.

CGT: Yes, correct. In the late 19th century, many Sequoia Sempervirens were planted in and around Houston.

...

I mean, come on; I already spew enough bullshit, I don't need an automated friend to help out!

datadrivenangel · 3h ago
This may be an issue with default settings:

"Modern LLMs now use a default temperature of 1.0, and I theorize that higher value is accentuating LLM hallucination issues where the text outputs are internally consistent but factually wrong." [0]

0 - https://minimaxir.com/2025/05/llm-use/

dewarrn1 · 4h ago
So, in reference to the "reasoning" models that the article references, is it possible that the increased error rate of those models vs. non-reasoning models is simply a function of the reasoning process introducing more tokens into context, and that because each such token may itself introduce wrong information, the risk of error is compounded? Or rather, generating more tokens with a fixed error rate must, on average, necessarily produce more errors?
ActorNightly · 2h ago
Its a symptom of asking the models to provide answers that are not exactly in the training set, so the internal interpolation that the models do probably runs into edge cases where statistically it goes down the wrong path.
scudsworth · 2h ago
bdangubic · 2h ago
"self-driving cars are getting more and more powerful but the number of deaths they are causing is rising exponentially" :)
dimal · 2h ago
I wish we called hallucinations what they really are: bullshit. LLMs don’t perceive, so they can’t hallucinate. When a person bullshits, they’re not hallucinating or lying, they’re simply unconcerned with truth. They’re more interested in telling a good, coherent narrative, even if it’s not true.

I think this need to bullshit is probably inherent in LLMs. It’s essentially what they are built to do: take a text input and transform it into a coherent text output. Truth is irrelevant. The surprising thing is that they can ever get the right answer at all, not that they bullshit so much.

kelseyfrog · 26m ago
In the same sense that astrology readings, tarot readings, runes, augury, reading tea leaves are bullshit - they have oracular epistemology. Meaning comes from the querant suspending disbelief, forgetting for a moment that the I Ching is merely sticks.

It's why AI output is meaningless for everyone except the querant. No one cares about your horoscope. AI shares every salient feature with divination, except the aesthetics. The lack of candles, robes, and incense - the pageantry of divination means a LOT of people are unable to see it for what it is.

We live in a culture so deprived of meaning we accidentally invented digital tea readings and people are asking it if they should break up with their girlfriend.

elpocko · 46m ago
Or maybe we could stop anthropomorphizing tech and call the "hallucinations" what they really are: artifacts introduced by lossy compression.

No one is calling the crap that shows up in JPEGs "hallucinations" or "bullshit"; it's commonly accepted side effects of the compression algorithm that makes up shit that isn't there in the original image. Now we're doing the same lossy compression with language and suddenly it's "hallucinations" and "bullshit" because it's so uncanny.