> “If you look at the models before they are fine-tuned on human preferences, they’re surprisingly well calibrated. So if you ask the model for its confidence to an answer—that confidence correlates really well with whether or not the model is telling the truth—we then train them on human preferences and undo this.
Now that is really interesting! I didn't realize RLHF did that.
davesmylie · 3h ago
Obviously it's well over a year since this article was posted and if anything I've anecdotally noticed hallucinations getting more, not less, common.
Possibly/probably with another years experience with LLMs I'm just more attuned to noticing when they have lost the plot and are making shit up
DoctorOetker · 24m ago
In the article it is argued that the brainfarts could be beneficial for exploration of new ideas.
I don't agree. The "temperature" parameter should be used for this. Confabulation / bluff / hallucination / unfounded guesses are undesirable at low temperatures.
ipv6ipv4 · 3h ago
There is ample evidence that hallucinations are incurable in the best extant model of intelligence - people.
add-sub-mul-div · 2h ago
Someday we'll figure out how to program computers to behave deterministically so that they can complement our human abilities rather than badly impersonate them.
alkyon · 2h ago
The more accurate word would be confabulation
Wowfunhappy · 2h ago
You lost this battle, sorry. It's not going to happen.
Both terms are "inaccurate" because we're talking about a computer program, not a person. However, at this point "hallucination" has been firmly cemented in public discourse. I don't work in tech, but all of my colleagues know what an AI hallucination is, as does my grandmother. It's only a matter of time until the word's alternate meaning gets added to the dictionary.
alkyon · 1h ago
Maybe I lost this battle, but also in science the terminology evolves. If you replace AI hallucination with AI confabulation even your grandmaother would get it right. I also don't agree that both terms are equally inaccurate.
peterashford · 1h ago
Correct. This is the way language works. It's annoying when you know what words mean but this is the way it is.
alkyon · 1h ago
Obviously, hallucination is by definition a perception, so it incorrectly anthropomorphizes AI models. On the other hand, the term confabulation involves filling in gaps with fabrication, exactly what LLMs do (aka bullshitting).
more_corn · 1h ago
What an absurd prediction.
techpineapple · 4h ago
I wonder if it would be better to have 1 “perfect” LLM trying to solve problems or 5 intentionally biased LLM’s.
d00mB0t · 4h ago
I'm so tired of these rich dweebs pontificating to everyone.
Now that is really interesting! I didn't realize RLHF did that.
Possibly/probably with another years experience with LLMs I'm just more attuned to noticing when they have lost the plot and are making shit up
I don't agree. The "temperature" parameter should be used for this. Confabulation / bluff / hallucination / unfounded guesses are undesirable at low temperatures.
Both terms are "inaccurate" because we're talking about a computer program, not a person. However, at this point "hallucination" has been firmly cemented in public discourse. I don't work in tech, but all of my colleagues know what an AI hallucination is, as does my grandmother. It's only a matter of time until the word's alternate meaning gets added to the dictionary.