Ask HN: Does acceptance of Wikipedia as reliable source foreshadow same for AI?
4 bookofjoe 11 4/13/2025, 3:48:01 PM
When Wikipedia started it was denounced as an unacceptable reference source compared to standards like Encyclopedia Britannica. Now its information is regularly cited here, generally without dispute. The output/product of AI/LLM is denounced here as hallucination-prone and unacceptable for reference. Will that too pass just as Wikipedia's early days engendered fierce debate as to its suitability?
Neither is a primary source though, but the advantage of linking to an encyclopedia is that it provides a plain-English summarization of the primary sources, as well as pertinent reference to those sources.
Large language models have historically been incapable of providing sources, but newer models are gaining the ability to do so, which is making them as useful as encyclopedias. Until everyone starts using sourced output from LLMs, we get to see who is blindly trusting hallucinations. (https://www.infodocket.com/2024/12/06/report-media-personali...)
Today it is better in terms of comprehensiveness but it is still poorly written compared to the work of professional writers and editors. In part because good writing is not valued and well-written passages get arbitrary edits from amateurs.
https://www.nature.com/articles/438900a (2005)
FWIW, Perplexity Pro now provides numerous primary source references/citations for its answers.
Is is "good enough" for the general public, but that is not the same thing.
You generally can't even cite encyclopedias in high school research reports.
"Source of reliable information" is one of them.
"Source of how a topic has changed over time" is another.
"Source of what disputes are more common regarding specific parts of each article" is another.
And so on...
Even if most people don't actually do that and trust every small bit of information (like the original amount of hair of some classical composer had), some other people will in fact track and try to understand the path of that information and whether is truthful or not, relevant or not.
Maybe one day LLMs will allow that kind of thing as well. I don't know. Currently they don't offer that choice.
Does that answer your question?
Wikipedia has a very dedicated fact-checking team who try to enforce accuracy, and (at least for non-political articles) most agree they do very well. Perhaps someone will develop a reliable automated fact-checker, then it can be applied to LLM output to “bless” it or point out the mistakes.