AI hallucinations: Why LLMs make things up (and how to fix it)

What “hallucination” means

  • Many argue hallucinations are not a special mode but simply “outputs not fit for purpose” from a probabilistic text generator; in a sense, all outputs are hallucinations, some just align with reality.
  • Others insist hallucinations should be reserved for fabricated or non‑grounded claims, especially invented citations, APIs, or features.
  • Some frame LLMs as modeling language, not the world; grammar and style are accurate even when facts are wrong.

Can hallucinations be fixed or only mitigated?

  • Strong view: hallucinations are inevitable in this architecture; they’re a property of probabilistic modeling and imperfect data, so they can only be managed, not eliminated.
  • Counter‑view: calling them “inevitable” is premature; better architectures, confidence estimation, and multi‑stage checking may drastically reduce them.
  • Debate over whether hallucinations are “bugs”: some say yes (result not matching user intent), others say no (it’s expected behavior, like Bloom filter false positives or network latency).

Techniques to reduce hallucinations

  • Retrieval‑augmented generation (RAG) and domain‑restricted context: treat the LLM as a natural‑language interface over vetted documents; separate fact retrieval from wording.
  • External verification and secondary models: fact‑check outputs against context, or use trained “factuality” checkers.
  • Prompting and instruction: explicitly discourage making things up, use chain‑of‑thought, or force tool use (e.g., code execution).
  • Sampling changes: newer truncation samplers (e.g., min_p, entropy‑based methods) aim to exploit logprob signals to avoid unlikely, error‑prone continuations.

Confidence, uncertainty, and self‑knowledge

  • Several comments stress the need for reliable “I don’t know” behavior; others note current confidence is about token likelihood, not ground‑truth.
  • Some research suggests internal states encode truthfulness cues that can be probed, but this is still dataset‑ and method‑dependent and far from a general solution.

Use cases, risk tolerance, and product fit

  • For low‑stakes tasks (summaries, drafts, simple Q&A), current error rates are seen as comparable to or better than many humans.
  • For high‑stakes domains (law, medicine, safety), hallucinations are considered unacceptable unless outputs are mediated by trained professionals and rigorous checks.
  • Distinction emphasized between “LLM-as-pattern-generator” and “commercial Q&A service”: in the latter, hallucinations are plainly product failures.

Terminology and user expectations

  • “Hallucination” as a term is contested; alternatives suggested include “error,” “inaccuracy,” or “confabulation.”
  • Some see anthropomorphic language as misleading; others see it as necessary shorthand for non‑experts.
  • A recurring concern is that users over‑trust fluent text; traditional cues for effort and competence are now cheaply faked, so education and UI affordances (e.g., uncertainty indicators) are seen as crucial.