12 January 2026
Why LLMs hallucinate
Hallucinations aren’t random — they’re a predictable failure mode of how language models generate text.
A large language model generates text by predicting the next token that looks most likely given the prompt and its training. That means it can sound confident even when it’s wrong.
The core reason
An LLM is not “retrieving facts” by default. It’s modelling plausibility. If your prompt nudges it toward an answer, it will produce one — even if the right response should be “I don’t know.”
Common triggers
- Missing context: the answer depends on facts not in the prompt.
- Ambiguous questions: multiple interpretations with no grounding.
- Overly broad requests: “give me everything about…”
- Incentives to be helpful: the model is optimised to respond.
What to do about it
Hallucination risk drops when you:
- constrain the task,
- provide grounded context (documents, policies, product data), and
- make abstention acceptable (e.g., “say you don’t know”).
In other words: treat hallucinations like a product quality attribute you can design for.
Want the quick version?
Grab the free guide: What is AI, GenAI, and how to avoid hallucination.
Get the guide