What Is AI Hallucination? How to Spot It and Stop It
TL;DR: AI hallucination is when a model generates confident but wrong information. It happens because language models predict plausible text, not truthful text. The most reliable way to catch hallucinations is to compare answers across multiple AI models. When models disagree, something is wrong.
AI Hallucination, Explained Simply
You ask an AI model a question. It gives you a clear, detailed answer. The problem? The answer is completely made up.
That's AI hallucination. The model didn't misunderstand your question. It generated text that sounds right based on statistical patterns in its training data, but the facts are wrong. It might invent a legal statute, cite a study that doesn't exist, or confidently state that a product has features it lacks.
This isn't a rare edge case. In our evaluation of real queries, factual accuracy scores between the best and worst model on the same question differed by an average of 5.8 points on a 10-point scale. One model gets it right. Another invents the answer. And both sound equally confident.
Why Do AI Models Hallucinate?
Language models don't "know" anything. They predict the most probable next word based on patterns learned from training data. When the training data is rich (common topics, well-documented facts), the predictions tend to be accurate. When it's thin (rare languages, niche regulations, recent events), the model fills in gaps with plausible-sounding fiction.
There's no internal fact-checker. The model that gives you a perfect answer about Python syntax uses the exact same mechanism to fabricate a Botlikh language phrase or invent a tax law that doesn't exist. It has no way to distinguish between "I know this" and "I'm guessing."
Real Examples: When AI Gets It Wrong
We tested 32 AI models on identical questions across 8 domains. Here's what we found:
- Tax law: Three models correctly identified Massachusetts' source-state taxation rules for 457(b) deferred compensation. One model stated the exact opposite, which could cost a retiree thousands of dollars. See the full comparison.
- Rare languages: Asked for common phrases in Botlikh (a language spoken by ~5,000 people), one model honestly refused. Another confidently invented phrases that don't exist in any documented source. See the full comparison.
- Product specifications: When asked if the MikroTik GCC6010 has a built-in RADIUS server, three models correctly said no. One model claimed it does, which would lead to a failed network deployment. See the full comparison.
- Historical facts: Two models accurately identified details about Darlene Diebler Rose's WWII internment. One model confused key facts about the camps and locations. See the full comparison.
These are not cherry-picked failures. They come from routine queries that professionals ask every day. For more documented cases, see our full breakdown of when AI gets it wrong.
How to Catch AI Hallucinations
The single most effective technique is multi-model verification. Instead of trusting one model's answer, send the same question to several models and compare their responses.
When four models agree and one disagrees, you know where to focus your verification effort. When all models disagree, you know the topic needs human expertise.
This is exactly what Trust Score measures. Our Ensemble Disagreement metric quantifies cross-model consensus on every query. High agreement increases confidence. Significant disagreement flags potential hallucination.
Other detection strategies include:
- Ask for sources. If a model cites a specific study or statute, verify it exists.
- Test edge cases. Hallucination risk increases with rare topics, recent events, and specific numbers.
- Check consistency. Ask the same question in different ways. If the answers contradict each other, at least one is fabricated.
For a step-by-step process, see our guide on how to fact-check AI answers.
How Trust Score Measures This
Trust Score evaluates every AI response across 7 metrics. Two are directly relevant to hallucination detection:
- Factual Accuracy (FA): Are the facts verifiable and correct? This is the hardest metric for AI models, with scores ranging from 0.0 to 8.9 across our 2,637 evaluations. See the full FA rankings.
- Ensemble Disagreement (ED): When multiple models answer the same question, do they agree? This metric is unique to multi-model platforms and directly measures the signal that catches hallucinations.
Frequently Asked Questions
What is AI hallucination?
AI hallucination is when a language model generates information that sounds confident and plausible but is factually incorrect. This includes invented statistics, fake citations, wrong dates, and fabricated product specifications.
Why do AI models hallucinate?
AI models predict the most likely next word in a sequence based on training data patterns. They have no mechanism for verifying whether their output is true. When training data is sparse on a topic, models fill gaps with statistically plausible but invented information.
How can you detect AI hallucination?
The most effective method is multi-model verification: asking the same question to multiple AI models and comparing their answers. When models disagree, it flags potential hallucination. Trust Score measures this through the Ensemble Disagreement metric, which tracks cross-model consensus.
Which AI models hallucinate the most?
Based on Trust Score evaluations of 32 models, factual accuracy scores range from 0.0 to 8.9 on a 10-point scale. Larger models from OpenAI, Google, and Anthropic tend to score higher on factual accuracy, while smaller or specialized models show more variation.