Level 1 · Chapter 5.2

Detecting Hallucinations
& Errors

AI systems are phenomenal liars. They confidently assert information they invented. This is called "hallucination." Learn the red flags, master fact-checking techniques, and develop the discipline of verification before trusting any critical output.

Watch the Lecture

What Are Hallucinations

Hallucination is when an AI system confidently asserts information that is false or unsupported. The output is fluent, confident, and wrong. The system is not intentionally lying. It is pattern-matching text and generating statistically likely continuations without any mechanism for verifying whether those continuations are true.

Hallucinations are one of the hardest problems in modern AI. They happen because language models optimize for fluency and coherence, not truth. A fluent-sounding false statement is more likely to be generated than an uncertain correct statement.

Hallucinations Are Common

Do not expect hallucinations to be rare. They are common enough that you should assume any factual claim needs verification. This is not cynicism. This is professionalism. Assume nothing. Verify everything that matters.

Hallucination Red Flags

Specific Numbers Without Sources

If the output includes specific statistics or numbers ("87% of companies report improved productivity" or "the market was valued at $2.3B in 2024"), ask: where did that number come from? If there is no source, it might be hallucinated.

Confident Assertions About Recent Events

Language models have knowledge cutoffs (e.g., April 2024). If the output makes specific claims about events after that date, it is likely hallucinating. "In 2026, Company X announced..." when your model does not have 2026 data.

Proprietary or Hard-to-Verify Information

Claims about internal company information, private databases, or non-public information cannot be verified by you and are likely hallucinated. "According to [internal database], Company X has 50,000 employees" is probably made up.

Inconsistency Within the Same Output

If the output says one thing here and contradicts itself there, something is wrong. Maybe one part is hallucinated. Maybe the system changed its approach mid-response. Either way, inconsistency is a red flag.

Overly Confident About Uncertain Topics

For subjective matters, current events, or rapidly changing fields, watch out for false confidence. "The best approach is X" when the reality is "it depends." Uncertainty should sound uncertain.

Missing or Vague Sources

If you ask "Who said this?" or "Where is this from?" and the answer is vague ("some research shows..." "it is commonly believed..."), the system might be hallucinating. True sources are specific.

Techniques for Verification

Cross-Reference Against Authoritative Sources

For factual claims, verify against authoritative sources: academic databases, government sources, reputable news organizations, official company websites. For financial information, check SEC filings. For scientific claims, check peer-reviewed journals.

Check Multiple Sources

Do not verify against just one source. Check 2-3 authoritative sources. If they agree, you have confidence. If they disagree, investigate further.

Question Your Own Domain Knowledge

If you are an expert in a domain and the AI output contradicts your knowledge, the AI is likely wrong. Trust domain expertise. But also be humble: maybe you are outdated or the AI has information you do not.

Request Sources and Reasoning

Ask the model: "Where did you get that information?" or "What is the source for this?" If it cannot provide sources, be skeptical. Also ask: "Walk me through your reasoning" to understand how it reached conclusions.

Fact-Check Specific Claims

Do not assume everything is correct because most of it seems right. Take specific factual claims and verify them. This is especially important for: statistics, dates, names, technical specifications, and claims about third parties.

Fact-Checking Workflow

Step 1: Identify Verifiable Claims Look through the output and identify factual claims that can be verified. Ignore opinions and subjective analysis for now. Focus on facts.

Step 2: Categorize by Importance Which facts matter most? Prioritize verifying the critical claims first. Do not spend hours fact-checking minor details.

Step 3: Verify Using Authoritative Sources For each important claim, check against authoritative sources. Look for 2-3 independent confirmations.

Step 4: Note Discrepancies If you find any claim the output got wrong, note it. This tells you the output cannot be fully trusted and needs review.

Step 5: Assess Overall Reliability Based on how many claims checked out versus how many were wrong, decide: is this output reliable? Does it need fact-checking review? Should it be rejected?

Domain-Specific Verification

Technical Claims

For code, test it. For system design, check it against known best practices. For statistical claims, verify the methodology and assumptions.

Medical or Health Claims

Always verify against authoritative medical sources. Incorrect medical information is dangerous. Use PubMed, Mayo Clinic, CDC, etc.

Legal Claims

Verify against authoritative legal sources. AI frequently gets legal details wrong. Use official government sources, legal databases, or consult attorneys.

Historical Claims

For historical facts, use academic sources and multiple sources that agree. Be especially careful about recent history and controversial historical events.

Knowing Your Own Limits

You cannot personally verify everything. In domains where you are not expert, you need to either: consult with domain experts, use specialized verification resources, or accept more uncertainty in your evaluation.

The key is being honest about what you can and cannot verify. If you are not sure, say so. Do not pretend to have expertise you do not have.

Key Takeaway

Hallucinations are common. The solution is systematic verification. Build a habit of checking important facts before trusting them. Learn the red flags that indicate hallucinations. Master fact-checking workflows for your domains. Never assume that fluent writing indicates accurate information.

The standard for important outputs is: verify before use. Make this your default. Your credibility depends on catching hallucinations before they cause problems.

Chapter Details
Reading Time ~42 minutes
Difficulty Beginner
Prerequisites Chapter 5.1

Lesson 5 Chapters
5.1 Evaluation Framework
5.2 Detecting Hallucinations Current
5.3 Identifying Bias 5.4 Quality Decisions