Why AI-cited pitch decks still get facts wrong (Even with RAG)

The pitch deck trust crisis reflects a mismatch between what LLMs optimize for (plausible, complete narratives) and what business decisions require (verifiable, auditable truth).

Pitch decks used to lie with cherry-picked metrics and vague math. Now they lie with citations, thanks to LLM hallucination turning "sound right" into "look true."

AI hallucination, errors, and misinformation aren't edge cases; they're the default. Use AI to generate slides, and your deck can become a polished fraud without anyone intending it.

This post connects three things that people keep treating as separate problems:

  1. everyday summarization mistakes (like your report vs. ChatGPT summary comparison),
  2. the industry’s “fix” for those mistakes (Retrieval-Augmented Generation / RAG), and
  3. why even RAG can still amplify AI misinformation - just with better formatting and “sources.”

1. AI and the art of persuasion

When a human analyst is unsure, they hedge. When a model is unsure, it often completes. That’s the core mechanic behind:

  • AI hallucination: the model produces a claim that isn’t true.
  • AI errors: real facts get distorted (wrong units, wrong emphasis, wrong context).
  • AI misinformation: the output misleads a reader - even if it sounds professional.

And the most dangerous version isn’t “nonsense.” It’s a clean paragraph with confident tone, structured bullets, and numbers that look audited.

→ That’s how you get AI spreading misinformation at scale.


2. A real example: AI errors in a financial report

Most people imagine AI hallucination as the model inventing something out of thin air. But the more common failure mode is more dangerous:

You give the model a real report and ask for a summary

→ It returns something that looks professional - yet drifts from the source in ways that create AI misinformation.

How AI summary breaks trust

When an LLM summarizes, it compresses rather than shortens, and that compression introduces 3 predictable risks:

  • Unit + framing distortion (numbers stay similar but the meaning changes)
  • Fabricated structure (labels and metrics that never appear in reports)
  • Context drift + omission (what matters disappears or gets reinterpreted)

Below are screenshots of exactly what that looks like in practice:

1. Unit/label risk: When the source table is labeled one way (e.g., “in millions”), the summary may restate values in a different unit format. Even when the math is correct, the format shift creates audit confusion.

Wrong unit of money

2. Claim inflation risk: A summary can introduce labels like "GAAP operating income" that makes it look authoritative to human eyes - even when those terms don't appear in the source.

Making up a fake data

3. Context drift: The model can reinterpret product lines or timelines and turn “one sentence in a particular context” into a different claim.

Wrong context: Model Y is not a model, it’s a name of a new vehicle

4. Omission risk: The most important part of AI errors is what disappears - key constraints, qualifiers, or expansion plans that never make it into the summary.

Lack of important information

To reduce AI hallucination, the industry introduced RAG as a fix. But having numbers isn't proof, and having citations isn't safety. The real question: does RAG eliminate this risk or does it make the same AI errors look more credible?


3. Retrieval-Augmented Generation: A definition

Retrieval-Augmented Generation (RAG) is an AI framework designed to optimize the output of Large Language Models (LLMs) like ChatGPT by anchoring them to specific, external data.

→ Instead of the AI relying purely on its internal, "hazy" training data, RAG allows the model to look up facts in real-time before generating a response.

  • Without RAG: the model answers like a confident intern who “kind of remembers” the topic.
  • With RAG: the model is handed excerpts from your files first, then writes using those excerpts.

How it works?

A typical RAG pipeline consists of two primary stages that transform your query into a slide-ready claim with a lower AI error rate:

  1. Retrieval: Identify and select relevant documents from a "knowledge base," such as your company's historical financial data or legal repositories.
  2. Generation: The retrieved text is "fed" to the language model (like GPT-4) along with your original prompt. The LLM then synthesizes this information to create a response grounded in those specific documents.
Figure 1: RAG Work Process

The RAG Promise

The RAG Promise is the industry's attempt to fix the trust gap in generative AI. Leading providers market Retrieval-Augmented Generation as the "solution" to AI misinformation, built on three pillars:

  • Grounding in a "Closed Universe": Instead of the AI guessing from memory, the system is restricted to answering from a curated knowledge base.
  • Fact-First Retrieval: The system retrieves actual document snippets and "injects" them into the prompt, forcing the model to act like it’s taking an "open-book exam".
  • Linked Hallucination-Free Citations: Providers promise that every claim is anchored by a verifiable link to the source document, ensuring users can "rely upon outputs with confidence".

The reality is not “good enough”

Despite these promises, research shows RAG is not a "silver bullet". In real-world tests, even RAG-based tools hallucinated 17% to 33% of the time, often citing real sources to support completely false propositions.

  • Over 1 in 6 queries caused Lexis+ AI and Ask Practical Law AI to output misleading or false information.
  • Westlaw hallucinated substantially more: one-third of responses contained a hallucination.
Figure 2: Hallucination rates between AI Generated tools

This shows that even when AI provides citations, it doesn't protect you from hallucinations - it just makes AI misinformation look smarter → That's why "we use RAG" is not a guarantee of AI reliability.


4. Back to the main question: Why pitch decks amplify AI misinformation?

Pitch decks are short, high-stakes, and optimized for persuasion. They’re also full of exactly what LLMs are most likely to distort:

  • numbers without full context
  • unit conversions
  • implied causal claims (“because A, therefore growth”)
  • selective summarization (“key highlights”)
  • citation theater (links that look legit)

This is how generative AI misinformation happens in decks: not as one giant lie, but as dozens of tiny plausible claims that nobody has the time to audit → And the more “polished” the deck looks, the harder it is to feel the danger.


Summary

The pitch deck trust crisis reflects a mismatch between what LLMs optimize for (plausible, complete narratives) and what business decisions require (verifiable, auditable truth).

Eliminating AI hallucination altogether is unrealistic. LLM hallucination can still occur in plain summarization, and it can persist even with “grounded” approaches like RAG. Citations and numbers can make AI misinformation look smarter, not safer.

Which means:

  • When a deck can’t trace each claim back to a source → it can’t be trusted in high-stakes contexts.
  • When a deck can show sources → humans still need to verify that the sources actually support the claims.

Therefore:

AI governance must be designed in layers: retrieval, citation validity, claim traceability, and human review must work together. Because no single model can be judge, witness, and narrator at once

→ Closing this gap requires architectural safeguards, not better prompts.


Footnote

[1] OpenAI. Retrieval Augmented Generation (RAG) and semantic search for GPTs. January 5, 2026. https://help.openai.com/en/articles/8868588-retrieval-augmented-generation-rag-and-semantic-search-for-gpts[[help.openai](https://help.openai.com/en/articles/8868588-retrieval-augmented-generation-rag-and-semantic-search-for-gpts)]

[2] Magesh, V., Suriyakumar, V. G., Ho, D. E. Hallucination-Free? Assessing the Reliability of Leading AI-Powered Legal Research Tools. Journal of Empirical Legal Studies, 2025; 0

–27. https://dho.stanford.edu/wp-content/uploads/Legal_RAG_Hallucinations.pdf[viblo+1](https://viblo.asia/p/tim-hieu-ve-retrieval-augmented-generation-rag-Ny0VGRd7LPA)

Want to get on the
LayerProof waitlist early?

Contact us