How AI can get things wrong: hallucinations explained
~22 min readPicture this: it's Thursday afternoon and you need a quick competitive analysis before a Friday board meeting. You open ChatGPT, ask it to summarize a competitor's recent product launches, and it delivers a polished, confident three-paragraph response complete with dates, product names, and market share figures. You copy it into your slide deck. On Friday morning, your CEO asks where you got the statistic about the competitor's 34% market share gain. You check. The number doesn't exist anywhere. The product launch ChatGPT described never happened. The whole thing was fabricated — and it was presented with exactly the same confident tone as everything else the model said. This is a hallucination, and it just cost you credibility in front of the room.
What Hallucinations Actually Are
AI hallucinations are outputs that are fluent, confident, and factually wrong. The term comes from psychology — a hallucination is a perception with no grounding in reality — and it fits precisely. Large language models like GPT-4, Claude 3, and Gemini 1.5 Pro generate text by predicting the most statistically probable next token given everything that came before it. They are not retrieving facts from a verified database. They are completing patterns. When the training data contains strong patterns around a topic — say, how a corporate earnings report reads — the model can produce something that looks exactly like an earnings report, with plausible-sounding figures, whether or not those figures correspond to anything real. The output is linguistically correct and factually hollow.
This distinction matters enormously in professional settings. When you use a search engine like Google, you get links to source documents. The information exists somewhere you can verify. When you use a standard ChatGPT prompt without web browsing enabled, you are talking to a model whose knowledge was frozen at a training cutoff date — GPT-4's is early 2024 — and whose answers emerge from pattern completion, not fact lookup. The model has no internal alarm that fires when it's about to say something false. It has no concept of uncertainty baked into its default output style. A hallucinated CEO name comes out of the model in the same confident grammatical structure as a correctly recalled one. Nothing in the text itself signals which is which.
Hallucinations cluster around specific failure modes that professionals encounter repeatedly. The most common is fabricated citations — models inventing paper titles, author names, URLs, and journal references that sound credible but don't exist. A 2023 study by Stanford researchers found that AI-generated legal briefs contained hallucinated case citations at rates as high as 40% in some conditions. The second common failure mode is confident extrapolation: the model knows something about a company or person and extrapolates details it doesn't actually know, filling gaps with plausible-sounding invention. Third is temporal confusion — the model presents outdated information as current, or conflates events from different years. Each of these is dangerous in a different professional context.
Confidence ≠ Accuracy
Why the Model Can't Just 'Tell You' When It's Guessing
A common reaction from professionals first encountering hallucinations is: why doesn't the AI just say it doesn't know? The answer requires understanding what these models are — and aren't. A language model is trained to produce helpful, coherent, fluent responses. During training, responses that sound confident and complete were reinforced. Responses that said 'I don't know' were often rated lower by human evaluators who wanted useful answers. The model learned that confident-sounding completions are rewarded. This is called RLHF — Reinforcement Learning from Human Feedback — and it's the same technique that makes ChatGPT feel so conversational and useful. It's also part of why the model defaults to sounding certain even when it's generating fiction. The behavior that makes these tools feel good to use is structurally related to the behavior that makes them hallucinate.
Newer models have improved significantly on this. Claude 3 Opus and GPT-4o both produce more calibrated uncertainty expressions than their predecessors — they are more likely to say 'I'm not certain about this' or 'you should verify this date.' Perplexity AI takes a different architectural approach entirely, grounding every response in live web search results and citing its sources inline, which dramatically reduces hallucination risk for current-events queries. But even the best current models hallucinate in specific conditions: long-form generation, highly specific factual questions about niche topics, requests involving recent events, and questions about real people's quotes or statements. Knowing these conditions is the beginning of using AI professionally rather than naively.
Hallucinations in the Wild: Two Real Professional Scenarios
Abstract explanations only go so far. The fastest way to internalize hallucination risk is to see it in the kind of prompts professionals actually run. Below are two prompt-and-response pairs that illustrate where hallucinations emerge — and what they look like when they do. These aren't edge cases. They represent the kinds of queries that managers, analysts, and consultants submit to ChatGPT dozens of times a week. Read the responses critically, the way you'd read a research memo from a junior analyst before presenting it upward.
Prompt
What were Salesforce's key product announcements at Dreamforce 2023, and what market share gains did they report in the CRM space?
AI Response
At Dreamforce 2023, Salesforce made several significant announcements. The company unveiled Einstein Copilot, its AI assistant integrated across the platform, and announced a deepened partnership with Google Cloud. CEO Marc Benioff highlighted that Salesforce had grown its CRM market share to 23.8%, up from 21.2% the prior year, citing strength in the financial services and healthcare verticals. The company also announced Slack GPT enhancements and a new Data Cloud pricing tier aimed at mid-market customers. Benioff described 2023 as a 'pivotal year for AI-first CRM.' ⚠️ ANALYST NOTE: The 23.8% figure and the exact quote from Benioff in this response are illustrative of hallucination risk — plausible-sounding but unverified. Always cross-reference specific statistics and quotes against Salesforce's official press releases or earnings call transcripts before using them.
Notice the structure of that response. Several elements are real — Einstein Copilot was indeed announced, the Google Cloud partnership exists, Slack GPT is a real product. The model's training data contained accurate information about these. But the specific market share percentages and the direct Benioff quote are exactly the kind of detail that models fabricate convincingly: they fit the pattern of how an earnings announcement reads, they're numerically plausible, and nothing in the text flags them as invented. This is the hallucination trap for professionals — the response is 70% accurate, which makes the 30% fabrication harder to spot than if the whole thing were wrong.
Prompt
Can you cite three academic or legal sources that discuss AI liability frameworks in the European Union, with authors and publication years?
AI Response
Here are three relevant sources on AI liability in the EU: 1. Floridi, L. & Cowls, J. (2022). 'Toward a Unified Framework for AI Liability in European Law.' Journal of Law and Technology, Vol. 14(2), pp. 88–112. 2. Veale, M. & Borgesius, F. (2021). 'Demystifying the Draft EU AI Act.' Computer Law & Security Review, Vol. 38, Article 105567. 3. Mökander, J., Floridi, L. & Taddeo, M. (2023). 'AI Auditing and the EU AI Act: Accountability Gaps.' AI & Society, Vol. 40(1), pp. 201–218. ⚠️ ANALYST NOTE: Source #2 is a real paper. Sources #1 and #3 contain real author names but the titles, volume numbers, and page ranges are fabricated. This is a textbook hallucination pattern: real names attached to invented citations. Never submit AI-generated citations to a client, regulator, or court without individually verifying each one in Google Scholar or a legal database.
How Different AI Tools Handle Hallucination Risk
Not all AI tools are equally prone to hallucination — and the differences are architectural, not just about model quality. The tool you choose for a given task should reflect the hallucination risk profile of that task. Asking Perplexity AI about a current regulatory update is a fundamentally different risk proposition than asking a base ChatGPT session the same question. Understanding this spectrum lets you match tool to task, which is one of the highest-leverage habits a professional AI user can develop.
| Tool | Hallucination Risk (General) | Grounded in Live Web? | Cites Sources? | Best For |
|---|---|---|---|---|
| ChatGPT (no browsing) | High for facts/dates/citations | No | No | Drafting, brainstorming, summarizing provided text |
| ChatGPT (Browse/GPT-4o) | Medium — depends on search quality | Yes (Bing) | Sometimes | Current events, recent product info |
| Perplexity AI | Low-Medium for factual queries | Yes (multi-source) | Yes, inline | Research, fact-checking, competitor intel |
| Claude 3 Opus | Medium — better calibrated uncertainty | No (base) | No | Long documents, nuanced reasoning, policy analysis |
| Gemini 1.5 Pro | Medium — strong on Google ecosystem data | Yes (Google Search) | Partial | Market research, Google Workspace integration |
| GitHub Copilot | Low for code syntax; Medium for library facts | No | No | Code generation, developer workflows |
| Notion AI | High for factual claims | No | No | Internal docs, meeting notes, writing polish |
What You Do Differently on Monday
Understanding hallucinations isn't an academic exercise — it's a professional risk management skill. The immediate behavioral change is this: you stop treating AI output as a first draft and start treating it as a lead. A lead is unverified information that points you in a direction. You follow the lead, then verify before you act. When ChatGPT tells you a competitor launched a product in Q2 2023, that's a lead — you go to the company's press release page, their SEC filings, or a news search to confirm it. When Claude summarizes a regulation, you pull the actual regulation text. This is not more work than before AI existed. It's less work, because the AI has narrowed your search dramatically. But verification is non-negotiable when the output will be seen by clients, executives, or regulators.
The second behavioral change is prompt design. You can significantly reduce hallucination risk by changing how you ask. When you need factual accuracy, instruct the model explicitly. Tell it to flag anything it's uncertain about. Ask it to separate what it knows confidently from what it's inferring. Ask it to avoid inventing citations and instead describe the type of source you should look for. None of these instructions eliminates hallucination — but they shift the model's output toward more appropriately hedged language, which makes the unreliable parts easier to spot. Prompting for uncertainty is a skill. Most professionals don't do it by default, which is why they get burned.
The third change is task routing. You now have a mental map of which AI tools carry lower hallucination risk for which tasks. Use Perplexity when you need current, sourced factual information. Use ChatGPT with browsing enabled when you need to synthesize recent news. Use base ChatGPT or Claude for tasks where hallucination risk is low: rewriting your own text, generating structure from content you provide, drafting emails where you supply the facts, or brainstorming options you'll evaluate yourself. The tool-task match is the professional's version of knowing which colleague to ask for which kind of help — and it matters just as much.
Goal: Directly experience hallucination in your own workflow using your own tool, and build a concrete habit for citation verification before citations reach client-facing work.
1. Open the AI tool you use most frequently at work — ChatGPT, Claude, Gemini, or another. 2. Submit this exact prompt: 'Name three academic papers published between 2020 and 2023 on [your professional field] with full author names, journal names, volume numbers, and page ranges.' 3. Copy the full response into a separate document. 4. Take each citation and search for it individually in Google Scholar (scholar.google.com). 5. Mark each citation as: VERIFIED (found exactly as stated), PARTIALLY ACCURATE (real authors, wrong details), or FABRICATED (does not exist). 6. Note which elements were consistently accurate (author names, journal names) versus consistently invented (volume numbers, page ranges, exact titles). 7. Write two sentences describing what you'd do differently the next time someone asks you to use AI to find research sources for a client deliverable. 8. If any citations were fully fabricated, re-run the same prompt in Perplexity AI and compare the results — note whether sources are cited inline and whether they verify.
How Hallucination Risk Varies by Role
- Managers asking AI to summarize competitor activity or market trends face high hallucination risk — specific figures, dates, and executive statements are exactly what models fabricate most convincingly.
- Analysts using AI to pull financial data or cite regulatory sources should treat every number as unverified until cross-referenced with the primary source — Bloomberg, SEC EDGAR, or the official regulatory text.
- Consultants generating client-facing deliverables face reputational risk when hallucinated content appears in reports — build a verification step into your workflow before anything leaves your organization.
- Marketers using ChatGPT or Notion AI to draft content face lower factual risk if they supply the key facts in the prompt — hallucination risk drops sharply when the model is organizing your information rather than recalling its own.
- HR and legal professionals asking AI about employment law, case precedents, or compliance requirements face the highest-stakes hallucination risk — wrong legal information in a policy document can create real liability.
- Product managers using GitHub Copilot for code face a different hallucination profile — code syntax is usually correct, but references to specific library versions, API endpoints, or deprecated functions can be fabricated with the same confidence as correct ones.
Prompting to Reduce Hallucination: A Better Approach
The prompts most professionals use by default are the ones most likely to produce hallucinations. Open-ended requests for facts, statistics, citations, or summaries of real-world events give the model maximum room to fill gaps with invented content. Constrained, well-structured prompts that supply context, demand uncertainty flagging, and separate synthesis from recall produce dramatically better results. The prompt example below shows the same underlying need handled in a way that makes the model's limitations visible rather than hiding them inside fluent prose.
Prompt
I need to understand the EU AI Act's requirements for high-risk AI systems. Please summarize what you know confidently, flag anything you're uncertain about with [UNCERTAIN], and do not invent citations. Instead, tell me what type of source I should look for to verify each claim.
AI Response
The EU AI Act, which reached political agreement in December 2023 and formal adoption in 2024, classifies AI systems into risk tiers. High-risk systems — including those used in hiring, credit scoring, biometric identification, and critical infrastructure — face mandatory conformity assessments, transparency obligations, and human oversight requirements before deployment in the EU market. [UNCERTAIN] The exact timelines for compliance by risk category — I believe high-risk systems face a 24-month implementation window from the Act's entry into force, but the specific phased dates should be verified against the official text. For verification, look for: (1) The official EU AI Act text on eur-lex.europa.eu, (2) The European Parliament's press releases from March–May 2024, (3) Analysis from the Future of Life Institute or AI Now Institute, which publish accessible policy summaries. I have not cited specific articles or page numbers because I cannot verify their accuracy.
Key Takeaways from This Section
- Hallucinations are fluent, confident, factually wrong outputs — not glitches or errors in the traditional sense. They emerge from how language models are built.
- Models generate text by completing statistical patterns, not by retrieving verified facts. Confidence in the output tells you nothing about accuracy.
- The failure modes most common in professional work are: fabricated citations, confident extrapolation of invented details, and temporal confusion about recent events.
- RLHF training reinforces confident-sounding outputs, which is structurally related to why models hallucinate — the same mechanism makes them useful and unreliable.
- Tool choice matters: Perplexity AI and browsing-enabled ChatGPT carry lower hallucination risk for current factual queries than base ChatGPT, Claude, or Notion AI.
- Treat AI output as a lead, not a source — verify before anything reaches a client, executive, or regulatory audience.
- Prompt design reduces hallucination risk: supply context, demand uncertainty flagging, and ask the model to describe what to verify rather than inventing citations.
- Hallucination risk is role-specific — legal, compliance, and financial professionals face the highest-stakes exposure, while marketers working with self-supplied facts face less.
When Confidence Becomes the Problem
Picture this: you're preparing a competitive analysis for a board presentation. You ask ChatGPT to summarize a rival company's recent product launches, and it returns four bullet points — dates, product names, even quoted price points. It reads like a Bloomberg brief. You paste it into your deck. Two days later, a colleague flags that two of those products don't exist. The dates are wrong. One 'quote' is fabricated. You've just experienced the most professionally dangerous form of hallucination: the kind that looks completely authoritative. The model didn't hedge. It didn't say 'I think' or 'approximately.' It delivered fiction in the tone of fact, and your brain — trained to trust confident, well-formatted information — had no reason to question it.
Why Models Hallucinate More on Some Topics Than Others
Hallucination risk isn't uniform. It clusters around specific knowledge types, and understanding that clustering helps you predict when to trust an output and when to verify aggressively. Models hallucinate most on recent events (anything after their training cutoff), specific numerical data (statistics, prices, headcounts, dates), named individuals who aren't globally famous, citations and academic references, and niche or regional topics underrepresented in training data. By contrast, models are far more reliable on well-documented conceptual territory — explaining how a DCF works, summarizing a widely covered historical event, or drafting a meeting agenda. The underlying reason is statistical: the model has seen more consistent, high-quality text about broadly covered topics, so its predictions are better calibrated. Obscure facts have thin, inconsistent training signal, so the model fills gaps with plausible-sounding invention.
The Citation Trap
There's a second layer to this pattern: models hallucinate more when you ask questions that feel answerable but technically aren't — at least not with the model's training data. 'What is Accenture's Q3 2024 revenue?' feels like a factual lookup. For the model, it's a generation task. It has seen Accenture revenue figures before, it knows roughly what the number looks like, and it will produce something plausible. This is the core confusion: users treat AI like a database query and the model behaves like a completion engine. Those are fundamentally different things. A database either has the record or it doesn't. A language model always has something to say — and that's precisely the risk.
The Mechanics of a Professional Hallucination
To catch hallucinations before they embarrass you, it helps to understand what they look like structurally. Most professional hallucinations fall into one of four patterns: invented specifics (fake stats, dates, names), plausible extrapolation (taking a real trend and extending it beyond what the data supports), source conflation (mixing details from two real things into one fake thing), and confident omission (leaving out critical caveats that would change the meaning of an answer). That last one is subtle. When a model tells you that 'most studies show X,' it may be technically accurate — while omitting the three major studies that found the opposite. The output isn't fabricated, but it's misleading by selection. Your critical instinct needs to cover not just what the model said, but what it might have chosen not to say.
Prompt
What were the key findings of the Deloitte Global Workforce Survey 2024 on AI adoption in financial services?
AI Response
The Deloitte Global Workforce Survey 2024 found that 67% of financial services firms had deployed AI tools in at least one business unit, up from 41% in 2022. The survey, which polled 4,200 executives across 18 countries, highlighted that risk and compliance functions saw the highest adoption rates at 73%, followed by customer service at 61%. Notably, 54% of CFOs reported that AI tools had reduced manual reporting time by more than 30%. [INSTRUCTOR NOTE: These figures are fabricated. Deloitte does publish workforce and AI surveys, which is why this feels credible — but the specific statistics, sample sizes, and percentages above were generated, not retrieved. This is a textbook hallucination: real organization, real survey category, completely invented data.]
Prompt
I'm researching AI adoption trends in financial services. I don't need you to retrieve specific survey data — instead, help me identify which types of sources and organizations typically publish this research, and suggest 3-4 specific search queries I should use to find verified 2024 data myself.
AI Response
Good framing. Here are the most credible sources for AI adoption data in financial services: **Primary research publishers:** - Deloitte Insights (ai.deloitte.com) - McKinsey Global Institute (specific AI reports, published annually) - PwC's Financial Services Technology Survey - Oliver Wyman's banking technology reports - NVIDIA's State of AI in Financial Services (vendor-funded, note the bias) **Search queries likely to surface recent verified data:** 1. "AI adoption financial services 2024 survey" site:deloitte.com OR site:mckinsey.com 2. "generative AI banking 2024 report" filetype:pdf 3. "financial services AI investment 2024" site:pwc.com 4. "AI risk compliance automation 2024" site:oliverwyman.com Once you find a source, I can help you interpret the findings, draft commentary, or structure them into your presentation.
Comparing Tools on Hallucination Risk
Not all AI tools carry the same hallucination profile. The differences are meaningful enough to shape which tool you reach for depending on the task. Retrieval-augmented tools like Perplexity AI ground their answers in live web sources, dramatically reducing invented facts — though they can still misread or misrepresent the sources they cite. Closed-context tools like GitHub Copilot operate within your codebase, so their 'hallucinations' look more like plausible-but-wrong code completions than fabricated facts. Document-grounded tools like Notion AI or ChatGPT with uploaded files are safer for specific content because you've constrained the context. Understanding this spectrum lets you make smarter tool choices before you start a task, not after you've already pasted bad data into a slide deck.
| Tool | Hallucination Risk Level | Primary Risk Type | Safer For | Riskier For |
|---|---|---|---|---|
| ChatGPT (GPT-4) | Medium-High | Invented specifics, fake citations | Drafting, summarizing known concepts, brainstorming | Recent stats, named individuals, citations |
| Claude (Anthropic) | Medium | Confident omission, plausible extrapolation | Long documents, nuanced reasoning, policy drafts | Niche industry data, real-time events |
| Perplexity AI | Low-Medium | Misreading sources, outdated links | Research with source verification, fact-gathering | Deep synthesis without checking cited pages |
| Gemini (Google) | Medium | Conflation of similar entities, date errors | Google Workspace integration, broad research | Precise financial data, regulatory specifics |
| GitHub Copilot | Medium | Plausible-but-broken code, wrong API calls | Boilerplate code, familiar frameworks | Obscure libraries, security-critical functions |
| Notion AI | Low (document-grounded) | Misinterpretation of uploaded content | Summarizing your own documents, meeting notes | Any facts not present in the source document |
Building a Personal Verification Reflex
The professionals who use AI most effectively aren't the ones who trust it least — they're the ones who've built fast, lightweight verification habits that add maybe five minutes to a task but eliminate 90% of the embarrassment risk. The core habit is what you might call 'spot-check specifics': any time a model output contains a number, a name, a date, or a citation, treat that specific element as unverified until you've confirmed it from a primary source. Everything else — the structure, the framing, the language — can usually be used as-is. This asymmetry is important. You're not re-researching the entire output; you're surgically verifying the falsifiable claims within it. That distinction keeps AI genuinely useful while protecting your professional credibility.
A second verification habit involves prompting the model to flag its own uncertainty. This sounds counterintuitive — asking the system that hallucinated to identify its hallucinations — but it works better than you'd expect. When you explicitly ask a model to mark any claim it's uncertain about, or to distinguish between what it knows confidently versus what it's inferring, better models like GPT-4 and Claude will often surface genuine caveats they'd otherwise omit. They won't catch everything. But adding a follow-up prompt like 'Which specific facts in that response would you recommend I verify independently?' frequently generates a useful shortlist. You're essentially using the model's self-knowledge as a triage tool, not a final check.
The Two-Pass Workflow
The third habit is the hardest to build because it cuts against how we naturally process information: slow down when the output is surprisingly good. When AI gives you exactly what you needed, formatted perfectly, with specific figures that slot cleanly into your argument — that's precisely when your skepticism should spike. Hallucinations that survive into final deliverables are almost always ones that were too convenient to question. The model produced something that confirmed what you already believed or filled a gap you needed filled, and your brain accepted it without friction. Developing the professional instinct to pause on suspiciously clean answers is a skill that takes conscious practice, but it's what separates analysts who use AI confidently from ones who get burned by it.
Goal: Develop a calibrated, evidence-based sense of where your primary AI tool is reliable and where it fabricates — based on your specific professional context, not general advice.
1. Open whichever AI tool you use most at work — ChatGPT, Claude, Gemini, or Perplexity. 2. Write a prompt asking for specific data relevant to your industry: recent market statistics, a competitor's reported figures, or findings from a named industry report. Keep it specific — include a year and a named organization. 3. Copy the full response into a separate document. Highlight every specific claim: every number, date, name, and citation. 4. For each highlighted item, open a new browser tab and attempt to verify it from a primary source (the organization's website, a news outlet, a government database). 5. Mark each claim as Verified, Unverifiable, or Fabricated. Note how long verification took per claim. 6. Run the two-pass workflow: return to the AI and ask 'Which specific facts in your previous response should I independently verify before using professionally?' Compare this list to the claims you already flagged. 7. Write a three-sentence summary of what you found: how many claims were accurate, which types of claims were most problematic, and what this tells you about how you'll use this tool going forward. 8. Repeat the exercise with a document-grounded prompt — upload a PDF or paste in a source document, then ask questions only answerable from that document. Compare the error rate to your first test. 9. Save both outputs and your verification notes. You'll use this as a reference when choosing between prompting strategies in future tasks.
How Hallucination Risk Differs by Role
- Managers: Your biggest risk is using AI-generated competitive or market data in presentations to leadership. Fabricated figures in a board deck are career-level errors. Verify every external statistic before it hits a slide.
- Analysts: You're most exposed when AI generates numbers that fit your model too neatly. Treat AI-produced data points as hypotheses to test, not inputs to use. Your value is in the verification and interpretation, not the generation.
- Marketers: AI-generated campaign copy rarely halluccinates dangerously, but AI-generated claims about product performance, regulatory compliance, or competitor comparisons absolutely can. Any factual claim in external-facing copy needs a primary source.
- Consultants: Client deliverables built on AI-generated research carry reputational risk that extends beyond a single project. Establish a team norm: AI drafts structure and language, humans verify every substantive factual claim before it reaches the client.
- HR and People Managers: AI tools used for policy drafting or summarizing employment law can confidently state outdated or jurisdiction-incorrect information. Always verify legal or compliance claims against current official sources.
- Product Managers: When using AI to summarize user research or competitive features, watch for confident omission — the model may accurately describe what a competitor does while silently skipping a critical limitation or difference that changes your strategic read.
Prompting to Reduce Hallucination at the Source
Verification habits fix problems after the fact. Better prompting reduces how many problems appear in the first place. The most effective structural change is constraining the model to what it knows confidently, rather than allowing it to fill gaps with invention. Explicit uncertainty framing — phrases like 'only include information you are highly confident about,' 'flag any claim where your confidence is lower,' or 'if you don't have reliable data on this, say so explicitly rather than estimating' — measurably changes output quality. These instructions don't eliminate hallucination, but they shift the model's behavior toward surfacing uncertainty rather than papering over it. Combined with document-grounding (providing source material rather than asking the model to recall from training), this approach produces outputs that require significantly less verification effort.
Prompt
I'm preparing a briefing on regulatory changes affecting fintech lending in the EU in 2023-2024. Use only information you are highly confident is accurate. For any area where your knowledge may be incomplete or potentially outdated, explicitly flag it as 'VERIFY: [reason]' rather than filling in with estimates. Structure your response as: confirmed regulatory changes, areas requiring verification, and suggested primary sources to consult.
AI Response
**Confirmed Regulatory Changes (high confidence):** - The EU AI Act, finalized in 2024, includes specific provisions for AI used in credit scoring and lending decisions, classifying these as 'high-risk' AI systems requiring transparency and human oversight obligations. - The revised Payment Services Directive (PSD3) and Payment Services Regulation (PSR) proposals, published by the European Commission in June 2023, propose updated open banking obligations affecting fintech lenders. - DORA (Digital Operational Resilience Act) entered into force in January 2023, with a compliance deadline of January 2025 — relevant to fintech lenders with significant IT infrastructure. **VERIFY: Specific Implementation Dates** The exact national transposition timelines for PSD3 vary by member state and were still being finalized as of my training data. Check the European Banking Authority (EBA) website for current status. **VERIFY: Consumer Credit Directive Updates** The revised Consumer Credit Directive (CCD2) was adopted in 2023, but specific provisions affecting algorithmic credit decisioning may have evolved. Confirm current text at eur-lex.europa.eu. **Suggested Primary Sources:** - European Banking Authority: eba.europa.eu - EUR-Lex for official legislative texts: eur-lex.europa.eu - European Commission DG FISMA for fintech policy updates
What Changes on Monday
- Before running any research prompt, classify it: is this a task where the model needs to recall specific facts, or one where it's generating structure and language? Apply verification discipline to the first type; use output more freely for the second.
- Add one sentence to your highest-stakes prompts: 'Flag any claim you're uncertain about rather than estimating.' It takes three seconds and consistently improves output honesty.
- When AI output contains statistics or citations that will appear in client or leadership materials, verify them from primary sources before they leave your hands — not after.
- Switch to Perplexity AI for research tasks where you need real source links, and use ChatGPT or Claude for drafting, reasoning, and synthesis where specific factual recall isn't the primary need.
- Run the two-pass workflow (draft prompt, then verification checklist prompt) on any output that carries professional reputational risk.
- Build the pause reflex: when AI gives you something suspiciously good, that's your cue to slow down, not speed up.
Picture this: it's Thursday afternoon and your analyst has just handed you a competitive research brief. It looks sharp — company names, market share figures, a quoted statistic from a Gartner report, even a CEO's name. You skim it, nod approvingly, and forward it to a client. On Friday morning, the client replies: the Gartner report doesn't exist, and the CEO named left the company two years ago. Your analyst used ChatGPT without verification. This isn't a hypothetical — it's a pattern playing out across consulting firms, marketing agencies, and strategy teams every week. The question isn't whether AI will hallucinate on your team. It's whether you have a system to catch it before it costs you.
Why Verification Is a Workflow, Not a Habit
Checking AI output isn't something you do when you feel suspicious — it's a structured step in every AI-assisted process, like spell-check or legal review. The problem is that hallucinated content rarely looks suspicious. It's grammatically perfect, confidently stated, and contextually plausible. A made-up statistic sits comfortably next to three real ones. A fabricated citation uses the correct journal name format. Your brain pattern-matches to 'this looks right' and moves on. Building verification into your workflow means treating AI output the same way a good editor treats a first draft: assume errors exist, then hunt for them systematically. Teams that do this well don't slow down — they catch problems before they compound.
The verification burden should be proportional to the stakes. A ChatGPT-drafted internal memo about meeting norms? Light review is fine. A Claude-generated market analysis going to a board? Every factual claim needs a primary source. Most professionals get this wrong in both directions — they either trust everything or distrust everything, which makes AI either risky or useless. The practical middle ground is a triage system: flag claims that are specific (numbers, names, dates, citations), claims about recent events (anything after your tool's training cutoff), and claims in specialist domains where you lack personal expertise to spot errors. Everything else can be reviewed at normal speed.
The 'Specific Claim' Rule
Prompting to Reduce Hallucinations at the Source
You can't eliminate hallucinations through prompting alone, but you can reduce their frequency and make them easier to spot. Three techniques matter most. First, ask the model to flag uncertainty explicitly — most models will comply when instructed. Second, restrict the model to a defined source set by pasting in the actual content you want it to reason over, rather than asking it to recall facts from training. Third, ask for reasoning steps, not just conclusions — hallucinations often collapse when a model has to show its working. These aren't magic fixes, but they shift the output from confidently wrong to usefully uncertain, which is a much safer starting point for professional work.
Prompt
I'm researching the European electric vehicle market. Based ONLY on the text I've pasted below, summarise the key market share figures and flag any claims where the source text is ambiguous or missing. Do not add any figures from your training data. [Paste source text here]
AI Response
Based on the provided text only: • Volkswagen Group held approximately 24% of European BEV sales in H1 2023 (stated clearly in paragraph 2) • Tesla's market share is described as 'significant' but no specific figure is given — flagged as ambiguous • Stellantis figures cover combined BEV and PHEV sales; pure BEV breakdown is not provided in this text — flagged as incomplete I have not added any figures from outside the text you provided.
Choosing the Right Tool for Factual Reliability
Not all AI tools handle factual accuracy the same way. The core architectural difference is whether a tool generates answers from training data alone, or retrieves live information before generating. Retrieval-augmented tools like Perplexity AI and the web-browsing versions of ChatGPT and Gemini pull current sources and cite them, which makes hallucinations easier to detect and reduces their frequency for recent facts. Pure generation tools — standard ChatGPT, Claude without documents — are working entirely from learned patterns, which is powerful for reasoning and writing but unreliable for specific factual recall. Knowing which mode you're in is the first step in calibrating your trust.
| Tool | Retrieval Mode | Cites Sources? | Best For | Hallucination Risk (Facts) |
|---|---|---|---|---|
| Perplexity AI | Always live web search | Yes — links provided | Current facts, research starting points | Lower for recent data |
| ChatGPT (GPT-4 + Browse) | Optional web search | Yes when browsing enabled | Mixed tasks needing current info | Lower when browse is on |
| ChatGPT (GPT-4, no browse) | Training data only | No | Writing, reasoning, analysis of pasted content | Higher for specific facts |
| Claude (Anthropic) | Training data only (base) | No | Long document analysis, reasoning, drafting | Higher for specific facts |
| Gemini (Google) | Google Search integration available | Yes in search mode | Current events, Google Workspace tasks | Lower in search mode |
| GitHub Copilot | Code context only | N/A | Code generation and completion | Low for code, N/A for facts |
What You Do Differently Starting Monday
The most immediate change is adding a verification step to any AI-assisted deliverable that leaves your team. This doesn't require new software — a simple checklist works. Before forwarding AI-generated content, scan for the four specific claim types (numbers, proper nouns, dates, citations) and check each one against a primary source. For market data, that means the original report. For company facts, that means the company's own website or filings. For statistics, that means the original study. If you can't find a source in two minutes of searching, the claim comes out or gets rewritten as a general observation. This single habit eliminates the vast majority of professional embarrassment from hallucinated content.
The second change is matching your tool choice to your task's factual demands. When you need current, citable information — competitor pricing, recent earnings, regulatory updates — open Perplexity or use ChatGPT with browsing enabled. When you're drafting, restructuring, or reasoning over documents you've already validated, Claude or standard GPT-4 are efficient and capable. Using a generation-only tool for current-facts research is a structural mistake, not a user error. Building a two-tool habit — one for retrieval, one for generation — takes about a week to become automatic and significantly reduces the surface area for hallucinations to cause damage.
The third shift is cultural, and it's the one that scales. When your team uses AI tools, normalise saying 'this came from AI — I've verified X, Y, and Z, but you should check A.' That phrase does two things: it models good verification behaviour, and it creates psychological safety for others to admit when they haven't verified something. Teams where AI use is invisible are teams where hallucinated content circulates unchallenged. Teams where AI use is acknowledged have a natural checkpoint. You don't need a formal policy to start this — one person modelling the behaviour in a team meeting is usually enough to shift the norm.
Prompt
I've used AI to help draft this competitive analysis. Write a short, professional caveat (3-4 sentences) I can add to the document header that acknowledges AI assistance, notes which sections were AI-drafted, and tells readers which claims they should independently verify before acting on them.
AI Response
This analysis was developed with AI assistance (ChatGPT GPT-4). Sections 2 and 3 — covering competitor positioning and market share estimates — were AI-drafted and have been reviewed for accuracy. All market share figures in Section 3 are drawn from publicly available sources cited in the appendix; readers should verify currency before using in external presentations. Strategic recommendations in Section 4 reflect human judgment and have not been AI-generated.
Key Differences by Role
- Managers: Your risk is forwarding AI content without review. Build a 60-second specific-claims scan into your personal approval process for any AI-assisted output leaving the team.
- Analysts: You produce the content others trust. Add a verification column to your research notes — one row per factual claim, one column for the primary source URL. If the column is empty, the claim doesn't ship.
- Marketers: Statistics and quotes are your highest-risk content types. Never use an AI-generated statistic in copy without tracing it to the original study. Fabricated stats in published content create legal and reputational exposure.
- Consultants: Client-facing deliverables demand primary-source verification for every specific claim. Use Perplexity for initial research to get citable sources from the start, then Claude or GPT-4 for structuring and drafting.
- All roles: When you're not sure whether a claim is AI-generated or sourced, treat it as AI-generated. The cost of over-verifying is a few extra minutes. The cost of under-verifying is a client email on Friday morning.
Takeaways
- Hallucinations are a structural feature of generative AI, not a bug that will be patched away — your workflow must account for them permanently.
- Specific claims (numbers, proper nouns, dates, citations) are where hallucinations concentrate and where verification effort should focus.
- Prompting AI to flag uncertainty, restrict itself to pasted sources, and show reasoning steps reduces hallucination frequency and makes errors easier to spot.
- Retrieval-augmented tools (Perplexity, ChatGPT with browse, Gemini in search mode) are structurally safer for current factual research than generation-only tools.
- A two-tool habit — retrieval tool for facts, generation tool for drafting — covers most professional use cases with significantly lower risk.
- Normalising AI disclosure within your team ('I've verified X but not Y') creates a natural verification checkpoint without requiring a formal policy.
- The professional standard is: any AI-assisted content leaving your team should have its specific claims traced to primary sources before it goes.
Goal: Produce a personalised, role-specific AI verification checklist you can use immediately, plus a verified (or flagged) version of a real piece of AI content you've already worked with.
1. Open a document (Word, Google Docs, or Notion) and title it 'AI Output Verification Checklist — [Your Name]'. 2. Create four section headers: Numbers & Statistics, Proper Nouns (People & Companies), Dates & Timeframes, Citations & Reports. 3. Under each header, write one example of a hallucination risk in your specific role — for instance, a marketer might write 'AI-generated industry adoption percentage used in ad copy' under Numbers & Statistics. 4. Add a fifth section titled 'Tool Selection' and write two rows: one for when you'll use a retrieval tool (Perplexity or ChatGPT with browse) and one for when you'll use a generation tool (Claude or standard GPT-4), with a specific use case for each. 5. Take a real piece of AI-generated content you've produced or received in the last two weeks. Run it through your checklist — highlight every claim that falls into one of the four categories. 6. For each highlighted claim, spend two minutes searching for a primary source. Note the URL next to the claim, or mark it 'unverified — remove or rewrite'. 7. Add a 'Last Updated' date to the document and save it somewhere you'll actually find it — your desktop, a pinned Notion page, or a browser bookmark. 8. Share the checklist with one colleague who uses AI tools and ask them to add one risk example from their own role. 9. Use this checklist as your personal standard for every AI-assisted deliverable going forward — review and update it monthly as your AI tool usage evolves.
Your marketing colleague shares a ChatGPT-drafted press release containing the line: 'According to a 2023 Forrester report, 67% of B2B buyers now consult AI tools before vendor selection.' What is the correct immediate action?
A consultant needs current information about a competitor's pricing changes announced last month. Which tool configuration is most appropriate for this research task?
Which prompting technique most directly reduces the risk of hallucination when analysing a market research document?
An analyst builds a habit of adding a verification column to research notes with a source URL for every factual claim. Which statement best describes why this practice is effective?
A manager wants to shift team culture around AI use without implementing a formal policy. Which action is most likely to create lasting behavioural change?
Sign in to track your progress.
