Case Study 4.1: The Confident Fabrication — How Alex Lost a Client Pitch
Background
Alex had been marketing manager at Clearbrook, a mid-sized e-commerce company selling outdoor lifestyle equipment, for three years. When the company decided to explore expanding its brand into subscription boxes — a model several direct competitors had already launched — Alex was asked to lead a competitive analysis and pitch the strategic recommendation to the executive team.
It was a high-stakes presentation. The CEO, CFO, and VP of Product would all be in the room. If Alex's analysis was compelling, the project would be greenlit and Alex would likely lead it. If it was not, someone else would own the space.
Alex had been using ChatGPT for about four months at this point, mostly for email drafts, social media content ideas, and campaign taglines. The tool had become a reliable time-saver. Alex had not yet encountered a significant error from it and had developed what turned out to be excessive confidence in its factual outputs.
The Preparation
With two days to prepare the pitch, Alex decided to use ChatGPT to accelerate the competitive research. The prompt was:
"Give me a detailed competitive analysis of the subscription box market in the outdoor/lifestyle category. Include market size, growth rates, the top five competitors, their subscriber counts, their price points, their key differentiators, and any recent notable developments in the space."
ChatGPT returned a substantial, impressive-looking response. It named specific competitors, cited specific subscriber numbers ("approximately 180,000 active subscribers"), mentioned specific price points ($49.99/month for one competitor, $34.99/month for another), and cited what appeared to be market research: "According to a 2023 IBIS World report, the subscription box market in the US grew 17.4% year-over-year, with the outdoor and lifestyle segment representing approximately $1.2 billion in annual revenue."
The response was formatted clearly, organized by competitor, with a confident, authoritative tone throughout. It looked exactly like the kind of competitive briefing an analyst might produce. Alex was impressed and relieved — this would have taken days of research to compile manually.
Alex copied the key figures into a slide deck, built out the recommendation section, and added a market opportunity framing based on the cited market size and growth rate.
One slide in particular read:
The Market Opportunity The U.S. outdoor subscription box segment represents a $1.2B market growing at 17.4% YOY. Leading competitors including [Competitor A] (180K subscribers) and [Competitor B] (95K subscribers) demonstrate proof-of-concept at scale. Clearbrook's brand recognition and existing customer base of 240,000 positions us to capture 2-3% market share within 24 months.
Alex did not independently verify any of these figures. They sounded right. The market felt that big. The growth rate seemed plausible for a category that had been growing visibly. And the format was so professional-looking that it was easy to treat it as having been researched.
The Pitch
The presentation started well. Alex walked through the strategic rationale clearly, and the format and structure were strong. The CEO was nodding. Then the CFO interrupted.
"Before we get to the financial model, I want to understand the market size basis. Where does the $1.2 billion figure come from?"
Alex said: "From an IBIS World report on the subscription box market."
The CFO frowned. "I pulled an IBIS World subscription box report last month for a different project. I don't recall that figure, and I don't think they break it out by outdoor specifically. Can you pull up the original report?"
Alex did not have the original report. There was no original report. The citation had been invented by ChatGPT.
"I'll need to track it down," Alex said. "I can follow up after the meeting."
The CFO moved on, but the tenor of the room had changed. The CEO asked about the subscriber count for the competitor Alex had cited — 180,000 active subscribers. "That seems high. Last I heard, they were much smaller. How recent is this data?"
Alex did not know. The session ended with the executive team declining to make a decision until the competitive data was verified and sourced.
The Aftermath
Over the next two days, Alex attempted to verify every figure in the presentation. The results were sobering:
- The $1.2 billion market figure did not appear in any identifiable IBIS World report. IBIS World did have reports on subscription boxes but categorized the data differently.
- The 17.4% YOY growth figure appeared in no verifiable source.
- The subscriber count for the first competitor was approximately 45,000 — not 180,000. The 180,000 figure was four times the verifiable estimate.
- The subscriber count for the second competitor was unverifiable because the company was private and had not disclosed subscribers.
- One of the five "competitors" named by ChatGPT had ceased operations 18 months earlier.
When Alex went back to ChatGPT and asked where it had gotten these figures, the response was vague: "I should note that specific subscriber counts and market statistics may not be precisely accurate, as my training data has limitations and companies often do not disclose exact figures publicly."
The presentation was rescheduled. Alex spent the next week doing the research properly — using IBIS World directly, checking companies' press releases, looking at SEC filings for public competitors, and triangulating subscriber estimates from available data. The real numbers told a somewhat different story: the market was large but growing more slowly than the ChatGPT estimate, and the leading competitors were smaller than claimed.
The eventual pitch, with verified data, was approved. But the delay cost three weeks, Alex's credibility took a noticeable hit, and the incident was referenced — with some diplomatic handling — in Alex's next performance review.
Analysis: What Went Wrong and Why
The core failure: treating Zone 3 output as Zone 1. Market statistics, specific competitor data, and research citations are all Zone 3 tasks — high risk of fabrication, requiring independent verification. Alex implicitly treated this output as Zone 1 — high reliability, use directly. This miscalibration was the root cause.
Why the miscalibration happened: Several factors contributed.
-
Recency bias from good experiences. Alex had used ChatGPT reliably for months on Zone 1 tasks (email drafts, taglines, content ideas) without errors. This created an implicit assumption that the tool was generally reliable that did not account for task-specific reliability variation.
-
The professionalism heuristic. The ChatGPT output looked like analyst-quality research. Bullet points, competitor-by-competitor breakdown, specific numbers with decimal places and units. The format triggered the association "this looks like what professional research looks like, therefore it is professional research." This is the fluency-accuracy gap in action.
-
Time pressure. Two days to prepare a major pitch is tight. The temptation to treat a fast, good-looking output as research rather than as a starting point for research was real.
-
The citation anchor. The specific mention of "IBIS World report" was particularly dangerous. It was not just a statistic — it was a citation that appeared to root the statistic in a real, verifiable source. Alex did not think to verify whether that citation was real before treating the statistic as sourced.
The cascading effect. One error does not always stay isolated. Once Alex accepted the $1.2B market figure, the 17.4% growth rate, and the competitor subscriber counts as real, the financial modeling built on top of them was also wrong. The "2-3% market share in 24 months" calculation was predicated on a market size that did not exist. A single set of hallucinated inputs propagated throughout the analysis.
What Should Have Happened
At prompt time: Alex should have recognized that competitive market statistics and specific company data are Zone 3 tasks. A mental note to verify all specific figures before including them in any deliverable.
At output time: Rather than copying the figures directly into a slide, Alex should have opened a new document and listed every specific factual claim as a "to verify" item: market size, growth rate, each competitor's subscriber count, each pricing figure, the IBIS World citation.
At verification time: Even with two days of time pressure, 90 minutes of verification effort would have caught the errors. Searching IBIS World for the report directly, checking competitor company pages and press releases for subscriber data, and cross-referencing the market size against other sources would have revealed the fabrications.
At prompt refinement time: A better prompt would have helped: "I'm going to look up market research myself. Help me identify what specific data points I should be looking for and what sources typically publish subscription box market data." This uses AI as a research-planning tool rather than a research-generating tool for Zone 3 tasks.
Calibration Updates Alex Made
After this incident, Alex made several lasting calibration changes:
-
The verification rule for statistics. Any specific numerical claim — percentages, market sizes, subscriber counts, growth rates — goes on a verification list before being used in any deliverable. No exceptions for time pressure.
-
The citation existence check. Any named source (report, study, publication) gets Googled before it becomes part of a presentation. If it cannot be found, it does not get cited.
-
The competitor data protocol. For competitive analysis, Alex now starts with primary sources: company websites, press releases, SEC filings for public companies, industry trade press, and market research subscriptions. AI can help structure the analysis and identify what to look for, but the data itself comes from verified sources.
-
The Red Flag addition. "ChatGPT fabricates specific statistics and citations in competitive analysis" was added to Alex's personal Red Flag list. It remained there permanently.
The mistake was costly in the short term. In the long term, it produced a calibration update that prevented more costly errors down the road.
Key Lessons
-
Fluency is not accuracy. Professional-looking, well-structured output with specific numbers and citations can still be entirely fabricated. The quality of the presentation is no indicator of the quality of the underlying facts.
-
Zone 3 tasks require verification regardless of time pressure. Time pressure is the most common reason people skip verification — and the reason the stakes are highest when they do.
-
Citations require existence verification, not just plausibility checking. "IBIS World report" sounds like a real source. The only way to know it is a real source is to find it.
-
Trust calibration errors compound. One set of fabricated inputs ripples through every analysis built on top of them. The sooner you catch a Zone 3 error, the smaller the compounding effect.
-
Experience with reliable AI tasks does not generalize. Reliable performance on Zone 1 tasks tells you nothing about Zone 3 task reliability. Each zone must be calibrated independently.