> — Adapted from a principle in The Black Swan by Nassim Nicholas Taleb
Learning Objectives
- Define survivorship bias and explain why it is a structural feature of evidence collection, not just individual carelessness
- Identify survivorship bias operating in at least four different domains
- Apply Abraham Wald's insight: the evidence you DON'T see is often more important than the evidence you do
- Analyze how publication bias creates survivorship bias in scientific knowledge at industrial scale
- Add the survivorship bias lens to your Epistemic Audit
In This Chapter
- Chapter Overview
- 5.1 Abraham Wald and the Missing Bullet Holes
- 5.2 The Business Success Literature: Studying Winners, Ignoring Losers
- 5.3 The Architecture of Ancient Knowledge: What Survived Isn't Representative
- 5.4 Publication Bias: The File Drawer Problem
- 5.5 Startup Mythology and the Denominator Problem
- 5.6 What It Looked Like From Inside
- 5.7 Active Right Now: Where Survivorship Bias May Be Operating
- 5.8 Mutual Funds and the Advertising of Survivors
- 5.9 The Antidote: Learning to See What Isn't There
- 5.10 Practical Considerations: Living With Biased Evidence
- 5.11 Chapter Summary
- Spaced Review
- What's Next
- Chapter 5 Exercises → exercises.md
- Chapter 5 Quiz → quiz.md
- Case Study: Abraham Wald and the Art of Seeing What Isn't There → case-study-01.md
- Case Study: Publication Bias in Antidepressant Research → case-study-02.md
Chapter 5: Survivorship Bias at Scale
"The living tell their tales. The dead do not." — Adapted from a principle in The Black Swan by Nassim Nicholas Taleb
Chapter Overview
In 1943, the U.S. military had a problem. Its bombers were being shot down over Europe at alarming rates. To improve their chances, the military wanted to add armor to the planes — but armor is heavy, and adding it everywhere would make the planes too sluggish to fly effectively. They needed to know where to put the armor.
The obvious approach: examine the bombers that returned from missions, catalog where they had been hit, and armor those areas. The Statistical Research Group at Columbia University was brought in to analyze the data. Abraham Wald, a mathematician who had fled Austria after the Nazi annexation, was assigned to the problem.
The returning bombers showed a clear pattern. They had bullet holes concentrated in the fuselage, the fuel system, and the wings. The engines had relatively few hits. The military's conclusion seemed straightforward: armor the fuselage, fuel system, and wings — the areas taking the most damage.
Wald's conclusion was the opposite: armor the engines.
His reasoning was elegant and devastating. The data came only from planes that had survived. The planes that had been hit in the engines hadn't come back. The bullet holes on the returning planes showed where a bomber could be hit and still fly home. The areas with no bullet holes — the engines — were the areas where hits were fatal.
The military was looking at the survivors and drawing conclusions about what killed the non-survivors. The evidence they had was precisely the evidence that didn't matter. The evidence that mattered — the location of hits on downed planes — was at the bottom of the English Channel. It was literally invisible, irrecoverable, and yet it contained exactly the information the military needed to save lives.
Wald's insight saved an unknown number of aircrew members. The armor was placed on the engines, where it could prevent fatal damage. The returning bombers' bullet holes — the visible evidence that had almost led to the wrong decision — were, for the first time, correctly understood as evidence of survivability, not vulnerability.
This story has been told many times, and it deserves every retelling. It is the single clearest illustration of a pattern that corrupts knowledge production across every domain.
This is survivorship bias: the systematic error of drawing conclusions from data that survived a selection process while ignoring the data that was filtered out. And it is the fourth major entry mechanism for wrong ideas in knowledge production.
In this chapter, you will learn to: - Recognize survivorship bias as a structural feature of how evidence is collected and preserved - Apply Wald's insight: ask what evidence is missing, not just what evidence is present - Identify survivorship bias in scientific publishing, business strategy, historical knowledge, and everyday reasoning - Understand publication bias as survivorship bias operating at industrial scale in science - Add the "missing evidence" lens to your Epistemic Audit
🏃 Fast Track: If you're familiar with survivorship bias and publication bias, skim to section 5.4 (Publication Bias: The File Drawer Problem) for the science-specific treatment, and then section 5.6 for the cross-domain analysis. Exercise B.2 tests application.
🔬 Deep Dive: After this chapter, explore the Cochrane Collaboration's work on publication bias detection methods, and read Ben Goldacre's Bad Pharma for the most thorough account of how publication bias distorts medical evidence.
5.1 Abraham Wald and the Missing Bullet Holes
Let's stay with Wald's insight a moment longer, because it contains — in compressed form — the entire logic of survivorship bias.
The Structure of the Error
The military analysts who wanted to armor the fuselage weren't stupid. They were doing exactly what empirical analysis is supposed to do: collecting data, identifying patterns, and drawing conclusions. The data was real. The patterns were genuine. The bullet holes were exactly where they were.
The error was not in the analysis of the data. It was in the selection of the data. The sample — returning bombers — was not representative of all bombers. It was representative of surviving bombers, which is a fundamentally different population. The difference between the two populations — the planes that came back and the planes that didn't — contained exactly the information the military needed.
This is the deep structure of survivorship bias: the process that selects which evidence reaches you is invisible, but it determines the shape of everything you see.
Consider a more precise formulation. Every body of evidence you encounter has been through at least one selection filter. The question is: does the filter correlate with the thing you're trying to learn about? If the filter is random — if evidence is missing at random — the bias is minimal. But if the filter is systematic — if the evidence that's missing is systematically different from the evidence that survived — then every conclusion you draw from the surviving evidence is distorted in a specific, predictable direction.
Wald's bombers illustrate the strongest case: the filter (survival) was directly correlated with the variable of interest (vulnerability). Planes that were hit in vulnerable locations didn't survive, so they weren't in the sample. The surviving evidence was anti-informative about vulnerability — the very thing the military wanted to know.
This anti-informative property is what makes survivorship bias so dangerous. It's not just that the evidence is incomplete. It's that the pattern of incompleteness systematically distorts conclusions in a specific direction. And the direction is predictable: survivorship bias always makes things look better than they are, because the failures have been removed from the record.
When you study successful companies, you see the strategies of winners. You don't see the identical strategies of companies that failed using the same approaches. When you read published research, you see the studies that produced significant results. You don't see the studies that produced null results and were never published. When you study ancient civilizations, you study the ones that left written records. You don't study the equally complex civilizations that didn't.
In every case, a selection process has filtered the evidence before it reaches you. And in every case, the filtered-out evidence may be more informative than the evidence that survived.
💡 Intuition: Imagine you're trying to understand why some people live to be 100. You interview centenarians and ask about their habits. One says she drank a glass of whiskey every day. Another says he never exercised. A third says she smoked for 50 years. You might conclude that these behaviors are compatible with longevity — or even cause it.
But you haven't interviewed the vastly larger number of people who drank whiskey every day, never exercised, or smoked for 50 years and died at 60. The centenarians are the survivors. Their habits tell you about what's compatible with extreme luck, not about what causes longevity. The evidence you need is in the cemetery, not in the nursing home.
The Wald Insight as Diagnostic Template
Wald's approach can be formalized into a diagnostic template that applies to any domain:
-
Identify the evidence in front of you. What data do you have? (Bullet holes on surviving planes. Published studies. Successful companies. Ancient texts.)
-
Identify the selection process. What filter produced this evidence? (Only surviving planes returned. Only positive studies are published. Only successful companies are studied. Only durable records survived.)
-
Ask what the filter excluded. What evidence would have been generated by the cases that didn't survive? (Bullet holes on downed planes — concentrated on engines. Null studies — showing treatments don't work. Failed companies — with the same characteristics as successes. Oral cultures — with different social structures.)
-
Determine whether the filter correlates with your variable of interest. Does the selection process systematically exclude the evidence you most need? If yes, your conclusions from the surviving evidence are not just incomplete but actively misleading.
-
Adjust your conclusions accordingly. If the missing evidence would change your conclusions, state that uncertainty explicitly rather than treating the surviving evidence as the full picture.
This template — Wald's template — is one of the most powerful general-purpose diagnostic tools in epistemology. We'll return to it repeatedly throughout the book.
What Made Wald Exceptional
Wald's genius was not mathematical — the logic of his insight is accessible to anyone. His genius was seeing the selection process. While everyone else looked at the data and asked "where are the bullet holes?", Wald looked at the data and asked "where are the bullet holes not?"
This is the fundamental skill that this chapter aims to develop: the ability to look at evidence and ask, "What evidence am I not seeing? What got filtered out? And would the filtered-out evidence change my conclusions?"
🧩 Productive Struggle
Before reading the next section, try this exercise: Think of a recent conclusion you drew from evidence. Now ask: "What evidence might I be missing? What got filtered out before it reached me?" List at least three possible sources of missing evidence.
The difficulty of this exercise — the struggle to see what's not there — is itself the point. Survivorship bias is hard to detect precisely because the missing evidence is, by definition, absent.
5.2 The Business Success Literature: Studying Winners, Ignoring Losers
Perhaps no domain is more thoroughly corrupted by survivorship bias than the business success literature.
Consider the genre: books that study successful companies to extract the secrets of their success. In Search of Excellence (Peters and Waterman, 1982), Built to Last (Collins and Porras, 1994), Good to Great (Collins, 2001). These books are perennial bestsellers. They are assigned in MBA programs. They have shaped management thinking for decades.
They are also, structurally, survivorship bias factories.
The Method and Its Flaw
The typical method: select companies that have achieved exceptional performance, study their characteristics, and identify the common factors. Good to Great selected 11 companies that made the transition from good to great, studied their management practices, and identified factors like "Level 5 Leadership," "the Hedgehog Concept," and "a Culture of Discipline."
The problem: the study doesn't include companies that had the same characteristics but didn't become great. Without this comparison group — the denominator, the base rate, the non-survivors — you can't determine whether the identified factors actually caused success or whether they're simply common among companies in general (including companies that failed).
The Aftermath
The survivorship bias in Good to Great was not just theoretical. Of the 11 companies identified as "great" in 2001:
- Circuit City filed for bankruptcy in 2008 and was liquidated
- Fannie Mae required a government bailout in 2008
- Wells Fargo became embroiled in a massive fraud scandal involving millions of fake accounts
- Several others significantly underperformed the market in subsequent years
This doesn't mean Collins's observations were wrong about what those companies were doing at the time. It means that the factors he identified were not sufficient to predict or guarantee sustained success — because the methodology couldn't distinguish between factors that cause success and factors that are merely present during success.
This pattern — retrospective identification of "key success factors" from a survivor-biased sample — is sometimes called the halo effect (a term Phil Rosenzweig uses in his excellent critique, The Halo Effect). When a company is doing well, everything about it looks good: its strategy is "bold," its culture is "strong," its leadership is "visionary." When the same company falters, the same characteristics are relabeled: the strategy was "reckless," the culture was "cult-like," the leadership was "out of touch." The evaluations are driven by the outcome, not by independent assessment of the characteristics.
The same critique applies to virtually all business success literature. Studying winners and ignoring losers guarantees that you will "discover" factors that are correlated with survival but not necessarily causes of it. It's Wald's bullet holes on the returning bombers: you're studying where the survivors were hit, not where the non-survivors were hit.
🔄 Check Your Understanding (try to answer without scrolling up)
- What is the fundamental flaw in studying only successful companies?
- What would a non-survivorship-biased study of business success look like?
Verify
1. Without studying companies that had the same characteristics but failed, you can't determine whether the identified factors caused success or are merely common among all companies (including failures). 2. It would include both successful and unsuccessful companies, compare their characteristics, and identify factors that are significantly more common among the successful group — controlling for the base rate.
5.3 The Architecture of Ancient Knowledge: What Survived Isn't Representative
Survivorship bias doesn't just affect modern fields. It shapes our understanding of the entire human past.
The Literacy Bias
The survivorship bias in historical knowledge is so pervasive that it shapes our most basic theories about human civilization — often without anyone noticing.
Our knowledge of ancient civilizations is overwhelmingly biased toward literate cultures. We know vastly more about ancient Egypt, Greece, Rome, China, and Mesopotamia than about contemporary civilizations in sub-Saharan Africa, pre-Columbian South America, Southeast Asia, or the Pacific Islands — not because these civilizations were less complex or less interesting, but because the literate civilizations left written records that survived.
The result: our picture of the ancient world is dominated by a small number of literate, urban, state-organized societies. This biases our theories of social organization, political development, economic evolution, and cultural achievement toward the characteristics of these surviving-evidence civilizations. Theories of "how civilizations develop" are really theories of "how civilizations develop when they happen to leave the kind of evidence that survives for millennia."
The Building Bias
When architectural historians study "great buildings," they study buildings that survived — cathedrals, pyramids, temples, castles. These are the most durable construction types, built with the most resources, for the most powerful patrons. They are not representative of the buildings that actually constituted the built environment. The vast majority of historical buildings — houses, workshops, markets, inns, warehouses — were built of wood, mud brick, or other perishable materials and have long since disappeared.
Our theories of architectural history, construction technique, and urban form are built almost entirely on the evidence that survived, which represents a tiny, unrepresentative fraction of what was built. It's as if we tried to understand modern transportation by studying only Rolls-Royces — the vehicles that get preserved in museums — while ignoring the Toyotas and bicycles that move most people.
The Music and Art Bias
The survivorship bias in our understanding of historical culture is profound. When we speak of "classical music," we mean the tiny fraction of composed music from previous centuries that survived: compositions that were written down, preserved in institutions (churches, courts, conservatories), and deemed worthy of continued performance by later generations. The vast majority of music — folk songs, improvised performances, dance music, work songs, lullabies, religious chants from minority traditions — was never notated and has been lost.
Our understanding of "what music was" in previous centuries is therefore dominated by the most formally structured, institutionally preserved, elite-patronized compositions. It's as if future historians tried to understand 21st-century music by studying only the works performed by major symphony orchestras — missing hip-hop, pop, electronic music, folk, jazz, and everything that wasn't formally notated and institutionally archived.
The same applies to visual art (we study what was collected and preserved, not what was created), literature (we study what was published and reprinted, not what was written), and philosophy (we study thinkers whose works survived copying and translation, not the full landscape of philosophical thought).
The Great Man Bias
History is written about people whose stories survived. These tend to be the powerful, the victorious, the literate, and the well-connected. The perspectives of the poor, the defeated, the illiterate, and the marginalized are systematically underrepresented — not because their lives were less significant, but because their evidence didn't survive the selection process.
This creates a systematic distortion in our understanding of how historical change occurs — a distortion that has shaped historiography, political theory, and our collective understanding of what it means to be human in organized society. "Great man" theories of history — the idea that history is driven by exceptional individuals — are partly an artifact of survivorship bias: the stories of exceptional individuals are the ones that survive. The collective actions, social movements, and structural forces that drive most historical change leave less durable evidence and are therefore less visible in the surviving record.
📜 Historical Context: The field of subaltern studies, pioneered by historians like Ranajit Guha, explicitly addresses this survivorship bias by attempting to reconstruct the perspectives and agency of people whose voices were filtered out of the historical record. The challenge is immense — how do you study what isn't there? — but the recognition that the surviving evidence is biased is itself a crucial epistemic advance.
5.4 Publication Bias: The File Drawer Problem
If survivorship bias in business literature and ancient history are concerning, publication bias in science is terrifying — because scientific knowledge is supposed to be our most reliable form of knowledge, and publication bias systematically distorts it.
The Mechanism
The file drawer problem, named by psychologist Robert Rosenthal in 1979, works like this:
- A researcher conducts a study testing whether Drug X reduces symptom Y.
- If the result is positive (Drug X works, p < 0.05): the study is published in a journal, cited by other researchers, and incorporated into the evidence base.
- If the result is negative (Drug X doesn't work, p > 0.05): the study is filed away in the researcher's drawer, never published, never cited, and never incorporated into the evidence base.
The result: the published literature systematically overrepresents positive findings and underrepresents null findings. If ten independent research groups each test Drug X, and only one group (by chance) gets a positive result, only that one positive study gets published. The scientific literature then shows that "Drug X works" — based on one published study — while the nine contradicting studies are invisible.
This is Wald's missing bombers applied to science. The evidence that survived (published studies) is not representative of all evidence (all studies conducted). The evidence that was filtered out (null results) may be more informative than the evidence that survived.
The Scale of the Problem
Publication bias is not a marginal phenomenon. Empirical studies of publication bias have found:
-
In medicine: Studies with positive results are roughly twice as likely to be published as studies with negative results. In pharmaceutical research, the ratio is even more skewed — industry-funded studies with positive results for the sponsor's drug are published at dramatically higher rates than studies with negative or mixed results.
-
In psychology: The Reproducibility Project (2015) attempted to replicate 100 published psychological studies. Only about 36% replicated with effects in the same direction and of similar magnitude. Publication bias — the selective publishing of surprising, positive results — is a major contributor to this replication failure.
-
In economics: Research suggests that published economics studies systematically overestimate effect sizes compared to pre-registered studies, consistent with publication bias selecting for larger, more dramatic effects.
-
In education: Systematic reviews have found that published educational intervention studies report larger effects than unpublished studies of the same interventions.
The pattern is consistent across fields: publication acts as a selection filter that systematically removes evidence against the hypothesis (null results) while preserving evidence for it (positive results). The surviving evidence — the published literature — is biased toward confirming hypotheses rather than testing them.
A Concrete Example: The Antidepressant Evidence
The most vivid documented case of publication bias in medicine involves antidepressant drugs. In 2008, researchers Ewan Turner, Annette Matthews, and colleagues published a landmark analysis comparing FDA registration data (which includes all submitted trials, both published and unpublished) with the published literature for twelve antidepressant drugs.
The results were stark:
- Of 74 FDA-registered studies, 38 had positive results and 36 had negative or questionable results — roughly a coin flip.
- Of the 38 positive studies, 37 were published.
- Of the 36 negative studies, only 3 were published as negative. Another 11 were published but with results spun to appear positive. The remaining 22 were never published at all.
The published literature showed 94% of antidepressant studies as positive. The full evidence — including the unpublished studies — showed approximately 51% as positive. An entirely different picture.
A physician reading only the published literature would conclude that antidepressants are dramatically effective for nearly everyone. A physician with access to the full evidence base would conclude that antidepressants are modestly effective on average, with significant variation across patients. The first conclusion drives prescribing practices. The second would lead to more careful, individualized treatment decisions.
This is not an abstract statistical curiosity. The survivorship-biased published evidence shaped prescribing guidelines, patient expectations, and insurance coverage decisions affecting millions of people. The evidence that wasn't published — the bombers that didn't return — contained information that would have materially changed clinical practice.
🔄 Check Your Understanding (try to answer without scrolling up)
- In the antidepressant example, what percentage of studies appeared positive in the published literature versus the full evidence base?
- Why is this an example of survivorship bias specifically (rather than some other type of bias)?
Verify
1. 94% appeared positive in the published literature; approximately 51% were positive in the full evidence base. 2. It is survivorship bias because the positive studies "survived" the publication selection process while the negative studies were filtered out. The resulting published evidence base is a biased sample of all conducted studies, systematically excluding the evidence that would weaken the conclusion.
The Funnel Plot: Detecting Publication Bias Visually
Statisticians have developed a simple visual tool for detecting publication bias: the funnel plot. When you plot the effect size of published studies against their sample size, a field without publication bias produces a symmetrical funnel shape — small studies scatter widely (because they're less precise), and large studies cluster tightly around the true effect.
When publication bias is present, the funnel is asymmetric: small studies with negative results are missing (they were filed away), while small studies with positive results — including those that are positive by chance — are over-represented. The funnel tilts.
{Diagram: Two funnel plots side by side. Left: "Symmetric funnel — no publication bias detected." Studies scatter symmetrically around a vertical line. Right: "Asymmetric funnel — publication bias likely." The lower-left corner (small negative studies) is empty, while the lower-right corner (small positive studies) is full.
Alt-text: Two scatter plots labeled "Funnel Plots." The left plot shows dots distributed symmetrically around a central line, forming a symmetric funnel shape wider at the bottom. The right plot shows the same general shape but with a gap in the lower-left quadrant and excess dots in the lower-right quadrant, indicating that small studies with negative results are missing while small studies with positive results are over-represented.}
Funnel plot analysis applied to many fields of medical and psychological research has consistently found asymmetry — the signature of publication bias operating at scale.
⚠️ Common Pitfall: Publication bias is not primarily caused by researcher fraud or misconduct. It is a structural feature of the publication system. Journals prefer to publish surprising, positive results because they attract readers and citations. Reviewers prefer to recommend studies with clean, significant findings. Researchers prefer to submit their most impressive results. Each actor is behaving rationally within the incentive system. The bias emerges from the structure, not from individual dishonesty.
🔗 Connection: Publication bias is the scientific instantiation of the streetlight effect (Chapter 4): the published literature illuminates what has been tested and found positive, while leaving the much larger space of null and negative results in darkness. It also interacts with the authority cascade (Chapter 2): published positive findings acquire citation momentum and authority, while the unpublished null findings cannot push back because they are invisible.
5.5 Startup Mythology and the Denominator Problem
The technology industry has its own version of survivorship bias, and it shapes billions of dollars in investment decisions.
The Narrative
The standard startup success narrative goes like this: a visionary founder, working out of a garage, defies conventional wisdom, pivots through adversity, and builds a billion-dollar company. Steve Jobs, Bill Gates, Mark Zuckerberg, Jeff Bezos — the stories are familiar, inspirational, and deeply misleading.
They are misleading because they omit the denominator. For every startup that becomes a household name, thousands of startups with equally passionate founders, equally innovative ideas, and equally heroic perseverance failed. The failure rate for venture-backed startups is roughly 75–90%, depending on the definition of failure. The survival rate for startups in general is even lower.
The Distortion
When the media, business schools, and motivational speakers study only the survivors, they extract "lessons" that are artifacts of selection, not causes of success:
-
"Follow your passion" — The survivors followed their passion. So did the failures. Passion is not a differentiating factor; it is a base-rate characteristic of anyone who starts a company.
-
"Disrupt the incumbents" — The survivors disrupted. Many failures also tried to disrupt and were destroyed. Disruption is a strategy that works spectacularly when it works and catastrophically when it doesn't. Studying only the successes tells you nothing about when it works.
-
"Move fast and break things" — The survivors moved fast. So did Theranos, WeWork, FTX, and thousands of other companies that moved fast and broke themselves. Speed is not inherently a virtue; it depends entirely on the context.
-
"The founder's vision was key" — Every founder has a vision. The difference between a "visionary founder" and a "delusional founder" is often just the outcome. We retroactively label the winners as visionaries and the losers as delusional, but the ex ante characteristics may be indistinguishable.
-
"Persistence is the key to success" — The survivors persisted and eventually succeeded. But so did the failures — they persisted and eventually went bankrupt. Persistence is a necessary condition for success (you can't succeed if you quit) but not a sufficient one. The survivorship bias makes it look sufficient because we never see the persistent failures.
The Control Group That Doesn't Exist
The fundamental problem with startup success narratives is the absence of a control group. You cannot identify the "key factors" that cause success by studying only successes, any more than you can identify what causes cancer by studying only people who have cancer (you need to compare them to people who don't).
What would a non-biased study of startup success look like? You would need to: 1. Identify a large cohort of startups at founding — before outcomes are known 2. Track all of them, successes and failures alike, over a long period 3. Measure their characteristics comprehensively at founding and over time 4. Compare the characteristics of those that succeeded with those that failed 5. Control for luck, timing, market conditions, and other confounding variables
This kind of study is expensive, time-consuming, and unglamorous. It would probably find that the most predictive factors are market timing, access to capital, and luck — not the heroic narratives of visionary founders. Which is precisely why it's rarely done: the results wouldn't sell books.
The Investment Consequence
Survivorship bias in startup mythology has real economic consequences. Venture capital investment decisions are influenced by pattern-matching against successful founders — but the patterns are drawn from survivors, not from the full population. This can lead to systematic overinvestment in founder profiles that match the survivor pattern (young, male, Stanford-educated, charismatic) and underinvestment in founder profiles that are equally capable but don't match the biased template.
🔍 Why Does This Work?
Survivorship bias in startups is particularly powerful because the stakes (potential massive returns) create strong motivation to find patterns, and the survivors' stories are compelling narratives that satisfy the human need for explanation (see Chapter 6: The Plausible Story Problem). The combination of high stakes, narrative satisfaction, and missing denominator data makes startup survivorship bias nearly irresistible.
5.6 What It Looked Like From Inside
Consider the perspective of a medical researcher in 2005:
- You have conducted a well-designed study testing whether a promising drug reduces blood pressure. After careful analysis, your result is: no significant effect. P = 0.34.
- You submit the study to a journal. The reviewers respond: "While the methodology is sound, the null result is not sufficiently novel or interesting for our readership. We suggest submission to a specialty journal."
- You submit to two more journals. Same response. Your null result is methodologically valid but unpublishable — not because it's wrong, but because it's not interesting enough.
- Meanwhile, a colleague at another university tests the same drug with a different sample and, through the randomness inherent in any study, gets a positive result (p = 0.04). Their study is published in a major journal. Their finding is cited. The drug appears to work.
- You know that the published evidence is misleading — because your unpublished study contradicts it. But you have no venue for your evidence. It sits in your file drawer. You move on to the next project.
From inside this system, no individual is behaving badly. You conducted good research. The journals applied their standard criteria. Your colleague's study was legitimate. The problem is that the system — the entire publication infrastructure — systematically selects for positive results, creating a biased evidence base that overestimates the efficacy of treatments and the size of effects.
This is why publication bias is so pernicious: it operates without anyone making a conscious decision to bias the evidence. It is an emergent property of the incentive structure, not a deliberate choice.
And the researcher who files away the null result is not simply being passive. They are making a rational career decision. Publishing a null result requires significant effort (writing up, submitting, revising, resubmitting) with minimal career reward (null results rarely attract citations, don't make headlines, and don't impress tenure committees). The researcher's time is better spent on the next study that might produce a publishable positive result. The structural incentives ensure that the file drawer fills steadily with evidence that would correct the field's biases — evidence that no one will ever see.
This pattern has a particularly perverse consequence for drug development. Pharmaceutical companies conduct multiple clinical trials of each drug. If three of five trials show positive results and two show negative results, the company may publish the three positive trials and file the two negative ones. A physician reviewing the published literature sees three studies supporting the drug and none against it. The true evidence — a roughly 60/40 split — is invisible. The physician prescribes the drug based on survivorship-biased evidence, and the patient may receive a treatment that is less effective than the literature suggests.
This is not speculation. The antidepressant example documented exactly this pattern. And it extends far beyond antidepressants: reviews of FDA-registered trials across multiple drug categories have found similar patterns of selective publication.
📐 Project Checkpoint
Your Epistemic Audit — Chapter 5 Addition
Return to your audit target and ask:
What evidence in your field DIDN'T survive? What studies weren't published? What failures aren't discussed? What perspectives are missing from the record?
Is there a publication bias problem? For the core claims in your field, are they supported primarily by published studies? Are there likely unpublished null results that would change the picture?
What's the denominator? When your field cites examples of success (successful treatments, successful companies, successful policies), what's the base rate? How many failures exist for every success?
What selection processes filter the evidence before it reaches you? Map the filters: publication, media coverage, institutional memory, training curricula, conference presentations. Which of these systematically remove certain types of evidence?
Wald's question: If you could see the evidence that was filtered out — the unpublished studies, the failed examples, the lost perspectives — would your field's core conclusions change?
Add 300–500 words to your Epistemic Audit document.
5.7 Active Right Now: Where Survivorship Bias May Be Operating
AI model benchmarks. When AI companies report model performance, they typically report the best results across many training runs and evaluation configurations. The unsuccessful runs — models that failed to train, configurations that underperformed — are not reported. The published benchmarks represent the survivors of an optimization process, and their performance may substantially overstate what a typical deployment would achieve.
COVID-19 treatment evidence. During the pandemic, the urgency to find treatments led to thousands of studies, many of which were small, poorly controlled, and subject to severe publication bias. Positive results were published rapidly; negative results often weren't. The result: several treatments (hydroxychloroquine, ivermectin) appeared effective in early published studies but failed in larger, more rigorous trials. The early evidence base was survivorship-biased — the positive studies survived to publication while the negative ones were in file drawers.
Influencer and creator success narratives. Social media platforms showcase their most successful creators — the ones who gained millions of followers, signed brand deals, or built businesses. The vastly larger population of creators who invested similar effort and achieved nothing is invisible. The result: aspiring creators systematically overestimate their chances of success based on a survivorship-biased sample of the visible winners.
Historical narratives of social movements. We study the social movements that succeeded — civil rights, women's suffrage, anti-apartheid. We rarely study the movements that had similar characteristics, similar passion, and similar organization but failed. This creates a biased understanding of what makes movements succeed, because our sample contains only the survivors.
5.8 Mutual Funds and the Advertising of Survivors
One of the purest real-world demonstrations of survivorship bias occurs in the financial industry.
Mutual fund companies advertise the performance of their best-performing funds. This is legal, accurate, and deeply misleading.
Here's why: mutual fund companies regularly close or merge their worst-performing funds. A fund that performs poorly for several years is quietly merged into a better-performing fund, and its track record disappears from the company's published statistics. The surviving funds — the ones that are still open and advertised — represent a biased sample of all funds the company has ever managed.
Studies have estimated that survivorship bias inflates the average reported return of mutual funds by roughly 1–2% per year — a significant amount in an industry where a 1% difference in annual returns can mean the difference between comfortable retirement and financial hardship over 30 years.
The investor who reads the fund company's marketing materials is seeing only the survivors. The failures have been quietly removed from the record, like the bombers that didn't come back from their missions.
This is particularly consequential because investors use past performance to make future decisions — despite the standard disclaimer that "past performance is not indicative of future results." When past performance data is survivorship-biased, the disclaimer is understated: the past performance data is not just unindicative — it is systematically misleading, showing a rosier picture than any investor should expect.
The Securities and Exchange Commission has required certain disclosures to mitigate this bias, but the fundamental dynamic persists: the funds that investors see advertised are the ones that survived, and the ones that survived are by definition the ones that performed well enough to remain open. The selection filter ensures that investors are always comparing their options against a biased baseline.
📊 Real-World Application: A 2023 analysis found that over a 20-year period, roughly half of all mutual funds were either closed or merged. The funds that survived — and whose track records were displayed to prospective investors — were systematically better than the full population of funds. An investor choosing based on published track records was making a decision based on survivorship-biased data, unknowingly comparing their options to a curated sample rather than the full field.
5.9 The Antidote: Learning to See What Isn't There
Detecting survivorship bias requires a specific cognitive habit: looking for the missing evidence. This is difficult because, by definition, the missing evidence isn't visible. But several strategies can help.
Strategy 1: Always Ask About the Denominator
When presented with examples of success, ask: "How many failures used the same approach?" When told that a treatment works in published studies, ask: "How many unpublished studies tested the same treatment?" When shown that successful companies share a characteristic, ask: "Do failed companies also share it?"
Strategy 2: Seek Pre-Registered Evidence
In science, pre-registration (declaring your hypothesis and analysis plan before collecting data) and registered reports (journals committing to publish the study regardless of results) are powerful antidotes to publication bias. When evaluating evidence, give more weight to pre-registered studies than to post-hoc analyses.
Registered reports are particularly powerful because they eliminate the publication selection filter entirely: the journal commits to publishing the study based on the methodology (before results are known), not on the results. This means that null results get published at the same rate as positive results, and the file drawer empties. Early evidence suggests that registered reports produce substantially different findings than traditional publications — with more null results and smaller effect sizes, exactly as the publication bias model predicts.
Strategy 3: Look for Systematic Reviews and Meta-Analyses
Systematic reviews that include unpublished studies, conference abstracts, and grey literature provide a less biased picture than any single published study. The Cochrane Collaboration in medicine and the Campbell Collaboration in social science specialize in minimizing survivorship bias in evidence synthesis. When a Cochrane review concludes that a treatment works, the conclusion is based on a deliberately comprehensive search for all evidence — including the file drawer. This makes Cochrane reviews substantially more reliable than individual published studies, though even they cannot fully overcome publication bias when the unpublished studies are truly inaccessible.
Strategy 4: Study Failures, Not Just Successes
The most informative evidence often comes from failures. Aviation safety improved dramatically not by studying successful flights (survivorship-biased) but by studying crashes and near-misses (the evidence that didn't "survive"). The Aviation Safety Reporting System (ASRS), which collects voluntary reports of near-misses without punishing the reporters, is one of the most successful de-biasing mechanisms ever created. It works by making the non-surviving evidence visible.
Medical knowledge advances through the study of disease, not just health. Engineering advances through the study of structural failures, not just successful buildings. If your field studies only successes, propose studying failures — the evidence will almost certainly be more informative.
🌍 Global Perspective: Different cultures handle survivorship bias in failure analysis differently. Japan's automotive industry pioneered the "5 Whys" methodology — repeatedly asking "why?" after every failure to find the root cause rather than the proximate blame. The U.S. military's After Action Review (AAR) process is designed to capture lessons from both successes and failures. The financial industry's "post-mortem" culture varies enormously: some firms rigorously study their failed investments while others quietly bury them. The common thread: organizations that systematically study their failures learn faster than organizations that study only their successes.
Strategy 5: Build a Mental Model of the Selection Process
For any body of evidence, ask: "What process selected this evidence for my attention?" Understanding the selection filter allows you to estimate what was filtered out. If you know that journals preferentially publish positive results, you can mentally adjust for the bias. The adjustment won't be precise, but even an approximate correction is better than none.
✅ Best Practice: When encountering any impressive-sounding statistic or case study, train yourself to ask: "What is this a survivor of?" This single question — applied consistently — is one of the most powerful epistemic tools in this book.
5.10 Practical Considerations: Living With Biased Evidence
We cannot eliminate survivorship bias. The historical record will always be incomplete. The published literature will always overrepresent positive findings. Business case studies will always oversample winners. The question is how to make decisions in a world where the available evidence is systematically biased.
The answer is calibrated confidence: holding conclusions with less certainty than the surviving evidence seems to warrant, because you know the evidence is biased. If published studies unanimously support a treatment, and you know that publication bias exists, the correct degree of confidence is lower than the published evidence alone would suggest — perhaps substantially lower. This isn't cynicism; it's calibration.
The analogy is insurance. You don't know that your house will burn down. But you know that houses sometimes burn down, and you adjust your behavior accordingly. You don't know which specific findings in your field are distorted by survivorship bias. But you know that some of them are, and you should adjust your confidence accordingly.
This is not an argument for paralysis or nihilism. You still need to make decisions based on the best available evidence. But "best available evidence" should be understood as "evidence that has passed through a known biased filter" — and your confidence should be calibrated to that understanding. A finding supported by pre-registered studies with null results included is more trustworthy than the same finding supported only by post-hoc published studies. A business strategy supported by analysis of both successful and failed implementations is more trustworthy than one based only on case studies of winners.
🪞 Learning Check-In
Pause and reflect: - Think of a belief you hold confidently. What evidence supports it? Now ask: is that evidence survivorship-biased? What evidence might have been filtered out? - In your field, what is the "file drawer"? What findings never get published, shared, or discussed? - How would your field's conclusions change if the non-surviving evidence suddenly became visible?
5.11 Chapter Summary
Key Arguments
- Survivorship bias is the systematic error of drawing conclusions from evidence that survived a selection process while ignoring what was filtered out
- Abraham Wald's WWII bomber insight captures the core logic: the evidence you don't see is often more important than the evidence you do
- The bias operates at massive scale in science (publication bias), business (success literature), history (literacy and durability bias), finance (fund performance reporting), and technology (startup mythology)
- Publication bias is the most consequential form because it distorts the scientific evidence base that all other fields rely on
- The antidote is not eliminating the bias (impossible) but learning to detect it and calibrate confidence accordingly
Key Debates
- Can pre-registration and registered reports solve publication bias, or will they create new problems?
- Is the business success literature redeemable, or is the methodology fundamentally flawed?
- How much should we discount published evidence to account for the file drawer?
Analytical Framework
- The "denominator question": How many failures for every success?
- The "selection filter" mapping: What process selected this evidence?
- The "Wald question": What would the missing evidence tell me?
- Calibrated confidence as the response to biased evidence
Spaced Review
Revisiting earlier material to strengthen retention.
- (From Chapter 3) How does survivorship bias interact with unfalsifiability? Can survivorship bias make a falsifiable theory appear unfalsifiable by selectively removing disconfirming evidence?
- (From Chapter 2) The authority cascade amplifies claims through citation. Publication bias amplifies claims through selective publishing. How are these mechanisms similar? How are they different?
- (From Chapter 1) At which stage(s) of the lifecycle of a wrong idea does survivorship bias typically operate?
Answers
1. Yes — if null results are systematically unpublished, the surviving evidence base may contain only confirmations, making the theory *appear* to have no counter-evidence. The theory isn't unfalsifiable in principle (Level 3–4), but publication bias has made it unfalsifiable *in practice* by removing the disconfirming data. 2. Both amplify positive signals while suppressing negative ones. Authority cascade amplifies through *social* mechanisms (prestige, deference); publication bias amplifies through *institutional* mechanisms (journal selection, reviewer preference). Both create the illusion of stronger consensus than the full evidence warrants. 3. Primarily at Stage 2 (Adoption) and Stage 3 (Entrenchment): survivorship bias shapes which evidence enters the field's awareness, causing wrong ideas to appear better-supported than they are. It also operates at Stage 5 (Resistance), because the published evidence base — biased toward the dominant view — can be cited to dismiss challengers.What's Next
In Chapter 6: The Plausible Story Problem, we'll examine the fifth entry mechanism: how the human need for narrative coherence causes us to accept explanations that are compelling but unsupported. You'll encounter evolutionary psychology "just-so stories," startup mythology (now seen through the lens of narrative rather than survivorship), and the troubling question of whether explanation and storytelling are the same thing.
Before moving on, complete the exercises and quiz to solidify your understanding.
Chapter 5 Exercises → exercises.md
Chapter 5 Quiz → quiz.md
Case Study: Abraham Wald and the Art of Seeing What Isn't There → case-study-01.md
Case Study: Publication Bias in Antidepressant Research → case-study-02.md
Related Reading
Explore this topic in other books
How Humans Get Stuck The Anchoring of First Explanations Applied Psychology Cognitive Biases Media Literacy Cognitive Biases and Heuristics Propaganda Cognitive Biases Pattern Recognition Survivorship Bias Science of Luck Survivorship Bias Introductory Statistics Designing Studies