During World War II, the American military had a problem. Bombers were being shot down over Europe at devastating rates, and the surviving aircraft were returning to base riddled with bullet holes. The military wanted to add armor to the planes, but...
Learning Objectives
- Define survivorship bias and recount Abraham Wald's WWII bomber analysis, identifying the structural insight that made his reasoning revolutionary
- Analyze how survivorship bias distorts business success literature, explaining why studying only successful companies tells you nothing reliable about the causes of success
- Explain how survivorship bias operates in music, architecture, and cultural memory, producing the illusion that the past was better than the present
- Evaluate the healthy survivor effect in medicine and clinical trials, recognizing how differential attrition biases the evidence base for treatment effectiveness
- Identify survivorship bias in military history, fund manager performance, and the scientific publication record, describing the specific selection mechanism in each case
- Apply the threshold concept -- The Evidence Destroys Itself -- to recognize that in many domains the process of success or survival systematically eliminates the evidence of failure, making it structurally impossible to learn from failure unless you deliberately seek out the dead
In This Chapter
- Business, Music, Medicine, Military History, Architecture, Finance, Science
- 37.1 The Planes That Didn't Come Back
- 37.2 Business Success Literature -- The Graveyard No One Visits
- 37.3 Music -- The Illusion That the Past Was Better
- 37.4 Architecture -- Only the Good Buildings Survived
- 37.5 Medicine -- The Healthy Survivor Effect
- 37.6 Military History -- Written by the Survivors
- 37.7 Fund Manager Performance -- The Disappearing Losers
- 37.8 Silent Evidence -- Taleb's Framework
- 37.9 Publication Bias -- The File Drawer Problem
- 37.10 Countermeasures -- Seeking the Dead
- 37.11 The Threshold Concept -- The Evidence Destroys Itself
- 37.12 The Pattern Library Checkpoint
- 37.13 The Silent Graveyard Surrounds Every Field
- Summary
Chapter 37: Survivorship Bias -- The Evidence You Never See
Business, Music, Medicine, Military History, Architecture, Finance, Science
"The cemetery of failed restaurants is very quiet." -- Nassim Nicholas Taleb, The Black Swan, 2007
37.1 The Planes That Didn't Come Back
During World War II, the American military had a problem. Bombers were being shot down over Europe at devastating rates, and the surviving aircraft were returning to base riddled with bullet holes. The military wanted to add armor to the planes, but armor is heavy. Adding it everywhere would make the bombers too slow and too fuel-hungry to complete their missions. The question was: where should the armor go?
The obvious answer seemed clear. Engineers at the Center for Naval Analyses examined the returning bombers and carefully catalogued where the bullet holes were concentrated. The fuselage was riddled. The wings had clusters of damage. The tail gunner positions were pockmarked. The engines, by contrast, showed relatively few hits. The conclusion seemed inescapable: armor the fuselage, the wings, and the tail, because those are the areas taking the most damage.
Abraham Wald, a mathematician from what is now Cluj-Napoca, Romania, who had fled the Nazis and was working with the Statistical Research Group at Columbia University, saw the problem differently. Wald looked at the same data -- the same distribution of bullet holes on the same returning bombers -- and reached the opposite conclusion.
The bullet holes on the returning planes, Wald realized, did not represent the places where bombers were vulnerable. They represented the places where bombers could be hit and still fly home. The planes that had been hit in the engines were not in the sample. They were at the bottom of the English Channel. The data the military was examining was not a random sample of all bullet damage. It was a biased sample -- biased by the very process of survival. The planes that made it back were, by definition, the planes that could survive the damage they had received. The planes that could not survive their damage were absent from the data, invisible, destroyed by the same process that generated the evidence.
Wald's recommendation: armor the engines. Armor the places where the returning planes were not hit, because those were the areas where a hit was fatal.
This story -- verified and well-documented in the historical record of the Statistical Research Group -- is the clearest illustration of survivorship bias in the intellectual canon. Survivorship bias occurs when we draw conclusions from the things that survived a selection process while ignoring the things that did not survive, leading to systematically wrong conclusions about what caused the survival.
The structural insight is this: the evidence you are looking at has already been filtered by the process you are trying to understand. The survivors are visible. The non-survivors are not. And if you mistake the visible survivors for a representative sample, every conclusion you draw will be distorted in the same direction -- toward overestimating the robustness, the skill, the quality, or the safety of whatever survived.
Wald understood this because he was a mathematician trained in sampling theory. But the insight extends far beyond bullet holes on bombers. It extends into every field that draws conclusions from what exists, what succeeded, what was recorded, or what survived -- which is to say, every field.
Fast Track: Survivorship bias is the systematic error of drawing conclusions from what survived a selection process while ignoring what did not survive. If you already grasp this core idea from the Wald example, skip to Section 37.5 (Medicine) for the healthy survivor effect, then read Section 37.8 (Silent Evidence) for Taleb's formal framework, Section 37.9 (Publication Bias) for the scientific implications, and Section 37.10 (Countermeasures) for practical remedies. The threshold concept is The Evidence Destroys Itself: in many domains, the process of survival systematically eliminates the evidence of failure, making it structurally impossible to learn from failure unless you deliberately seek out the dead.
Deep Dive: The full chapter develops survivorship bias across seven domains in concrete detail, extracts the shared deep structure through Taleb's silent evidence concept, connects it to the streetlight effect (Ch. 35), base rate neglect (Ch. 10), and signal and noise (Ch. 6), and examines countermeasures including pre-registration, the outside view, and base rate thinking. Read everything, including both case studies. Section 37.8 on silent evidence is where the chapter's deepest theoretical synthesis occurs.
37.2 Business Success Literature -- The Graveyard No One Visits
In 2001, Jim Collins published Good to Great, a book that would sell over four million copies and become one of the most influential business books ever written. Collins and his research team studied eleven companies that had made the transition from good performance to sustained great performance over a fifteen-year period. They identified the characteristics these companies shared: Level 5 leadership (humble but determined), the hedgehog concept (focusing on what you can be best at), a culture of discipline, getting the right people on the bus.
The methodology appeared rigorous. Collins used comparison companies -- firms in the same industries that did not make the good-to-great transition -- to isolate the distinguishing factors. The conclusions were presented with the confidence of scientific findings. CEOs, MBA students, and management consultants absorbed the lessons. The companies Collins profiled became exemplars: Fannie Mae, Circuit City, Wells Fargo, Gillette.
Within a decade, the exemplar list had crumbled. Fannie Mae was placed in government conservatorship during the 2008 financial crisis, its balance sheet toxic with subprime mortgage exposure. Circuit City went bankrupt in 2009. Wells Fargo became the center of one of the largest consumer fraud scandals in banking history, with employees creating millions of unauthorized accounts. The "great" companies were not merely declining. They were failing spectacularly, in ways that seemed to contradict every principle Collins had identified.
What went wrong? The answer is not that Collins was careless. The answer is that his methodology was built on survivorship bias.
Collins studied companies that had already succeeded. He then looked backward to identify what they had in common. But the only companies available for study were the ones that had survived long enough and performed well enough to appear in the dataset. The companies that had tried the same strategies and failed -- the companies with humble leaders who focused on a single concept and built cultures of discipline and still went bankrupt -- were invisible. They were in the graveyard. No one was studying them, because there was nothing to study: they had been dissolved, acquired, or forgotten.
This is the fundamental problem with business success literature. Tom Peters and Robert Waterman's In Search of Excellence (1982) profiled forty-three excellent companies; within two years of publication, a third of them were in financial difficulty. The traits that Peters and Waterman identified as distinguishing excellent companies -- a bias for action, closeness to the customer, autonomy and entrepreneurship -- were almost certainly present in an equal number of failed companies. But failed companies do not get profiled. They do not get studied. They do not write memoirs. They vanish.
The same bias infects startup culture. Entrepreneurship literature is dominated by the stories of founders who succeeded: Steve Jobs, Elon Musk, Sara Blakely, Jeff Bezos. Their characteristics are catalogued and prescribed. Take risks. Follow your passion. Drop out of college. Disrupt the industry. Ignore the doubters. The advice is drawn entirely from the survivors. The founders who took the same risks, followed the same passions, dropped out of the same colleges, disrupted the same industries, ignored the same doubters, and went bankrupt are not writing books. They are not giving TED talks. They are not available for study.
The result is that business advice systematically overstates the reliability of successful strategies, understates the role of luck and timing, and makes success look far more reproducible than it actually is. Every successful company is a bullet hole on a returning bomber. Every failed company is a plane at the bottom of the Channel. And the advice -- "do what the survivors did" -- is the equivalent of armoring the fuselage.
Connection to Chapter 10 (Base Rates): Survivorship bias in business literature is a specific failure of base rate reasoning. When someone points to a successful founder who dropped out of college and says "see, dropping out can work," they are ignoring the base rate: the proportion of all college dropouts who attempted entrepreneurship and failed. The success stories are visible. The failure stories are invisible. The base rate -- the denominator in the probability calculation -- is hidden by the selection process. Without the base rate, the observed success rate is meaningless.
🔄 Check Your Understanding
- In Abraham Wald's bomber analysis, why did the military's initial interpretation of the bullet hole data lead to exactly the wrong conclusion? What structural feature of the data created the error?
- Why does studying only successful companies tell you nothing reliable about the causes of success? What would you need to study instead, and why is that study difficult to conduct?
- How does survivorship bias in business literature interact with base rate neglect (Ch. 10) to make success strategies appear more reliable than they are?
37.3 Music -- The Illusion That the Past Was Better
"They don't make music like they used to."
This sentence, or something very close to it, has been uttered in every generation for as long as recorded music has existed. People in the 1990s said it about the 1970s. People in the 1970s said it about the 1950s. People in the 1950s said it about the 1930s. The conviction is sincere, widely shared, and almost certainly wrong -- not because old music was bad, but because survivorship bias has silently curated the past, removing everything mediocre and leaving only the best.
Consider medieval music. When people think of medieval music at all, they think of Gregorian chant, of troubadour songs, of the compositions of Hildegard von Bingen or Guillaume de Machaut. This music is beautiful, haunting, and sophisticated. It gives the impression that medieval people had extraordinary musical taste.
But the medieval music we know is an infinitesimal fraction of the medieval music that existed. Musicologists estimate that well over ninety percent of medieval compositions have been lost entirely -- not because they were destroyed by any deliberate process, but because the mechanisms of preservation were so fragile. Music had to be written down (in a period when musical notation was still evolving and literacy was rare), the manuscript had to survive fire, flood, war, insect damage, and simple neglect across centuries, and the manuscript had to end up in an institution -- a monastery, a cathedral, a university library -- that would preserve it. The music that survived this gauntlet was not a random sample. It was biased toward compositions that were considered important enough to copy, associated with institutions powerful enough to maintain archives, and written in notation systems that later generations could decode.
The vast majority of medieval music -- the tavern songs, the work songs, the popular ditties, the forgettable melodies, the experimental compositions that did not catch on -- is gone. Not lost in the sense that it might be found. Gone in the sense that the manuscripts were never made, or were made and rotted, or were made and burned, or were made and thrown away because no one thought they were worth keeping. The evidence was destroyed by the very process of cultural selection that determines what gets remembered.
When we compare "medieval music" (meaning the surviving masterpieces) to "modern music" (meaning the full, unfiltered range of everything being produced right now, including the terrible), the comparison is rigged. We are comparing the curated highlights of seven centuries against the unfiltered present. The past always wins this comparison, because the past has had its mediocrity removed by time.
The same bias operates in every cultural domain where time filters the record. We think classical literature was uniformly brilliant because we read only the works that survived out of the tens of thousands that were published. We think golden-age Hollywood produced only masterpieces because we watch only the films that are still distributed. We think the 1960s produced only great rock music because we hear only the albums that have remained in circulation. In each case, the silent graveyard -- the vast cemetery of forgotten mediocrity -- is invisible, and its invisibility makes the survivors look like the norm rather than the exception.
Retrieval Prompt: Pause before continuing. Can you explain in your own words why comparing old music to new music is an unfair comparison? What is the selection mechanism that makes the past look better? And can you identify another domain -- literature, film, fashion, technology -- where the same mechanism operates?
37.4 Architecture -- Only the Good Buildings Survived
The Parthenon has stood for nearly 2,500 years. The Roman Colosseum has endured for nearly two millennia. The great Gothic cathedrals of Europe have survived six, seven, eight centuries of war, weather, and neglect. Ancient buildings, it seems, were built to last. Modern buildings, by contrast, seem flimsy, temporary, disposable. The comparison appears to prove that ancient builders were more skilled, more conscientious, more committed to permanence than their modern counterparts.
The comparison proves nothing of the sort.
Ancient Greece did not produce only the Parthenon. It produced thousands of buildings -- houses, workshops, warehouses, markets, temples, walls -- the vast majority of which were built with timber, mudbrick, thatch, or low-quality stone. These buildings collapsed, rotted, burned, or were demolished within decades or centuries of their construction. They left little or no archaeological trace. The Parthenon survived because it was built with exceptional materials (Pentelic marble), exceptional engineering (its columns have a deliberate slight curvature to correct for optical illusion), and exceptional cultural significance that motivated successive civilizations to maintain or at least not demolish it. It is not representative of ancient Greek construction. It is the extreme outlier -- the one building in a thousand that was built well enough, maintained consistently enough, and valued sufficiently to survive.
The same logic applies to Roman architecture. The Romans built prolifically, and the vast majority of what they built is gone. The concrete apartment blocks (insulae) that housed most of Rome's population were notorious for their shoddy construction; the ancient writer Juvenal complained bitterly about collapsing buildings, and the architect Vitruvius noted that many structures failed within years of completion. The buildings that survived -- the Colosseum, the Pantheon, the aqueducts -- were the finest products of the empire's most skilled engineers, built with the best materials and maintained by imperial authority. They represent the ceiling of Roman capability, not the floor.
Gothic cathedrals provide perhaps the starkest example. The cathedrals of Chartres, Notre-Dame, Cologne, and Salisbury are awe-inspiring achievements. But the history of Gothic construction is also a history of catastrophic structural failures. The cathedral at Beauvais, which attempted to push Gothic engineering to its limits, suffered a partial collapse of its choir vault in 1284, just twelve years after completion. Numerous lesser churches and cathedrals collapsed entirely and were either rebuilt or abandoned. The engineering treatises of the period describe building failures as routine events. The cathedrals we admire today are not typical of medieval construction. They are the survivors of a building program in which failure was common and only the structurally sound endured.
The survivorship bias in architecture creates a false narrative about the trajectory of building quality. Ancient buildings seem better because the bad ones fell down. Modern buildings seem worse because all of them -- the excellent and the terrible -- are still standing. Give it five hundred years, and the only buildings from our era that will remain will be the best-engineered ones. Future observers will look at those survivors and declare, with equal conviction, that early twenty-first-century construction was of extraordinary quality.
Spaced Review (Ch. 33): Recall the lifecycle S-curve from Chapter 33 -- the universal pattern of growth, saturation, and eventual plateau or decline. Buildings follow their own S-curves: rapid construction, a period of use and maintenance, and eventual decay or demolition. Survivorship bias in architecture is the result of observing only the buildings that are still in the growth or plateau phase of their S-curve. The buildings that completed their S-curves -- that reached decline and collapsed -- are invisible. We see only the right tail of the distribution: the buildings that, for whatever reason, have not yet finished dying.
37.5 Medicine -- The Healthy Survivor Effect
In the early 2000s, a series of observational studies appeared to show that hormone replacement therapy (HRT) in postmenopausal women reduced the risk of heart disease. The evidence seemed compelling: women who took HRT had lower rates of cardiac events than women who did not. Doctors prescribed HRT widely, and millions of women took it with the expectation that it was protecting their hearts.
Then the Women's Health Initiative, a large randomized controlled trial, reported its results. HRT did not reduce heart disease risk. In fact, it slightly increased the risk of heart attack and stroke. The observational studies had been wrong -- not because the data was fabricated, but because survivorship bias had contaminated the comparison.
The mechanism was the healthy survivor effect (sometimes called the healthy user effect). Women who chose to take HRT were not a random sample of all postmenopausal women. They were, on average, healthier, wealthier, more health-conscious, and more engaged with the medical system than women who did not take HRT. They exercised more. They ate better. They had better access to medical care. They were, in short, the kind of women who were already at lower risk for heart disease -- and they were the kind of women who would be more likely to survive long enough to be included in the study.
The comparison -- HRT users versus non-users -- was contaminated by selection. The apparent benefit of HRT was not caused by the hormone therapy. It was caused by the fact that the women who took it were already healthier. The health preceded the treatment. But the study, observing only the survivors and the currently living, could not distinguish between the effect of the drug and the effect of the selection process that determined who took the drug.
The healthy survivor effect is pervasive in medical research. Any observational study that compares people who chose a treatment to people who did not choose it is vulnerable, because the choice itself is correlated with health, socioeconomic status, health literacy, and access to care. People who take vitamins are healthier than people who do not -- but the vitamins may not be the cause. People who follow their doctor's instructions have better outcomes than people who do not -- but the compliance may be a marker of general health consciousness rather than a cause of better outcomes. In each case, the "survivors" -- the people who are healthy enough, engaged enough, and privileged enough to choose the treatment -- are mistaken for evidence that the treatment works.
Clinical trials are designed to overcome survivorship bias through randomization: assigning treatments randomly so that the treatment and control groups are comparable at baseline. But even clinical trials are not immune. Differential attrition -- the tendency for sicker patients to drop out of trials before completion -- creates survivorship bias within the trial itself. If patients who experience side effects or worsening symptoms are more likely to discontinue the study, the remaining participants are a biased sample: they are the patients for whom the treatment was tolerable. The trial's results then overstate the treatment's effectiveness and understate its side effects, because the patients who had the worst experiences are no longer in the data.
Connection to Chapter 35 (Streetlight Effect): The healthy survivor effect is survivorship bias operating through the same structural logic as the streetlight effect: the evidence that is available (data from survivors, from compliant patients, from people who stayed in the trial) is systematically different from the evidence that is missing (data from those who died, dropped out, or were too sick to participate). Chapter 35 described the tendency to search where the data is; Chapter 37 describes why the data itself is already biased by a selection process that removed the most informative observations.
🔄 Check Your Understanding
- Explain the healthy survivor effect in your own words. Why do observational studies of health interventions systematically overestimate treatment benefits?
- How does differential attrition create survivorship bias within randomized clinical trials -- the very studies designed to eliminate such bias?
- What structural feature do the hormone replacement therapy example and the returning bombers example share? What is the common pattern?
37.6 Military History -- Written by the Survivors
"History is written by the victors" is one of those quotations that is attributed to Winston Churchill, though he almost certainly never said it. The attribution is itself a minor form of survivorship bias -- Churchill's words survived more robustly than those of less famous figures, so everything pithy tends to get attributed to him.
But the insight, whoever originated it, contains a deep truth about survivorship bias in historical knowledge. Military history is not merely incomplete. It is systematically incomplete in a particular direction: the strategies that won are overrepresented, the strategies that lost are underrepresented, and the strategies that led to total annihilation are absent entirely.
Consider the Mongol conquests of the thirteenth century. The Mongol military machine under Genghis Khan and his successors conquered the largest contiguous land empire in history, spanning from Korea to Hungary. The Mongol approach to warfare -- rapid cavalry movement, psychological warfare, sophisticated intelligence networks, feigned retreats, and the systematic destruction of cities that resisted -- is extensively documented because the Mongols won. Their strategies are studied in military academies. Their tactical innovations are celebrated.
But the peoples the Mongols conquered also had military strategies. The Khwarezmian Empire, which the Mongols destroyed in 1219-1221, had a defensive strategy of dispersing its forces across fortified cities. This strategy failed catastrophically against Mongol mobility. The civilizations of Central Asia had their own approaches to warfare that proved inadequate against this particular enemy. These strategies are documented only through the lens of their failure -- and many are not documented at all, because the civilizations that employed them were destroyed so thoroughly that their military treatises, their strategic debates, their tactical manuals were lost along with their cities.
The bias extends to every era. We know a great deal about Roman military strategy because Rome won, survived, and produced writers who documented its methods. We know relatively little about the military strategies of the peoples Rome defeated and absorbed -- the Samnites, the Gauls, the Carthaginians (with the partial exception of Hannibal, whose fame is a function of having nearly defeated Rome). The Carthaginian perspective on the Punic Wars is almost entirely lost. We know Carthaginian strategy only through Roman sources, which describe it from the perspective of the enemy.
The consequence is that military history creates a systematic illusion: the strategies that produced victory appear to be better strategies, because we have extensive evidence of their success. But "extensive evidence of success" is not the same as "evidence that the strategy was reliably superior." It may be that the winning strategy succeeded because of circumstances, timing, terrain, logistics, luck, or the specific weaknesses of the particular enemy it faced. Alternative strategies that might have worked better in different circumstances were tested by armies that were destroyed, and the evidence was destroyed with them.
This is survivorship bias operating at the civilizational level: the selection process of military conquest does not merely eliminate armies. It eliminates the entire knowledge base of the defeated civilization -- its records, its analyses, its self-understanding. The evidence does not merely fail to survive. It is actively destroyed by the very process that determines what survives.
Retrieval Prompt: Pause before continuing. You have now seen survivorship bias in four domains: military engineering (Wald), business, music, and military history. Before reading the next section, can you articulate the common structural pattern? In each case, what is the selection process? What is being filtered out? And why does the filtering make the survivors look better than they are?
37.7 Fund Manager Performance -- The Disappearing Losers
Every year, financial publications report on the performance of mutual funds and hedge funds. The reports invariably identify funds that have outperformed the market for three, five, or ten consecutive years. Financial advisors point to these track records as evidence of manager skill. Investors allocate their money accordingly. The implied message is clear: some fund managers are genuinely better than others, and past performance can identify them.
The data is real. The interpretation is contaminated by survivorship bias.
Here is the mechanism. In any given year, roughly half of actively managed funds underperform their benchmark index (the exact proportion varies, but this is the approximate long-run average). Funds that underperform significantly for several consecutive years face consequences: investors withdraw their money, management companies merge the fund into a better-performing fund, or the fund is closed entirely. When a fund is closed or merged, its track record often vanishes from the databases that researchers and financial journalists use to evaluate performance. The fund is not reported as having failed. It simply ceases to exist in the dataset.
The result is a database that systematically overstates the average performance of active fund managers. The funds that are still alive -- the funds whose track records can be examined -- are a biased sample. They are the survivors. The funds that performed worst have been removed from the sample by the very process of failure. It is as if you surveyed only the people who made it to age ninety and concluded that the human lifespan was ninety years.
The magnitude of the bias is significant. Studies of mutual fund survivorship bias have found that it inflates reported average returns by roughly one to two percentage points per year -- a seemingly small number that compounds dramatically over long periods. A ten-year track record that appears to show two percent annual outperformance may, after correcting for survivorship bias, show no outperformance at all.
The bias also distorts the apparent persistence of manager skill. When researchers examine whether fund managers who outperformed in the past continue to outperform in the future, the answer depends heavily on whether defunct funds are included in the analysis. With survivorship-bias-free data, the evidence for persistent skill is weak. With survivorship-biased data -- the data most commonly available to individual investors -- the evidence for persistent skill appears much stronger than it is.
The financial industry has strong incentives to maintain this bias. Fund companies are not required to publicize the performance records of funds they have closed. Marketing materials naturally feature the best-performing funds. Financial journalists, who need compelling stories about exceptional performance, gravitate toward the funds with the longest and most impressive track records -- which are, by definition, the survivors. The entire ecosystem of financial information is structured to make survivorship bias invisible.
Spaced Review (Ch. 35): Recall the streetlight effect from Chapter 35 -- the tendency to draw conclusions from the evidence that is available rather than the evidence that should be available. Fund manager survivorship bias is the streetlight effect applied to financial performance data. The light shines on the surviving funds. The dark contains the dead funds. Investors and journalists search under the light, find impressive track records, and conclude that active management works -- exactly as they would if they studied only the returning bombers and concluded that the fuselage was the vulnerable area.
37.8 Silent Evidence -- Taleb's Framework
Nassim Nicholas Taleb, in The Black Swan (2007) and Fooled by Randomness (2001), gave the general phenomenon a name that captures its essential character: silent evidence. The term refers to evidence that has been destroyed, hidden, or rendered invisible by the very process being studied. Silent evidence is not merely absent. It is absent in a way that is structurally invisible -- you cannot see it, you cannot know what it would have told you, and its absence systematically distorts whatever conclusions you draw from the evidence that remains.
Taleb illustrates the concept with a story about Diagoras of Melos, a Greek philosopher of the fifth century BCE. Diagoras was shown a collection of votive tablets in a temple, each left by a sailor who had prayed to the gods and survived a shipwreck. "See," his hosts said, "proof that the gods save those who pray." Diagoras replied: "Where are the tablets of those who prayed and drowned?"
The drowned sailors are the silent evidence. They prayed. They did not survive. They did not leave tablets. And their absence from the temple makes the prayer appear to be effective, because only the evidence of success is visible. The evidence of failure is at the bottom of the sea.
Taleb generalizes this insight into a principle: in any system where failure leads to elimination from the sample, the evidence of failure is destroyed by failure itself, and the remaining evidence will systematically overstate the probability of success, the effectiveness of the strategies employed by the survivors, and the reliability of the conditions that preceded survival.
The principle operates with particular force in domains where failure is catastrophic and irreversible. In business, bankruptcy eliminates the company and its data. In war, defeat eliminates the civilization and its records. In evolution, extinction eliminates the species and its genetic strategies. In cultural memory, forgetting eliminates the work and its reputation. In each case, the evidence is not merely hard to find (as with the streetlight effect). The evidence has been annihilated by the process under study. This is the crucial distinction between survivorship bias and the streetlight effect: the streetlight effect means we are not looking in the right place; survivorship bias means the right place no longer exists.
This is what the chapter's threshold concept calls the evidence destroying itself: the selection process that generates the sample simultaneously destroys the counter-evidence that would reveal the sample's bias. The mechanism is not concealment. It is annihilation.
The silent graveyard -- Taleb's metaphor for the accumulated mass of invisible failures -- is the complement to any visible success story. For every bestselling author, there are thousands of equally talented writers whose manuscripts were rejected and who stopped writing. For every successful drug, there are thousands of promising molecules that failed in trials and were abandoned. For every thriving city, there are thousands of settlements that were founded and abandoned. For every surviving religion, there are thousands of belief systems that attracted followers and then vanished. The graveyard is always larger than the city of the living. But the city is all we see.
Connection to Chapter 14 (Absence of Evidence): Chapter 14 examined the logical relationship between absence of evidence and evidence of absence. Survivorship bias adds a structural dimension: in many domains, the absence of evidence is not merely uninformative (as Chapter 14 argued in general) but actively misleading, because the evidence was destroyed by a non-random process. The absence of bullet holes on the engines was not random absence. It was the signature of fatal damage. The absence of failed companies from business databases is not random absence. It is the signature of bankruptcy. When absence is produced by a selection process, absence is evidence -- evidence of the selection process itself.
37.9 Publication Bias -- The File Drawer Problem
In 1979, the psychologist Robert Rosenthal published a paper with a title that would become famous in methodology circles: "The File Drawer Problem and Tolerance for Null Results." Rosenthal's argument was simple and devastating.
Scientific journals preferentially publish studies that find statistically significant results -- studies that report a positive finding, a confirmed hypothesis, an effect that meets the conventional threshold of statistical significance (p < 0.05). Studies that find nothing -- null results, negative results, results that fail to confirm the hypothesis -- are much harder to publish. They are, in Rosenthal's metaphor, consigned to the file drawer: written up, perhaps submitted, rejected, and then filed away where no one will ever see them.
The consequence is that the published scientific literature is a survivorship-biased sample of all the research that has been conducted. The studies that "survived" the publication process are disproportionately the ones that found something. The studies that did not survive -- the null results, the failed replications, the inconclusive findings -- are invisible. And their invisibility systematically inflates the apparent strength, reliability, and generalizability of whatever effects the surviving studies report.
Consider a concrete scenario. Suppose twenty research teams independently investigate whether a particular dietary supplement reduces the risk of heart disease. Suppose that the supplement has no actual effect. By the laws of statistics, roughly one of the twenty teams -- five percent, the conventional significance threshold -- will find a "statistically significant" result purely by chance. That one team publishes its finding. The other nineteen teams, having found nothing, file their results away. The published literature now contains one study showing that the supplement works and zero studies showing that it does not. A doctor reviewing the literature concludes that there is evidence the supplement is effective. The evidence is real. The conclusion is wrong. It is wrong because the published record is a survivorship-biased sample of all the evidence, and the bias has filtered out exactly the evidence that would have corrected the error.
This is not a hypothetical. John Ioannidis, an epidemiologist at Stanford, published a landmark paper in 2005 titled "Why Most Published Research Findings Are False." Ioannidis demonstrated mathematically that under realistic conditions -- small sample sizes, small effect sizes, multiple testing, financial or career incentives to find positive results, and publication bias against null results -- the majority of published research findings in many fields are likely to be false positives. Not wrong because of fraud. Wrong because the publication system is a survivorship filter that preferentially transmits false positives and suppresses the true negatives that would reveal them.
The file drawer problem is survivorship bias operating on the scientific record itself. The "selection process" is the editorial and peer-review system. The "survivors" are the studies that found significant results. The "dead" are the studies that found nothing. And just as the dead bombers could not report where they were hit, the dead studies cannot report what they found -- or rather, what they did not find.
The implications extend beyond individual studies. Meta-analyses -- studies that combine the results of multiple individual studies to estimate the true size of an effect -- are only as good as the studies they include. If the underlying studies are biased by publication bias, the meta-analysis will inherit the bias. A meta-analysis of published studies on the dietary supplement would find an apparently robust effect, because it would be aggregating the survivors. The dead studies -- the ones that found nothing -- would be absent from the analysis, and their absence would make the effect appear larger and more reliable than it is.
🔄 Check Your Understanding
- Explain the file drawer problem in your own words. What is the "selection process," who are the "survivors," and who are the "dead"?
- How can a meta-analysis -- which is supposed to provide more reliable evidence than any single study -- be contaminated by survivorship bias?
- Why does Ioannidis argue that most published research findings in some fields are likely false? What role does publication bias play in his argument?
- Connect the file drawer problem to Wald's bomber analysis. What is the equivalent of "armoring the fuselage" in the context of scientific evidence?
37.10 Countermeasures -- Seeking the Dead
Survivorship bias is not an error that can be corrected by being smarter or more careful within the existing frame. It requires deliberately changing the frame -- actively seeking out the evidence that the selection process has destroyed. The following countermeasures have proven effective in specific domains.
Countermeasure 1: Ask About the Dead
The most fundamental countermeasure is the simplest: for any success story, any body of evidence, any dataset, ask what is not here? Who failed? What was lost? What evidence was destroyed? What companies went bankrupt? What studies found nothing? What planes did not come back? What buildings fell down? What music was forgotten?
This is Wald's insight generalized: before drawing conclusions from the survivors, look for the non-survivors. If you cannot find them -- and often you cannot, because the evidence has been destroyed -- at least acknowledge that they existed and that their absence biases your conclusions.
Countermeasure 2: Base Rate Thinking
Base rate thinking is the practice of starting any analysis with the overall rate of success or failure in the relevant reference class -- the broad category to which the case belongs -- rather than starting with the specific case and its apparently distinctive features.
When evaluating a startup founder's strategy, base rate thinking asks: What is the overall success rate of startups? (Roughly ten to twenty percent survive past five years, depending on the industry and the definition of "success.") Only after anchoring on the base rate do you ask whether this particular founder's strategy shifts the probability.
When evaluating a fund manager's track record, base rate thinking asks: How many fund managers with similar starting conditions produced similar track records purely by chance? If the base rate of chance outperformance is high enough, the track record is not evidence of skill.
When evaluating a medical treatment based on observational data, base rate thinking asks: What is the baseline health trajectory of the type of patient who chooses this treatment? Only after establishing the baseline do you ask whether the treatment adds anything.
Base rate thinking is the antidote to survivorship bias because it forces you to consider the full population -- survivors and non-survivors -- rather than drawing conclusions from the survivors alone.
Connection to Chapter 10 (Base Rates): Chapter 10 examined base rate neglect -- the tendency to ignore background probabilities when evaluating specific cases. Survivorship bias is one of the mechanisms that makes base rate neglect feel natural: by filtering out the failures, the selection process removes the very evidence that would anchor your estimate of the base rate. When you see only the successes, the base rate of success appears to be one hundred percent. Base rate thinking deliberately reintroduces the denominator that survivorship bias has hidden.
Countermeasure 3: The Outside View
The outside view -- a concept developed by Daniel Kahneman and Amos Tversky -- is the practice of evaluating a project, a strategy, or a prediction by examining how similar projects, strategies, or predictions have performed historically, rather than focusing on the specific features of the current case.
The inside view asks: What are the specific features of this project that will determine its outcome? The outside view asks: What is the distribution of outcomes for projects like this one?
The outside view is a countermeasure to survivorship bias because it forces you to look at the full reference class, including the failures. When a city plans to build a new subway system and estimates that it will be completed on time and on budget, the inside view focuses on the specific engineering plans and management structures that make this project different. The outside view examines all comparable subway projects and notes that virtually every one has been late and over budget. The outside view includes the dead -- the projects that failed -- in the analysis.
Countermeasure 4: Pre-registration and Registered Reports
Pre-registration is the practice of publicly registering a study's hypotheses, methods, and analysis plan before collecting data. Registered reports go further: the journal agrees to publish the study based on the quality of the research design, regardless of whether the results are positive, negative, or null.
Pre-registration and registered reports are the most direct countermeasures to publication bias. They eliminate the file drawer by removing the incentive to suppress null results. If the journal has committed to publishing the study regardless of the outcome, the null results survive. The dead are no longer dead. They enter the record alongside the survivors, and the record becomes representative rather than biased.
The adoption of pre-registration has accelerated since the replication crisis in psychology began in the early 2010s. Major journals in psychology, medicine, and other fields now accept or encourage registered reports. The effect on the published record is already visible: registered reports are far more likely to report null results than traditional publications, suggesting that the traditional publication process was, as suspected, systematically filtering out negative findings.
Countermeasure 5: Survivorship-Bias-Free Databases
In finance, the creation of survivorship-bias-free databases -- databases that include the performance records of defunct funds alongside surviving funds -- has transformed the study of fund manager performance. The most widely used survivorship-bias-free database, maintained by the Center for Research in Security Prices (CRSP), includes data on funds that have been closed, merged, or liquidated. Research using these databases consistently finds weaker evidence of manager skill than research using databases that include only surviving funds.
The principle generalizes: in any domain where data disappears when the subject fails, creating and maintaining archives that include the failures alongside the successes is a direct countermeasure to survivorship bias. Business researchers who study failure -- not just success -- produce more accurate models of what actually drives organizational outcomes. Medical researchers who track patients who drop out of clinical trials -- not just patients who complete them -- produce more accurate estimates of treatment effects.
37.11 The Threshold Concept -- The Evidence Destroys Itself
Every chapter in this book contains a threshold concept -- an idea that, once grasped, permanently changes how you see the world. The threshold concept for survivorship bias is this: The Evidence Destroys Itself.
The insight is deeper than "we only see the survivors." It is that in many domains, the process of success or survival systematically eliminates the evidence that would reveal the true causes of success or survival. The elimination is not accidental. It is structural. It is built into the selection mechanism itself. Bankruptcy destroys the company and its records. Military defeat destroys the civilization and its knowledge. Cultural forgetting destroys the work and its reputation. Death removes the patient from the trial. The file drawer hides the null result. The fund closure erases the track record.
The consequences are profound. If the evidence of failure is systematically destroyed by the process of failure, then:
- Success looks easier than it is, because the failed attempts are invisible.
- Winning strategies look more reliable than they are, because the same strategies employed by losers are lost.
- Talent looks more common than it is, because the untalented who tried and failed are gone from the record.
- Risk looks smaller than it is, because the people who took the same risks and were destroyed are not available to testify.
- The past looks better than the present, because time has filtered out everything mediocre, leaving only the best.
- Treatment effects look larger than they are, because the patients who fared worst are no longer in the data.
These are not random errors in random directions. They are systematic errors in the same direction: toward overconfidence, overoptimism, and the false sense that the world is more predictable and more controllable than it actually is.
Before grasping this threshold concept, you see survivorship bias as a correctable error -- a statistical nuisance that more careful sampling could eliminate. You assume that the evidence available to you is roughly representative of all the evidence that exists. You draw confident conclusions from success stories, published studies, historical records, and performance track records.
After grasping this concept, you see survivorship bias as a structural feature of any domain where failure leads to elimination. You recognize that the evidence available to you has already been filtered by a selection process, and that the filtering has systematically removed the evidence most likely to contradict your conclusions. You begin to ask, automatically, of every body of evidence: What has been destroyed by the process that generated this evidence? What would the picture look like if the dead could speak?
How to know you have grasped this concept: When someone tells you a success story, your first thought is "Where are the people who tried the same thing and failed?" When you read a published study, your first thought is "How many unpublished studies found nothing?" When you admire an ancient building, your first thought is "How many ancient buildings fell down?" When you evaluate a fund manager's track record, your first thought is "How many fund managers with similar starting positions have been closed?" You have learned to see the graveyard -- not just the city of the living.
37.12 The Pattern Library Checkpoint
Add survivorship bias to your Pattern Library. Here is the entry:
Pattern: Survivorship Bias (Selection-Filtered Evidence) Structure: Conclusions are drawn from whatever survived a selection process (success, publication, cultural preservation, military victory, fund performance) while the non-survivors -- which constitute the majority of cases and contain the most important corrective information -- are invisible because the selection process destroyed them. The bias systematically makes success look easier, strategies look more reliable, risk look smaller, and the past look better than it was. Signature: Look for any dataset, record, or body of evidence that has been generated by a process that removes failures. If failures are absent not because they did not occur but because the process of failure eliminated them from the record, survivorship bias is operating. Countermeasures: Ask about the dead, base rate thinking, the outside view, pre-registration, survivorship-bias-free databases. Adjacent patterns: Streetlight effect (Ch. 35), base rate neglect (Ch. 10), signal and noise (Ch. 6), absence of evidence (Ch. 14), lifecycle S-curve (Ch. 33), narrative capture (Ch. 36).
Spaced Review Connection: Look back at your Pattern Library entries for the lifecycle S-curve (Ch. 33) and the streetlight effect (Ch. 35). Survivorship bias and the streetlight effect are related but structurally distinct. The streetlight effect says we are looking in the wrong place. Survivorship bias says the right place has been destroyed. The S-curve provides a temporal dimension: in the early stages of a system's lifecycle, survivorship bias is minimal (not much has been selected out yet). In the late stages, survivorship bias is maximal (centuries of selection have removed everything except the most robust survivors). The age of the evidence determines the severity of the survivorship bias. Can you identify a domain where the evidence is young enough that survivorship bias is relatively mild, and one where it is old enough that the bias is severe?
37.13 The Silent Graveyard Surrounds Every Field
This chapter has traced survivorship bias across seven domains: military engineering, business, music, architecture, medicine, military history, and finance -- with a detour through the scientific publication record. In every domain, the same structural pattern operates: a selection process filters out failures, the filtering is invisible, and the remaining evidence systematically overstates the reliability of success, the quality of the past, or the effectiveness of the strategies employed by the survivors.
The connections to Part VI's broader argument about how humans actually decide should be clear. Chapter 34 examined skin in the game -- how the distribution of consequences shapes the reliability of decisions. Chapter 35 examined the streetlight effect -- how the distribution of observable evidence shapes what we study. Chapter 36 examined narrative capture -- how the structure of stories shapes what we believe. This chapter has examined survivorship bias -- how the selection of what survives shapes what we know.
The pattern deepens. Chapter 38 will examine Chesterton's fence -- the failure to ask why a rule, tradition, or institution exists before removing it. Chesterton's fence is, in a sense, the inverse of survivorship bias: where survivorship bias makes us overvalue the survivors (because we cannot see what was lost), Chesterton's fence asks us to respect the survivors (because we may not understand what they are protecting against). The fence survived for a reason. The reason may be invisible. But its invisibility does not mean it is unimportant.
The question this chapter leaves you with is not whether survivorship bias is affecting your understanding of your own field, your own decisions, your own assessment of risk and success. It is. The question is: how large is the graveyard you cannot see? And what would you believe differently if you could walk through it?
Retrieval Prompt: Final check. Without looking back, can you (1) recount Wald's bomber insight and explain why it is the canonical example of survivorship bias, (2) give examples from at least four of the seven domains discussed, (3) explain Taleb's concept of silent evidence and the silent graveyard, (4) describe the file drawer problem and its relationship to publication bias, (5) name at least three countermeasures and explain how each addresses the structural problem, and (6) articulate the threshold concept -- The Evidence Destroys Itself -- in your own words? If you can do all six, you have grasped this chapter's core architecture. If not, revisit the sections where the gaps are.
Summary
Survivorship bias -- the systematic error of drawing conclusions from what survived a selection process while ignoring what did not survive -- operates identically across military engineering (Wald's bombers), business success literature (studying only the winners), music and cultural memory (only the best survives), architecture (only the strongest buildings endure), medicine (the healthy survivor effect), military history (written by the victors), finance (fund manager performance), and the scientific record (publication bias and the file drawer problem). The deeper pattern, identified by Taleb as silent evidence, is that in many domains the process of survival systematically destroys the evidence of failure -- making it structurally impossible to learn from failure unless you deliberately seek out the dead. Countermeasures include asking about the dead, base rate thinking, the outside view, pre-registration, and survivorship-bias-free databases -- but all require the conscious effort to look beyond the visible survivors to the invisible graveyard. The threshold concept -- The Evidence Destroys Itself -- reveals that survivorship bias is not a correctable sampling error but a structural feature of any domain where failure leads to elimination from the record. Success looks easier than it is. Strategies look more reliable than they are. The past looks better than the present. Risk looks smaller than it is. And every conclusion drawn from survivors alone is biased in the same direction: toward overconfidence in a world that is less predictable, less controllable, and less kind than the surviving evidence suggests.
Related Reading
Explore this topic in other books
Metacognition Planning Your Learning Metacognition Calibration Pattern Recognition Map vs Territory Applied Psychology Cognitive Biases Science of Luck Cognitive Biases and Luck Media Literacy Cognitive Biases Sports Betting Cognitive Biases in Betting