Further Reading: The Bootstrap and Simulation-Based Inference
Books
For Deeper Understanding
Tim Hesterberg, "What Teachers Should Know about the Bootstrap" (2015) This paper (published in The American Statistician) is perhaps the single best introduction to the bootstrap for instructors and advanced students. Hesterberg walks through the bootstrap idea with exceptional clarity, provides simulation evidence for when the bootstrap works and when it doesn't, and offers practical advice on choosing between percentile, basic, and BCa methods. Freely available online — start here if you want to go deeper.
Bradley Efron and Robert Tibshirani, An Introduction to the Bootstrap (1993) The definitive textbook, written by the inventor of the bootstrap and his frequent collaborator. Efron and Tibshirani cover the percentile method, the bias-corrected and accelerated (BCa) method, bootstrap hypothesis tests, regression diagnostics, and much more. The writing is remarkably clear for a technical monograph. Chapter 1 alone — "An Overview of Resampling Methods" — is worth reading for the historical perspective and the intuition it builds.
Philip Good, Resampling Methods: A Practical Guide to Data Analysis, 3rd edition (2006) A practical, applications-oriented treatment of bootstrap, permutation tests, and cross-validation. Good (yes, that's his name) focuses on when and how to use resampling methods in practice, with extensive examples from biology, medicine, and social science. Less mathematical than Efron and Tibshirani, more focused on getting the right answer for your data.
Larry Wasserman, All of Statistics: A Concise Course in Statistical Inference (2004) Chapter 8 covers the bootstrap with mathematical precision, including the theory behind why it works and when it fails. Wasserman's treatment of the "bootstrap consistency" theorem (the bootstrap converges to the true sampling distribution under mild conditions) is particularly illuminating for students who want to understand why the bootstrap works, not just that it works. Requires probability theory and calculus.
Charles Wheelan, Naked Statistics: Stripping the Dread from the Data (2013) Wheelan's chapter on statistical inference includes an intuitive discussion of resampling that doesn't require any mathematical background. His analogy comparing the bootstrap to "repeatedly shuffling a deck of cards" captures the spirit of the method. Previously recommended for hypothesis testing (Chapter 13) and confidence intervals (Chapter 12) — his resampling discussion ties those ideas together.
For the Philosophically Curious
Deborah Mayo, Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars (2018) Mayo provides a philosophical defense of frequentist methods (including the bootstrap and permutation tests) as tools for "severe testing" — procedures that are good at detecting when theories are wrong. Her discussion of how simulation-based methods relate to the theoretical foundations of inference is particularly relevant after this chapter. For students who want to think about what inference means, not just how to compute it.
Andrew Gelman and Jennifer Hill, Data Analysis Using Regression and Multilevel/Hierarchical Models (2007) Chapter 7 discusses the bootstrap in the context of regression analysis — a preview of how bootstrap methods will reappear in Chapters 22-23 of this textbook. Gelman and Hill argue that the bootstrap is particularly valuable for regression diagnostics and for constructing CIs when residual assumptions are violated.
Articles and Papers
Efron, B. (1979). "Bootstrap Methods: Another Look at the Jackknife." The Annals of Statistics, 7(1), 1-26. The paper that started it all. Efron introduces the bootstrap, proves its asymptotic validity, and compares it to the jackknife (an older resampling method). The writing is dense but rewarding — every statistician should read the introduction and Section 2 at least once. Available through JSTOR and many university libraries.
Efron, B. (2003). "Second Thoughts on the Bootstrap." Statistical Science, 18(2), 135-140. Twenty-four years after his original paper, Efron reflects on the bootstrap's impact and limitations. He discusses improvements (the BCa method), failures (small-sample bootstrap), and the surprising breadth of applications. A thoughtful retrospective from the method's creator.
Ernst, M. D. (2004). "Permutation Methods: A Basis for Exact Inference." Statistical Science, 19(4), 676-685. A clear, modern treatment of permutation tests, including their historical roots (Fisher, 1935), their relationship to parametric tests, and their advantages and limitations. Ernst emphasizes that permutation tests are "exact" in the sense that they don't rely on asymptotic approximations — the p-value is computed directly from the data.
Hesterberg, T. (2015). "What Teachers Should Know about the Bootstrap: Resampling in the Undergraduate Statistics Curriculum." The American Statistician, 69(4), 371-386. An essential resource for understanding the pedagogy of the bootstrap. Hesterberg demonstrates through simulation that the percentile bootstrap CI can have poor coverage for small samples and skewed data, and recommends the BCa method as a safer default. He also provides practical guidance on how many bootstrap samples ($B$) are needed (his recommendation: $B = 10{,}000$ for CIs, $B = 1{,}000$ for standard errors).
Lock, R. H., Lock, P. F., Morgan, K. L., Lock, E. F., & Lock, D. F. (2013). "Unlocking the Power of Data." Proceedings of the International Conference on Teaching Statistics (ICOTS9). The Lock family (literally — five members of the Lock family!) designed StatKey and the Statistics: Unlocking the Power of Data textbook, which places simulation-based inference at the center of introductory statistics. This paper explains their pedagogical philosophy: teach the bootstrap and permutation tests before formula-based methods, building intuition through simulation. Whether or not you agree with this ordering, their arguments for why simulation builds better statistical thinkers are compelling.
Chernick, M. R., & LaBudde, R. A. (2011). An Introduction to Bootstrap Methods with Applications to R. Wiley. A comprehensive guide to bootstrap methods with R code for every procedure. Covers topics beyond our scope — including the smoothed bootstrap, bootstrap for time series, and bootstrap model selection — but the first four chapters are accessible to students at our level and provide excellent additional examples.
Online Resources
Interactive Tools
StatKey: Bootstrap and Randomization Tests http://www.lock5stat.com/StatKey/ The single best interactive tool for learning the bootstrap and permutation tests. StatKey lets you enter your own data (or use built-in datasets), perform bootstrap resampling and permutation tests, and see the resulting distributions in real time. You can watch bootstrap samples being drawn, see which observations are included and excluded, and observe the bootstrap distribution growing as you add more iterations. Highly recommended for building visual intuition. This tool was previewed in Chapter 16's further reading — now is the time to use it in earnest.
Seeing Theory — Bootstrapping Module https://seeing-theory.brown.edu/frequentist-inference/ Brown University's beautiful interactive visualization includes a bootstrap module where you can draw samples from a population, resample with replacement, and watch the bootstrap distribution form. The visual representation of "sampling from the sample" makes the abstract concept concrete.
Art of Stat: Bootstrap Confidence Intervals https://artofstat.com/web-apps This suite of web apps includes tools for bootstrap CIs and permutation tests. The interface is simple and the output is well-labeled, making it ideal for students who want to practice without writing code.
Video Resources
StatQuest with Josh Starmer: "Bootstrapping Main Ideas" (YouTube) Josh Starmer's trademark energetic style makes the bootstrap accessible and memorable. He walks through the resampling process step by step, shows why with-replacement sampling matters, and builds up to bootstrap confidence intervals. His separate video on permutation tests is equally clear. Watch both in sequence.
3Blue1Brown: "But What Is a Random Variable?" (YouTube) While not specifically about the bootstrap, Grant Sanderson's visual approach to random variables and distributions provides essential background. Understanding that a statistic is itself a random variable — with a distribution that can be simulated — is the key conceptual leap in this chapter.
Khan Academy: "Resampling" (khanacademy.org) Sal Khan provides a clear, step-by-step introduction to resampling methods. The pace is slower than StatQuest, which makes it ideal for students who want extra time to process each step. The worked examples include both bootstrap CIs and permutation tests.
Crash Course Statistics: "P-Hacking and the Replication Crisis" (YouTube) This episode connects the bootstrap to the broader context of scientific reproducibility. While not directly about bootstrap methods, it explains why simulation-based inference has become increasingly popular as researchers seek alternatives to the misuse of p-values from formula-based tests.
jbstatistics: "Introduction to the Bootstrap" (YouTube) Jeremy Balka provides a mathematically precise yet accessible introduction. His treatment of the bootstrap standard error and its relationship to the formula-based SE is particularly clear. He also discusses the BCa method briefly, for students who want to go beyond the percentile method.
Technical Resources
SciPy Documentation: scipy.stats.bootstrap (SciPy ≥ 1.7)
https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.bootstrap.html
SciPy now includes a built-in bootstrap() function that computes bootstrap CIs using the BCa method (more sophisticated than the percentile method we covered). The function accepts any callable statistic and returns a confidence interval object. This is the production-grade tool for bootstrap analysis in Python.
from scipy.stats import bootstrap
import numpy as np
data = np.array([4.2, 3.8, 5.1, 4.7, 3.5, 6.3, 4.4, 5.8, 4.1, 3.9, 4.6, 5.0])
result = bootstrap((data,), np.median, n_resamples=10000,
confidence_level=0.95, method='percentile')
print(f"95% CI: ({result.confidence_interval.low:.2f}, "
f"{result.confidence_interval.high:.2f})")
scikit-learn: Bagging and Random Forests https://scikit-learn.org/stable/modules/ensemble.html The machine learning methods "bagging" (Bootstrap AGGregating) and "random forests" are direct descendants of Efron's bootstrap idea. Bagging trains multiple models on bootstrap samples of the training data and averages their predictions. Understanding the bootstrap from this chapter provides the foundation for understanding these AI methods in Chapter 26.
arch Python package: Bootstrap for Time Series
https://arch.readthedocs.io/en/latest/bootstrap/bootstrap.html
For students interested in financial or time series data, the arch package provides bootstrap methods designed for dependent data (block bootstrap, moving block bootstrap, circular bootstrap). These extend the independent-data bootstrap from this chapter to situations where the standard bootstrap would fail.
Historical Context
Fisher, R. A. (1935). The Design of Experiments. Oliver & Boyd. Fisher's foundational work includes the first description of what we now call the permutation test (he called it the "randomization test"). Fisher used the famous "Lady tasting tea" example to illustrate the logic of testing whether group labels matter — the same logic we used in Section 18.6. The bootstrap came 44 years later, but the permutation test predates it by nearly half a century.
Metropolis, N., & Ulam, S. (1949). "The Monte Carlo Method." Journal of the American Statistical Association, 44(247), 335-341. The paper that named Monte Carlo simulation. Metropolis and Ulam describe using random sampling to solve deterministic mathematical problems — an idea developed during the Manhattan Project for modeling neutron chain reactions. The bootstrap and permutation tests are modern applications of this same principle: using randomness to approximate difficult calculations.
What's Coming Next
Chapter 19 introduces the chi-square test — a formula-based method for analyzing categorical data. Where this chapter used simulation to test group differences, the chi-square test uses a clever comparison of observed vs. expected frequencies. Key resources to preview:
- StatQuest: "Chi-Square Tests" (YouTube) — visual walkthrough of the goodness-of-fit and independence tests
- Khan Academy: "Chi-Square Distribution" (khanacademy.org) — the distribution behind the test
- StatKey: Chi-Square Test module — you can compare the chi-square test to a simulation-based version, connecting Chapter 18's ideas to Chapter 19's formula-based approach
Chapter 20 introduces ANOVA — extending the two-group comparison from Chapter 16 to three or more groups. The bootstrap and permutation ideas from this chapter can also be applied to multi-group comparisons, providing a useful robustness check on ANOVA results.