> "I cannot seriously believe in [quantum mechanics] because the theory cannot be reconciled with the idea that physics should represent a reality in time and space, free from spooky actions at a distance."
Learning Objectives
- Analyze the EPR argument rigorously, identifying the assumptions of locality and realism
- Derive Bell's inequality (CHSH form) step by step, showing that local hidden variable theories are bounded by |S| <= 2
- Evaluate Aspect's (1982) and subsequent loophole-free (2015) experimental results demonstrating S = 2*sqrt(2)
- Implement the quantum teleportation protocol: Bell measurement, classical communication, and unitary correction
- Compare interpretations of QM (Copenhagen, many-worlds, Bohmian, QBism, consistent histories) and evaluate their empirical equivalence
In This Chapter
- 24.1 EPR and the Completeness Question
- 24.2 Hidden Variables: What Einstein Wanted
- 24.3 Bell's Inequality: The Decisive Test
- 24.4 The CHSH Inequality: Modern Form and Full Derivation
- 24.5 Aspect's Experiments and Loophole-Free Tests
- 24.6 Entanglement as a Resource
- 24.7 Quantum Teleportation
- 24.8 Superdense Coding
- 24.9 The No-Cloning Theorem
- 24.10 Interpretations of Quantum Mechanics
- 24.11 Summary and Progressive Project
- Key Equations Summary
Chapter 24: Entanglement, Bell's Theorem, and the Foundations of Quantum Mechanics
"I cannot seriously believe in [quantum mechanics] because the theory cannot be reconciled with the idea that physics should represent a reality in time and space, free from spooky actions at a distance." — Albert Einstein, letter to Max Born (1947)
"Anybody who's not bothered by Bell's theorem has to have rocks in his head." — N. David Mermin (paraphrasing Feynman)
This is the chapter where quantum mechanics gets personal.
Up to now, everything we have done — the Schrödinger equation, angular momentum algebra, perturbation theory, density matrices — has been technically demanding but philosophically manageable. You might have felt uneasy about superposition or measurement, but the formalism worked, the predictions were correct, and you could defer the deep questions. This chapter removes that option.
In 1935, Einstein, Podolsky, and Rosen argued that quantum mechanics must be incomplete — that a deeper, more sensible theory must exist beneath it. In 1964, John Stewart Bell proved a theorem that constrains every such deeper theory. In 1982, Alain Aspect tested Bell's theorem in the laboratory. And in 2015, three independent groups performed loophole-free tests that left no escape.
The verdict: nature is not locally realistic. At least one of the assumptions that Einstein considered non-negotiable — locality (no faster-than-light influences) or realism (physical quantities have definite values before measurement) — must be abandoned. This is not a theoretical curiosity. It is a fact about the universe, confirmed to many standard deviations. And it means that entanglement — the quantum phenomenon that Einstein dismissed as "spooky action at a distance" — is not a bug in the theory. It is the most profoundly non-classical feature of the natural world.
This chapter tells that story, from the EPR argument through Bell's theorem and its experimental verification, to the remarkable protocols (teleportation, superdense coding) that harness entanglement as a resource. We close with an honest, careful survey of the major interpretations of quantum mechanics. You deserve to know where the deepest questions stand — and where they remain genuinely open.
🏃 Fast Track: If you are already comfortable with the EPR argument (Section 24.1) and hidden variables (Section 24.2), skip to Section 24.3 — Bell's inequality — which is the mathematical heart of this chapter. Sections 24.4 (CHSH derivation) and 24.5 (experiments) are essential. Sections 24.6–24.9 (entanglement protocols) can be read in any order but are all needed for Chapter 25. Section 24.10 (interpretations) can be read independently.
24.1 EPR and the Completeness Question
The Setup: What Einstein Actually Argued
The EPR paper — "Can Quantum-Mechanical Description of Physical Reality Be Considered Complete?" (Einstein, Podolsky, and Rosen, 1935) — is one of the most important and most misunderstood papers in the history of physics. Let us state the argument precisely, because its precision is its power.
EPR begins with two definitions:
Completeness: A physical theory is complete if every element of physical reality has a counterpart in the theory.
Reality criterion: If, without in any way disturbing a system, we can predict with certainty (probability equal to unity) the value of a physical quantity, then there exists an element of physical reality corresponding to that quantity.
These seem almost too reasonable to dispute. If you can predict a measurement outcome with 100% certainty without touching the system, it seems perverse to deny that the system "really has" that property. And if a property is real but the theory doesn't describe it, the theory is incomplete.
The Argument
Now consider the Bohm version of the EPR setup (David Bohm reformulated the original EPR argument in terms of spin in 1951, making it cleaner). We prepare two spin-1/2 particles in the singlet state:
$$|\Psi^-\rangle = \frac{1}{\sqrt{2}}\bigl(|{\uparrow}\rangle_A|{\downarrow}\rangle_B - |{\downarrow}\rangle_A|{\uparrow}\rangle_B\bigr)$$
This is one of the four Bell states — the maximally entangled states of two qubits. You constructed these in Chapter 11 and computed their density matrices in Chapter 23. The singlet state has total spin zero: $\hat{\mathbf{S}}_{\text{total}} = \hat{\mathbf{S}}_A + \hat{\mathbf{S}}_B$, with $S_{\text{total}} = 0$.
We send particle $A$ to Alice (in her lab) and particle $B$ to Bob (in his lab, possibly light-years away). Alice and Bob can each choose to measure spin along any axis.
Here is the crucial feature of the singlet state: if Alice measures spin along any axis $\hat{n}$ and gets $+\hbar/2$, then Bob's spin along the same axis $\hat{n}$ is guaranteed to be $-\hbar/2$, and vice versa. This is a perfect anticorrelation, and it holds for any axis. (Verify this: expand the singlet in the $\hat{n}$ eigenbasis — the singlet is rotationally invariant, as you showed in Chapter 14's exercises.)
Now comes the EPR argument:
- Alice measures $S_z$ on particle $A$ and gets $+\hbar/2$.
- She now knows, with certainty, that Bob's particle has $S_z = -\hbar/2$ — without having disturbed Bob's particle in any way (she only touched her own particle, and the labs may be light-years apart).
- By the reality criterion, $S_z$ for particle $B$ is an element of physical reality.
But Alice could have chosen to measure $S_x$ instead. By the same argument:
- If Alice measures $S_x$ on particle $A$, she can predict $S_x$ for Bob's particle with certainty.
- Therefore, $S_x$ for particle $B$ is also an element of physical reality.
The conclusion: both $S_z$ and $S_x$ for Bob's particle are elements of physical reality simultaneously. But quantum mechanics says these cannot both have definite values — $\hat{S}_z$ and $\hat{S}_x$ do not commute, and no quantum state assigns definite values to both. Therefore, quantum mechanics is incomplete: there are elements of physical reality that it fails to describe.
💡 Key Insight: The EPR argument does not claim quantum mechanics is wrong. It claims quantum mechanics is incomplete — that there must be additional information ("hidden variables") that, when known, would assign definite values to all observables simultaneously. This is a much more subtle and harder-to-refute position than saying QM gives wrong predictions.
The Assumptions Behind EPR
The EPR argument relies on two critical assumptions, both of which seemed obvious to Einstein:
Locality: The outcome of a measurement on particle $A$ cannot depend on what measurement is performed on the distant particle $B$ (and vice versa), provided the measurements are space-like separated (i.e., no light signal could travel from one to the other in time).
Realism (counterfactual definiteness): Physical quantities have definite values whether or not they are measured. The moon is there even when nobody is looking.
Together, these assumptions constitute local realism. The EPR argument shows that if local realism holds, quantum mechanics must be incomplete.
Bohr's Response
Niels Bohr's response (also published in 1935, with the same title) rejected the EPR argument by rejecting the reality criterion. Bohr argued that you cannot speak of the "reality" of $S_x$ for Bob's particle unless you actually set up the apparatus to measure $S_x$. The choice of what to measure on Alice's particle determines what kind of experiment is being performed on the composite system, and this affects what can meaningfully be said about Bob's particle — not through any physical influence, but because the very concept of a physical property is defined only relative to a measurement context.
This is a deep philosophical position, and reasonable physicists found both EPR and Bohr's response compelling. For nearly thirty years, the debate seemed to be a matter of philosophical taste with no empirical consequences.
🔵 Historical Note: Einstein was not troubled by the predictions of quantum mechanics. He accepted them entirely. What troubled him was what the theory said about reality. He wanted a theory that described an objective, observer-independent reality evolving deterministically in space and time. Quantum mechanics, as interpreted by Bohr, seemed to deny this possibility. As Einstein put it: "Do you really believe the moon is not there when nobody looks?"
24.2 Hidden Variables: What Einstein Wanted
The Hidden Variable Program
What would a "complete" theory look like? Einstein imagined something like this: in addition to the quantum state $|\psi\rangle$, there exists a set of additional variables $\lambda$ (the "hidden variables") that, together with $|\psi\rangle$, determine the outcome of every measurement with certainty. The apparent randomness of quantum mechanics would then be like the randomness in classical statistical mechanics — not fundamental, but a consequence of our ignorance of $\lambda$.
More precisely, a local hidden variable (LHV) theory asserts:
- Hidden variables exist: Each particle carries hidden variables $\lambda$ drawn from some distribution $\rho(\lambda)$.
- Determinism: The outcome of measuring observable $\hat{A}$ on particle $A$ is a definite function $A(\hat{a}, \lambda)$ that depends on the measurement setting $\hat{a}$ and the hidden variables $\lambda$, but not on the distant measurement setting $\hat{b}$ or its outcome.
- Locality: Alice's result $A(\hat{a}, \lambda)$ does not depend on Bob's setting $\hat{b}$, and Bob's result $B(\hat{b}, \lambda)$ does not depend on Alice's setting $\hat{a}$.
The correlations observed in experiments arise because $A$ and $B$ share the common hidden variables $\lambda$, determined at the source when the particles were created.
Why Hidden Variables Seem Natural
Consider a mundane analogy. A factory produces pairs of gloves, one left and one right, and mails them in sealed boxes to Alice and Bob. When Alice opens her box and sees a left glove, she instantly knows Bob has a right glove — no matter how far away he is. There is no "spooky action at a distance." The correlation was established at the factory, and each glove had a definite handedness all along, encoded in the hidden variable $\lambda = \{\text{A: left, B: right}\}$ or $\lambda = \{\text{A: right, B: left}\}$.
This is exactly how classical correlations work, and it is what EPR-style reasoning suggests should explain quantum correlations. The anticorrelation of the singlet state would simply mean: at the source, particle $A$ was assigned spin-up and particle $B$ spin-down (or vice versa) along every axis simultaneously. The randomness of individual outcomes reflects our ignorance of which assignment was made.
⚠️ Common Misconception: "EPR just didn't understand quantum mechanics." This is historically and intellectually wrong. Einstein understood quantum mechanics as well as anyone alive. His objection was not to the predictions but to the interpretation. He accepted every experimental prediction of QM; he denied that the theory was a complete description of physical reality. This is a coherent and non-trivial philosophical position — and it took a theorem by Bell to show it has testable consequences.
Von Neumann's "Proof" and Its Flaw
In 1932, John von Neumann published what he claimed was a proof that no hidden variable theory could reproduce the predictions of quantum mechanics. This "proof" was accepted by the physics community for decades and used to dismiss the hidden variable program.
It was wrong. Von Neumann's proof contained a hidden assumption: that the expectation value of the sum of two observables equals the sum of the expectation values, $\langle A + B \rangle = \langle A \rangle + \langle B \rangle$, even for incompatible observables ($[\hat{A}, \hat{B}] \neq 0$). While this holds in quantum mechanics, there is no reason it must hold for hidden variable theories — the sum of two non-commuting operators is a different physical quantity from either one alone, and a hidden variable theory need not respect this additivity.
Grete Hermann pointed out this flaw in 1935, but her work was ignored. John Bell rediscovered the flaw independently in 1966 and published a clear analysis. Bell showed that von Neumann's theorem was weaker than claimed — and then went on to prove a much stronger and correct theorem that actually constrains hidden variable theories.
🔵 Historical Note: David Bohm constructed an explicit hidden variable theory in 1952 — a fully deterministic theory that reproduces every prediction of non-relativistic quantum mechanics. This theory (now called Bohmian mechanics or the de Broglie-Bohm theory) serves as an existence proof: hidden variable theories are not ruled out by von Neumann's theorem. However, Bohm's theory is non-local — it requires instantaneous influences between distant particles. Bell's theorem would later show that this non-locality is not a bug in Bohm's particular theory but a necessary feature of any hidden variable theory that reproduces quantum predictions.
24.3 Bell's Inequality: The Decisive Test
Bell's Stroke of Genius
John Stewart Bell's 1964 paper, "On the Einstein Podolsky Rosen Paradox," is four pages long and changed physics forever. What Bell did was astonishingly simple in concept: he showed that local hidden variable theories make a quantitative prediction — an inequality — that quantum mechanics violates.
This transformed the EPR debate from philosophy into experimental physics. It is no longer a matter of taste whether hidden variables exist. It is a matter of measurement.
The Original Bell Inequality
Consider Alice and Bob, each receiving one particle from an entangled pair. Alice can measure spin along axis $\hat{a}$ or $\hat{a}'$, getting results $A = \pm 1$ (in units of $\hbar/2$). Bob can measure along $\hat{b}$ or $\hat{b}'$, getting $B = \pm 1$.
In a local hidden variable theory, the outcomes are determined by the hidden variable $\lambda$:
$$A(\hat{a}, \lambda) = \pm 1, \quad B(\hat{b}, \lambda) = \pm 1$$
The correlation function is:
$$E(\hat{a}, \hat{b}) = \int d\lambda \, \rho(\lambda) \, A(\hat{a}, \lambda) \, B(\hat{b}, \lambda)$$
where $\rho(\lambda) \geq 0$ and $\int d\lambda \, \rho(\lambda) = 1$.
Bell's original inequality (1964) states: for three measurement directions $\hat{a}$, $\hat{b}$, $\hat{c}$:
$$|E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c})| \leq 1 + E(\hat{b}, \hat{c})$$
Proof: Since $A = \pm 1$ and $B = \pm 1$, we have $A^2 = B^2 = 1$. Now:
$$E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c}) = \int d\lambda \, \rho(\lambda) \, A(\hat{a}, \lambda)\bigl[B(\hat{b}, \lambda) - B(\hat{c}, \lambda)\bigr]$$
Taking the absolute value and using $|A| = 1$:
$$|E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c})| \leq \int d\lambda \, \rho(\lambda) \, |B(\hat{b}, \lambda) - B(\hat{c}, \lambda)|$$
Since $B = \pm 1$, we have $|B(\hat{b}, \lambda) - B(\hat{c}, \lambda)| = 1 - B(\hat{b}, \lambda) B(\hat{c}, \lambda)$ (verify by checking all four cases). Therefore:
$$|E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c})| \leq \int d\lambda \, \rho(\lambda) \bigl[1 - B(\hat{b}, \lambda) B(\hat{c}, \lambda)\bigr]$$
Now use the key locality assumption: since $B$ does not depend on Alice's setting, we can write $B(\hat{b}, \lambda) B(\hat{c}, \lambda) = A(\hat{b}, \lambda) B(\hat{b}, \lambda) \cdot [A(\hat{b}, \lambda) B(\hat{c}, \lambda)]$... but more directly, noting that $\int d\lambda \, \rho(\lambda) \, B(\hat{b}, \lambda) B(\hat{c}, \lambda)$ cannot be computed directly because we never measure both $\hat{b}$ and $\hat{c}$ on the same particle. However, in a hidden variable theory, $B(\hat{b}, \lambda)$ and $B(\hat{c}, \lambda)$ are both well-defined for each $\lambda$ (counterfactual definiteness). Using the relation $A(\hat{b}, \lambda) B(\hat{b}, \lambda)$ and the fact that $A(\hat{b}, \lambda)^2 = 1$:
$$|E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c})| \leq 1 + E(\hat{b}, \hat{c})$$
This is Bell's inequality. It is a consequence of locality, realism, and the existence of a joint probability distribution for all measurement outcomes. $\square$
💡 Key Insight: The crucial step in the proof is that $B(\hat{b}, \lambda)$ and $B(\hat{c}, \lambda)$ both exist simultaneously for each $\lambda$ — even though Bob cannot measure both $\hat{b}$ and $\hat{c}$ on the same particle. This is counterfactual definiteness: the insistence that unmeasured quantities have definite values. In quantum mechanics, this does not hold.
Quantum Violation
For the singlet state, quantum mechanics predicts:
$$E_{\text{QM}}(\hat{a}, \hat{b}) = \langle \Psi^- | (\hat{\boldsymbol{\sigma}}_A \cdot \hat{a})(\hat{\boldsymbol{\sigma}}_B \cdot \hat{b}) | \Psi^- \rangle = -\hat{a} \cdot \hat{b} = -\cos\theta_{ab}$$
where $\theta_{ab}$ is the angle between the measurement axes. (We derived this correlation function in Chapter 13 and verified it using the density matrix in Chapter 23.)
For the specific choice $\hat{a} = \hat{z}$, $\hat{b}$ at $45°$, $\hat{c}$ at $90°$ from $\hat{a}$:
$$E(\hat{a}, \hat{b}) = -\cos 45° = -\frac{\sqrt{2}}{2} \approx -0.707$$
$$E(\hat{a}, \hat{c}) = -\cos 90° = 0$$
$$E(\hat{b}, \hat{c}) = -\cos 45° = -\frac{\sqrt{2}}{2} \approx -0.707$$
Bell's inequality requires:
$$|E(\hat{a}, \hat{b}) - E(\hat{a}, \hat{c})| \leq 1 + E(\hat{b}, \hat{c})$$
$$\left|-\frac{\sqrt{2}}{2} - 0\right| \leq 1 + \left(-\frac{\sqrt{2}}{2}\right)$$
$$0.707 \leq 0.293$$
This is false. Quantum mechanics violates Bell's inequality.
✅ Checkpoint: Make sure you understand where locality entered the proof (the assumption that $A$ does not depend on Bob's setting), where realism entered (the assumption that $B(\hat{b}, \lambda)$ and $B(\hat{c}, \lambda)$ both exist), and why QM violates the inequality (entangled states produce stronger correlations than any LHV theory can).
24.4 The CHSH Inequality: Modern Form and Full Derivation
Why CHSH?
Bell's original inequality has a practical disadvantage: it requires perfect correlations, which real experiments never achieve (detectors are imperfect, sources are noisy). In 1969, John Clauser, Michael Horne, Abner Shimony, and Richard Holt derived a more experimentally accessible inequality: the CHSH inequality.
The CHSH Setup
Alice chooses between two measurement settings: $\hat{a}_1$ or $\hat{a}_2$. Bob chooses between two measurement settings: $\hat{b}_1$ or $\hat{b}_2$. Each measurement yields $\pm 1$.
Define the CHSH parameter:
$$S = E(\hat{a}_1, \hat{b}_1) + E(\hat{a}_1, \hat{b}_2) + E(\hat{a}_2, \hat{b}_1) - E(\hat{a}_2, \hat{b}_2)$$
Derivation of the CHSH Bound
Theorem (CHSH, 1969): For any local hidden variable theory, $|S| \leq 2$.
Proof: For each value of $\lambda$, define:
$$s(\lambda) = A_1 B_1 + A_1 B_2 + A_2 B_1 - A_2 B_2$$
where $A_i = A(\hat{a}_i, \lambda)$ and $B_j = B(\hat{b}_j, \lambda)$ for brevity. Factor:
$$s(\lambda) = A_1(B_1 + B_2) + A_2(B_1 - B_2)$$
Now, since $B_1, B_2 = \pm 1$, exactly one of the following is true: - $B_1 = B_2$: then $B_1 + B_2 = \pm 2$ and $B_1 - B_2 = 0$ - $B_1 = -B_2$: then $B_1 + B_2 = 0$ and $B_1 - B_2 = \pm 2$
In either case, one term vanishes and the other has magnitude 2. Since $|A_i| = 1$:
$$|s(\lambda)| = |A_1| \cdot |B_1 + B_2| + |A_2| \cdot |B_1 - B_2| = 0 + 2 = 2 \quad \text{or} \quad 2 + 0 = 2$$
Therefore $|s(\lambda)| = 2$ for every $\lambda$. Averaging over $\lambda$:
$$|S| = \left|\int d\lambda \, \rho(\lambda) \, s(\lambda)\right| \leq \int d\lambda \, \rho(\lambda) \, |s(\lambda)| = 2 \quad \square$$
💡 Key Insight: The CHSH bound $|S| \leq 2$ follows from nothing more than (1) the outcomes are $\pm 1$, (2) $A_i$ depends only on Alice's setting and $\lambda$, (3) $B_j$ depends only on Bob's setting and $\lambda$, and (4) the outcomes are determined by a shared $\lambda$ with distribution $\rho(\lambda)$. Any violation of $|S| > 2$ means at least one of these assumptions is wrong.
The Tsirelson Bound: How Much Can QM Violate?
What does quantum mechanics predict for $S$? For the singlet state and optimal measurement angles, Tsirelson (1980) proved:
$$|S_{\text{QM}}| \leq 2\sqrt{2} \approx 2.828$$
The maximum is achieved with the following settings (in the $xz$-plane):
$$\hat{a}_1 = \hat{z}, \quad \hat{a}_2 = \hat{x}, \quad \hat{b}_1 = \frac{\hat{z} + \hat{x}}{\sqrt{2}}, \quad \hat{b}_2 = \frac{\hat{z} - \hat{x}}{\sqrt{2}}$$
That is, Alice's axes are at $0°$ and $90°$; Bob's are at $45°$ and $-45°$ (or equivalently $135°$). The angles between the four pairs are:
| Pair | Angle | $E = -\cos\theta$ |
|---|---|---|
| $(\hat{a}_1, \hat{b}_1)$ | $45°$ | $-\frac{\sqrt{2}}{2}$ |
| $(\hat{a}_1, \hat{b}_2)$ | $45°$ | $-\frac{\sqrt{2}}{2}$ |
| $(\hat{a}_2, \hat{b}_1)$ | $45°$ | $-\frac{\sqrt{2}}{2}$ |
| $(\hat{a}_2, \hat{b}_2)$ | $135°$ | $+\frac{\sqrt{2}}{2}$ |
Therefore:
$$S = -\frac{\sqrt{2}}{2} - \frac{\sqrt{2}}{2} - \frac{\sqrt{2}}{2} - \left(+\frac{\sqrt{2}}{2}\right) = -2\sqrt{2} \approx -2.828$$
So $|S| = 2\sqrt{2}$, which exceeds the CHSH bound of 2 by about 41%.
📊 By the Numbers: The CHSH inequality neatly stratifies the possible correlation strengths: - $|S| \leq 2$: achievable by local hidden variable theories - $2 < |S| \leq 2\sqrt{2}$: achievable by quantum mechanics but not by LHV theories - $2\sqrt{2} < |S| \leq 4$: not achievable by quantum mechanics either (but permitted by some hypothetical "super-quantum" theories — see Popescu-Rohrlich boxes)
The fact that quantum mechanics does not saturate the algebraic maximum of $|S| = 4$ is itself a deep and not fully understood constraint.
Full Quantum Calculation
Let us compute $S$ in detail for the singlet state. Recall from Chapter 13 that for a spin-1/2 particle measured along axis $\hat{n} = (\sin\theta\cos\phi, \sin\theta\sin\phi, \cos\theta)$:
$$\hat{\boldsymbol{\sigma}} \cdot \hat{n} = \begin{pmatrix} \cos\theta & \sin\theta \, e^{-i\phi} \\ \sin\theta \, e^{i\phi} & -\cos\theta \end{pmatrix}$$
For the correlation function, we need:
$$E(\hat{a}, \hat{b}) = \langle \Psi^- | (\hat{\boldsymbol{\sigma}}_A \cdot \hat{a}) \otimes (\hat{\boldsymbol{\sigma}}_B \cdot \hat{b}) | \Psi^- \rangle$$
Using the identity (which you proved in Exercise 13.C.4):
$$\langle \Psi^- | \sigma_{A,i} \sigma_{B,j} | \Psi^- \rangle = -\delta_{ij}$$
we get:
$$E(\hat{a}, \hat{b}) = -\sum_i a_i b_i = -\hat{a} \cdot \hat{b} = -\cos\theta_{ab}$$
This is the key result: the quantum correlation depends only on the angle between the measurement axes, and it varies as $-\cos\theta_{ab}$ — a smooth sinusoidal function. Any LHV theory, by contrast, must produce correlations that satisfy $|S| \leq 2$, which constrains the correlation function to be "less curvy" — closer to a piecewise-linear zig-zag.
🔗 Connection: The quantity $E(\hat{a}, \hat{b}) = -\hat{a} \cdot \hat{b}$ for the singlet follows directly from the rotational invariance we proved in Chapter 14, combined with the partial trace and correlation machinery of Chapter 23. The formalism we built over the last thirteen chapters was exactly what we needed to reach this point.
24.5 Aspect's Experiments and Loophole-Free Tests
The Experimental Challenge
Bell's theorem (1964) showed that local realism makes a testable prediction. But testing it required:
- A reliable source of entangled particle pairs
- Efficient detectors capable of measuring individual particles
- The ability to choose measurement settings quickly enough that no signal could travel between Alice and Bob during the measurement
- Sufficient statistical precision to distinguish $S = 2$ from $S = 2\sqrt{2}$
First Generation: Freedman and Clauser (1972)
Stuart Freedman and John Clauser performed the first experimental test of a Bell inequality in 1972, using entangled photon pairs from an atomic cascade in calcium. They observed a violation of the Bell-CHSH inequality by 6 standard deviations.
However, the measurement settings were fixed — not switched randomly during the experiment. This left open the locality loophole: perhaps the measurement devices on Alice's side somehow communicated with Bob's side (or with the source) before the measurement was complete.
Aspect's Experiments (1982): Closing the Locality Loophole
Alain Aspect and his team in Orsay, France, performed the definitive early test. Their key innovation was rapid switching of measurement settings using acousto-optic modulators. The settings were changed every ~10 nanoseconds — far faster than light could travel between the two detectors (separated by 12 meters, corresponding to ~40 ns light travel time).
The setup used polarization-entangled photon pairs from an atomic cascade in calcium:
$$|\Psi\rangle = \frac{1}{\sqrt{2}}\bigl(|H\rangle_A|H\rangle_B + |V\rangle_A|V\rangle_B\bigr)$$
where $|H\rangle$ and $|V\rangle$ denote horizontal and vertical polarization. (This is the $|\Phi^+\rangle$ Bell state — equivalent to the singlet up to a local rotation.)
Aspect's results:
- Measured: $S_{\text{exp}} = 2.697 \pm 0.015$
- QM prediction: $S_{\text{QM}} = 2\sqrt{2} \approx 2.828$ (reduced by experimental imperfections)
- LHV bound: $S \leq 2$
- Violation: More than 40 standard deviations above the LHV bound
🧪 Experiment: Aspect's setup used two-channel polarizers (each photon is either transmitted or reflected, giving $+1$ or $-1$) rather than one-channel polarizers (transmitted or absorbed). This doubled the data rate and eliminated a significant systematic uncertainty. The rapid switching ensured that Alice's polarizer setting was chosen after the photon had left the source, closing the locality loophole for practical purposes.
The Detection Loophole
Aspect's experiment (and all photon-based experiments for decades after) had one remaining vulnerability: the detection loophole (also called the fair sampling loophole).
No photon detector is perfectly efficient. In Aspect's experiment, only about 5% of photon pairs were actually detected. If the detected pairs are a biased sample — if, say, the hidden variables determine which photons are detected — then a local hidden variable theory could fake a Bell violation even though the complete ensemble satisfies $|S| \leq 2$.
The detection loophole requires detector efficiency above a critical threshold. For the CHSH inequality with maximally entangled states, the threshold is $\eta_{\text{crit}} \approx 82.8\%$ (Eberhard, 1993).
Loophole-Free Tests (2015): The End of the Road
In 2015, three independent experiments finally closed all loopholes simultaneously:
1. Delft (Hensen et al., Nature, 2015): Used entangled electron spins in nitrogen-vacancy centers in diamond, separated by 1.3 km. Detection efficiency: ~96%. Space-like separation enforced by 1.3 km distance and fast random number generators. Result: $S = 2.42 \pm 0.20$, $p < 0.039$.
2. Vienna (Giustina et al., PRL, 2015): Used polarization-entangled photon pairs from SPDC (spontaneous parametric down-conversion) with superconducting nanowire detectors achieving ~75% system efficiency. Detectors separated by 58 meters with rapid setting switching. Result: $p < 3.7 \times 10^{-31}$ for the null hypothesis of local realism (using an Eberhard-type inequality optimized for asymmetric efficiency).
3. NIST Boulder (Shalm et al., PRL, 2015): Similar photonic setup with superconducting detectors, 184-meter separation. Result: $p < 2.3 \times 10^{-7}$.
These experiments closed the locality, detection, and freedom-of-choice loopholes simultaneously. Local realism is dead. No reasonable escape remains.
📊 By the Numbers: The 2015 Vienna experiment achieved a $p$-value of $3.7 \times 10^{-31}$ — roughly the probability of flipping a fair coin and getting heads 100 times in a row. By the standards of any experimental science, this is beyond decisive. The 2022 Nobel Prize in Physics was awarded to Alain Aspect, John Clauser, and Anton Zeilinger "for experiments with entangled photons, establishing the violation of Bell inequalities."
⚠️ Common Misconception: "Bell's theorem proves that quantum mechanics is non-local — that measurements on Alice's particle instantly affect Bob's particle." This is too strong. Bell's theorem proves that local realism is false. This means at least one of the following must be abandoned: (1) locality, (2) realism (definite pre-measurement values), or (3) the freedom to choose measurement settings independently (sometimes called "superdeterminism"). Different interpretations make different choices about which to give up — see Section 24.10.
24.6 Entanglement as a Resource
Having established that entanglement produces correlations no classical system can match, we now shift perspective. Rather than viewing entanglement as a philosophical puzzle, we treat it as a resource — something that can be produced, distributed, consumed, and used to accomplish tasks impossible with classical resources alone.
This resource perspective is the foundation of quantum information science (Chapter 25).
The Four Bell States
The maximally entangled states of two qubits — the Bell states or EPR pairs — form an orthonormal basis for the two-qubit Hilbert space:
$$|\Phi^+\rangle = \frac{1}{\sqrt{2}}\bigl(|00\rangle + |11\rangle\bigr)$$
$$|\Phi^-\rangle = \frac{1}{\sqrt{2}}\bigl(|00\rangle - |11\rangle\bigr)$$
$$|\Psi^+\rangle = \frac{1}{\sqrt{2}}\bigl(|01\rangle + |10\rangle\bigr)$$
$$|\Psi^-\rangle = \frac{1}{\sqrt{2}}\bigl(|01\rangle - |10\rangle\bigr)$$
where we use the computational basis $|0\rangle \equiv |{\uparrow}\rangle$ and $|1\rangle \equiv |{\downarrow}\rangle$.
Each Bell state is related to the others by local unitary operations (operations on a single qubit):
$$|\Phi^-\rangle = (\hat{\sigma}_z \otimes \hat{I})|\Phi^+\rangle, \quad |\Psi^+\rangle = (\hat{\sigma}_x \otimes \hat{I})|\Phi^+\rangle, \quad |\Psi^-\rangle = (i\hat{\sigma}_y \otimes \hat{I})|\Phi^+\rangle$$
🔗 Connection: You constructed the Bell states and computed their properties in Chapter 11. You computed the reduced density matrix of each Bell state in Chapter 23 and found it was the maximally mixed state $\rho_A = \frac{1}{2}\hat{I}$. This means each individual particle is maximally uncertain — all the information is in the correlations, none in the marginals.
Quantifying Entanglement
For a pure bipartite state $|\psi\rangle_{AB}$, the entanglement is quantified by the entanglement entropy:
$$E(|\psi\rangle) = S(\rho_A) = -\text{Tr}(\rho_A \log_2 \rho_A)$$
where $\rho_A = \text{Tr}_B(|\psi\rangle\langle\psi|)$ is the reduced density matrix. For the Bell states, $E = 1$ ebit (one bit of entanglement), the maximum for two qubits. For a product state, $E = 0$.
The entanglement entropy tells you how many Bell pairs' worth of entanglement a state contains. This is not just a mathematical curiosity — it determines the state's usefulness for quantum protocols.
24.7 Quantum Teleportation
Quantum teleportation, proposed by Bennett, Brassard, Crepeau, Jozsa, Peres, and Wootters in 1993, is the most dramatic demonstration that entanglement is a resource. It allows the transfer of an arbitrary quantum state from one location to another, using only shared entanglement and classical communication — without physically transporting the particle.
The Protocol
Resources: Alice and Bob share a Bell pair $|\Phi^+\rangle_{AB}$, prepared in advance. Alice also has a qubit in an unknown state $|\chi\rangle_C = \alpha|0\rangle_C + \beta|1\rangle_C$ that she wants to send to Bob.
The total three-qubit state is:
$$|\Psi\rangle = |\chi\rangle_C \otimes |\Phi^+\rangle_{AB} = (\alpha|0\rangle_C + \beta|1\rangle_C) \otimes \frac{1}{\sqrt{2}}(|00\rangle_{AB} + |11\rangle_{AB})$$
Step 1: Bell measurement. Alice performs a joint measurement on qubits $C$ and $A$ in the Bell basis. To see what happens, we rewrite the three-qubit state by expressing qubits $C$ and $A$ in the Bell basis.
The key identity (verify by direct expansion):
$$|0\rangle_C|0\rangle_A = \frac{1}{\sqrt{2}}(|\Phi^+\rangle_{CA} + |\Phi^-\rangle_{CA})$$
$$|0\rangle_C|1\rangle_A = \frac{1}{\sqrt{2}}(|\Psi^+\rangle_{CA} + |\Psi^-\rangle_{CA})$$
$$|1\rangle_C|0\rangle_A = \frac{1}{\sqrt{2}}(|\Psi^+\rangle_{CA} - |\Psi^-\rangle_{CA})$$
$$|1\rangle_C|1\rangle_A = \frac{1}{\sqrt{2}}(|\Phi^+\rangle_{CA} - |\Phi^-\rangle_{CA})$$
Substituting and collecting terms:
$$|\Psi\rangle = \frac{1}{2}\Bigl[|\Phi^+\rangle_{CA}(\alpha|0\rangle_B + \beta|1\rangle_B) + |\Phi^-\rangle_{CA}(\alpha|0\rangle_B - \beta|1\rangle_B)$$
$$\quad + |\Psi^+\rangle_{CA}(\beta|0\rangle_B + \alpha|1\rangle_B) + |\Psi^-\rangle_{CA}(-\beta|0\rangle_B + \alpha|1\rangle_B)\Bigr]$$
Step 2: Alice communicates her result. Alice's Bell measurement yields one of four outcomes, each with probability $1/4$. She sends 2 classical bits to Bob telling him which outcome she got.
Step 3: Bob applies a correction. Depending on Alice's result, Bob's qubit is in one of four states:
| Alice's result | Bob's state | Bob's correction |
|---|---|---|
| $\|\Phi^+\rangle_{CA}$ | $\alpha\|0\rangle + \beta\|1\rangle$ | $\hat{I}$ (nothing) |
| $\|\Phi^-\rangle_{CA}$ | $\alpha\|0\rangle - \beta\|1\rangle$ | $\hat{\sigma}_z$ |
| $\|\Psi^+\rangle_{CA}$ | $\beta\|0\rangle + \alpha\|1\rangle$ | $\hat{\sigma}_x$ |
| $\|\Psi^-\rangle_{CA}$ | $-\beta\|0\rangle + \alpha\|1\rangle$ | $i\hat{\sigma}_y$ |
After Bob's correction, his qubit is always in the state $\alpha|0\rangle + \beta|1\rangle = |\chi\rangle$. The teleportation is complete.
💡 Key Insight: Note what has happened. The state $|\chi\rangle = \alpha|0\rangle + \beta|1\rangle$ has been transferred from Alice to Bob, even though: - Alice never learned $\alpha$ and $\beta$ (her Bell measurement gives only 2 bits of information, while the state $|\chi\rangle$ requires an infinite amount of classical information to specify). - No physical particle traveled from Alice to Bob (only 2 classical bits were sent). - The protocol consumed one Bell pair — the entanglement is used up. - The original state $|\chi\rangle$ is destroyed at Alice's location (it must be, by the no-cloning theorem — see Section 24.9).
Experimental Realizations
Quantum teleportation has been demonstrated experimentally many times:
- Innsbruck (1997): Bouwmeester et al. teleported the polarization state of a photon over ~1 meter.
- Canary Islands (2012): Ma et al. teleported photon states over 143 km between two islands.
- Hefei-Shanghai (2017): Ren et al. teleported photon states from ground to the Micius satellite at 1,400 km altitude.
- Delft (2022): Hermans et al. teleported quantum states between non-neighboring nodes of a three-node quantum network.
⚠️ Common Misconception: "Quantum teleportation allows faster-than-light communication." It does not. Bob cannot extract any information from his qubit until he receives Alice's 2-bit classical message, which travels at the speed of light or slower. Before receiving Alice's message, Bob's qubit is in a maximally mixed state $\frac{1}{2}\hat{I}$ — it looks completely random. The classical communication is essential and enforces the speed-of-light limit.
24.8 Superdense Coding
Superdense coding (Bennett and Wiesner, 1992) is, in a sense, the mirror image of teleportation. While teleportation uses 1 Bell pair + 2 classical bits to transmit 1 qubit, superdense coding uses 1 Bell pair + 1 qubit to transmit 2 classical bits.
The Protocol
Alice and Bob share a Bell pair $|\Phi^+\rangle_{AB}$. Alice wants to send 2 classical bits to Bob.
Step 1: Depending on which 2-bit message Alice wants to send, she applies one of four operations to her qubit:
| Message | Alice's operation | Resulting state |
|---|---|---|
| $00$ | $\hat{I}$ | $\|\Phi^+\rangle$ |
| $01$ | $\hat{\sigma}_z$ | $\|\Phi^-\rangle$ |
| $10$ | $\hat{\sigma}_x$ | $\|\Psi^+\rangle$ |
| $11$ | $i\hat{\sigma}_y$ | $\|\Psi^-\rangle$ |
Step 2: Alice sends her qubit to Bob. (This is a single qubit — one physical particle.)
Step 3: Bob now has both qubits of a Bell pair. He performs a Bell measurement (a joint measurement in the Bell basis) and reads off the two-bit message.
The remarkable feature: by sending just one qubit, Alice transmits 2 classical bits of information. This is twice the classical capacity of a single qubit channel.
💡 Key Insight: The "extra" bit of information is not carried by the qubit Alice sends. It was already present in the pre-shared entanglement. Superdense coding demonstrates that shared entanglement + quantum communication > classical communication. Entanglement is a resource that enhances communication capacity.
Resource Accounting: Teleportation vs. Superdense Coding
| Protocol | Consumes | Transmits |
|---|---|---|
| Teleportation | 1 Bell pair + 2 cbits | 1 qubit |
| Superdense coding | 1 Bell pair + 1 qubit | 2 cbits |
These are dual protocols: each trades entanglement and one type of communication for the other. This duality is a deep feature of quantum information theory.
24.9 The No-Cloning Theorem
The no-cloning theorem (Wootters and Zurek, 1982; Dieks, 1982) states that there is no quantum operation that can take an arbitrary unknown quantum state $|\psi\rangle$ and produce two copies of it:
$$|\psi\rangle \not\to |\psi\rangle|\psi\rangle \quad \text{(for arbitrary } |\psi\rangle\text{)}$$
Proof
Suppose a cloning machine exists. It acts as a unitary transformation $\hat{U}$ on the state to be cloned and an initially blank target:
$$\hat{U}|\psi\rangle|0\rangle = |\psi\rangle|\psi\rangle$$
This must work for any state. Take two states $|\psi\rangle$ and $|\phi\rangle$:
$$\hat{U}|\psi\rangle|0\rangle = |\psi\rangle|\psi\rangle$$
$$\hat{U}|\phi\rangle|0\rangle = |\phi\rangle|\phi\rangle$$
Take the inner product:
$$\langle\psi|\phi\rangle \langle 0|0\rangle = \langle\psi|\phi\rangle^2$$
$$\langle\psi|\phi\rangle = \langle\psi|\phi\rangle^2$$
This equation has only two solutions: $\langle\psi|\phi\rangle = 0$ or $\langle\psi|\phi\rangle = 1$. So the machine can clone states only if they are either identical or orthogonal — it cannot clone an arbitrary unknown state. $\square$
Implications
The no-cloning theorem has far-reaching consequences:
-
Teleportation must destroy the original. If teleportation produced a copy without destroying the original, we could combine it with a measurement to violate the uncertainty principle. The no-cloning theorem enforces this.
-
Quantum information cannot be broadcast. You cannot send one quantum state to many recipients (no quantum "photocopier").
-
Quantum key distribution is secure. An eavesdropper cannot copy quantum states without disturbing them — the foundation of quantum cryptography (BB84 protocol).
-
Error correction is hard. You cannot simply copy a qubit for redundancy as you would a classical bit. Quantum error correction (Chapter 35) requires much more sophisticated techniques.
🔗 Connection: The no-cloning theorem connects to the formalism of Chapter 6 (measurement changes the state), Chapter 11 (entanglement cannot be created by local operations), and Chapter 23 (the impossibility of distinguishing non-orthogonal mixed states perfectly). It will be essential in Chapter 25 (quantum computing) and Chapter 35 (quantum error correction).
24.10 Interpretations of Quantum Mechanics
We have now established the experimental facts: entanglement is real, Bell inequalities are violated, local realism is false. But what does this mean about the nature of reality? Quantum mechanics tells us with exquisite precision what happens when we make measurements. It is silent about what is happening when we do not.
This section presents the major interpretations honestly and even-handedly. Each interpretation agrees with every experimental prediction of quantum mechanics. They differ in what they say about ontology — what exists, what is real, what happens between measurements. No experiment can currently distinguish them. This is not a failure of physics; it is an honest acknowledgment of an unsolved problem.
We present each interpretation, its strengths, its weaknesses, and its response to the puzzles raised by Bell's theorem and entanglement.
24.10.1 Copenhagen Interpretation
Central idea: The quantum state $|\psi\rangle$ is a tool for calculating probabilities of measurement outcomes. It is not a description of an objective physical reality. Measurement is a fundamental, irreducible process that cannot be analyzed further in quantum terms.
Key features: - The wave function is epistemic (represents knowledge, not reality) - Measurement causes "collapse": $|\psi\rangle \to |a_n\rangle$ upon observing eigenvalue $a_n$ - There is no description of what happens "between" measurements - Complementarity: wave and particle descriptions are complementary; both are needed but cannot be applied simultaneously - The classical world exists and is the precondition for quantum mechanics (measuring devices must be described classically)
Response to Bell/EPR: There is no paradox because there is no reality to be "spooky" about. Before measurement, the spins do not have definite values. The correlations are a feature of the quantum state, which describes the preparation procedure, not a hidden physical reality. No non-local influence occurs because there was no local reality to influence.
Strengths: - Operationally clear: tells you exactly how to calculate - Minimal metaphysical baggage - The pragmatic default used by most working physicists - Well-tested recipe for extracting predictions
Weaknesses: - The measurement problem: what counts as a "measurement"? Where is the boundary between quantum and classical? - Relies on classical concepts (measuring devices) that are supposedly derived from quantum mechanics — circular? - Vague about what exists when nobody is looking - Does not explain why outcomes are random — just declares it
Prominent advocates: Niels Bohr, Werner Heisenberg, Wolfgang Pauli, and (arguably) most contemporary physicists by default.
⚖️ Interpretation: The Copenhagen interpretation is sometimes accused of being "shut up and calculate." This is unfair to Bohr, who had deep philosophical reasons for his positions. But it is true that many physicists adopt Copenhagen in practice while being uncertain about its philosophical commitments.
24.10.2 Many-Worlds Interpretation (Everettian Quantum Mechanics)
Central idea: The wave function is real (ontic) and the only physical entity. There is no collapse. When a measurement is made, the universe "branches" — all possible outcomes actually occur in different branches of a universal wave function. What we perceive as a single definite outcome is our subjective experience within one branch.
Key features: - The Schrödinger equation applies universally and without exception — no collapse postulate - Measurement is just entanglement between the system and the apparatus (and the observer) - All branches are equally real - Probability arises from the "self-locating uncertainty" of an observer who does not know which branch they are in - The universal wave function $|\Psi\rangle_{\text{universe}}$ is the complete description of reality
Response to Bell/EPR: In many-worlds, every possible set of measurement outcomes occurs in some branch. There is no non-locality because nothing needs to "travel" — the correlations are simply features of the branch structure of the universal wave function. Bell violations are accommodated naturally: the universe does not need to be locally realistic because the local branches encode the correlations through their global structure.
Strengths: - Ontologically clean: one equation (Schrödinger), one entity (the wave function), no exceptions - Solves the measurement problem by eliminating collapse - No preferred role for observers or "consciousness" - Takes the wave function seriously as describing reality
Weaknesses: - The preferred basis problem: what determines which branches are "definite outcomes" and not some other superposition? - The probability problem: if all outcomes occur, why do relative frequencies match the Born rule? (Deutsch, Wallace, and others have proposed derivations, but they remain debated.) - Extreme ontological extravagance: an infinity of unobservable parallel universes - The concept of "branching" is harder to define precisely than it sounds - It is not clear what it even means for an observer to be "in" one branch
Prominent advocates: Hugh Everett III (originator), Bryce DeWitt (popularizer), David Deutsch, Sean Carroll, David Wallace, Lev Vaidman.
⚖️ Interpretation: Many-worlds is increasingly popular among theoretical physicists and philosophers of physics, particularly those working in quantum information and quantum gravity. Its supporters argue that the ontological extravagance (many worlds) is the price of theoretical elegance (one equation, no exceptions). Its critics argue that theoretical elegance is not enough — you need to explain why the Born rule holds, and this remains contested.
24.10.3 Bohmian Mechanics (de Broglie-Bohm Theory)
Central idea: Particles have definite positions at all times. The wave function $\psi$ is a real physical field that guides the particles via a "guidance equation." The randomness of quantum mechanics arises from ignorance of the exact initial positions.
Key features: - Particles always have definite positions: $\mathbf{q}_1(t), \mathbf{q}_2(t), \ldots$ - The wave function $\psi(\mathbf{q}_1, \ldots, \mathbf{q}_N, t)$ evolves by the Schrödinger equation (no collapse) - Particle velocities are determined by the guidance equation: $\dot{\mathbf{q}}_k = \frac{\hbar}{m_k}\text{Im}\frac{\nabla_k \psi}{\psi}$ - Measurement outcomes are determined by the actual particle positions - The Born rule $|\psi|^2$ is the equilibrium distribution, analogous to thermal equilibrium in classical statistical mechanics - Explicitly non-local: the velocity of particle $k$ depends instantaneously on the positions of all other particles through $\psi$
Response to Bell/EPR: Bohmian mechanics is explicitly non-local and makes no attempt to be otherwise. When Alice measures her particle, the wave function changes globally, instantly altering the guidance equation for Bob's particle. Bell's theorem is satisfied because the theory abandons locality (not realism). The particles always have definite properties — what is sacrificed is the idea that distant measurements cannot influence local particle dynamics.
Strengths: - Solves the measurement problem completely: measurements have definite outcomes because particles have definite positions - Recovers all predictions of standard QM (this has been proven rigorously) - Conceptually clear: particles exist, they move, there is no vagueness about what is "real" - Takes the formalism seriously: the wave function is real, the particles are real - No preferred role for observers
Weaknesses: - Explicitly non-local: the guidance equation involves instantaneous action at a distance (though this cannot be used to send signals — the "no-signaling" property holds) - Positions are privileged over other observables — why? - Difficult to extend to relativistic quantum field theory (though partial extensions exist) - The wave function lives in configuration space (3N dimensions for N particles), not in physical 3D space — what kind of "real" entity is that? - Requires a preferred foliation of spacetime (a universal "now"), which sits uneasily with special relativity
Prominent advocates: Louis de Broglie (originator), David Bohm (reviver), John Bell (sympathizer), Sheldon Goldstein, Detlef Dürr, Nino Zanghì.
⚖️ Interpretation: John Bell himself was sympathetic to Bohmian mechanics, calling it "the most serious of the interpretations." It serves a crucial role as an existence proof: it shows that a fully deterministic, realist theory that reproduces all of quantum mechanics is possible — at the price of non-locality. Bell's theorem guarantees this price must be paid.
24.10.4 QBism (Quantum Bayesianism)
Central idea: The quantum state is not a property of a physical system but a representation of an agent's personal beliefs about what will happen if they interact with the system. Quantum mechanics is a normative theory — it tells agents how to manage their expectations — not a descriptive theory of objective reality.
Key features: - Probabilities in QM are subjective (Bayesian), not objective frequencies - The quantum state $|\psi\rangle$ encodes an agent's beliefs, not the system's properties - Measurement is an action taken by an agent; the "outcome" is the agent's personal experience - There is no "view from nowhere" — quantum mechanics is always from the perspective of a particular agent - The Born rule is a normative consistency constraint (like the Dutch book argument in classical Bayesian probability) - There is no measurement problem because there is no objective collapse
Response to Bell/EPR: In QBism, there is no paradox because there was never an objective state to collapse or a real hidden variable to discover. Bell's theorem shows that certain kinds of agent beliefs (specifically, beliefs that conform to local hidden variable models) are inconsistent with quantum mechanics. This is a constraint on rational belief, not a statement about non-local influences in the physical world. Alice's measurement updates Alice's beliefs; it does not change anything about Bob's physical situation.
Strengths: - Dissolves the measurement problem: measurement is personal experience, not objective collapse - No non-locality: beliefs update locally - Takes probability in QM seriously as genuine uncertainty, not determinism-in-disguise - Consistent with special relativity (no preferred frame, no FTL influences) - Avoids the baggage of many worlds or Bohmian trajectories
Weaknesses: - What is the physical world, independent of agents? QBism is often accused of solipsism (though QBists deny this) - If quantum mechanics is just about agents' beliefs, why is it so extraordinarily successful at predicting objective experimental outcomes? - The status of intersubjective agreement (why do all agents agree on measurement statistics?) is not fully explained - Many physicists find it unsatisfying: "I want to know what the electron does, not what I should believe about the electron" - The agent-centered framework sits uncomfortably with cosmological applications (who was the agent during the Big Bang?)
Prominent advocates: Christopher Fuchs, Rüdiger Schack, N. David Mermin.
⚖️ Interpretation: QBism is the most radical of the mainstream interpretations — it denies that quantum mechanics describes an observer-independent reality. Its supporters see this as intellectual honesty; its critics see it as giving up on the explanatory ambition of physics. The debate is genuine and ongoing.
24.10.5 Consistent (Decoherent) Histories
Central idea: Quantum mechanics assigns probabilities to entire histories (sequences of events), not just to individual measurement outcomes. A "consistent" or "decoherent" set of histories is one for which probability rules (like the sum rule) apply without interference between branches.
Key features: - No collapse postulate and no observers required - Probabilities are assigned to "coarse-grained histories" — sequences of approximate properties at successive times - A set of histories is "consistent" if there is no quantum interference between different histories in the set (the decoherence condition) - Many different consistent sets exist, and the theory does not pick one as preferred - Classical physics emerges for history sets where decoherence is strong
Response to Bell/EPR: Consistent histories accommodates Bell violations by noting that certain sets of histories — specifically those that assign values to incompatible observables — are not consistent (they interfere). The EPR paradox arises from attempting to reason about a set of histories that is not decoherent. Within any consistent set, there is no paradox.
Strengths: - No measurement problem: the framework applies to any physical system, including the universe as a whole - Formally precise and mathematically rigorous - Compatible with quantum cosmology (no external observers needed) - Reduces to classical logic for decoherent histories
Weaknesses: - The multiplicity of consistent sets: which one describes "what actually happened"? The theory does not say. - Critics argue this amounts to "anything goes" — the theory is too permissive - The meaning of probability for a single universe (the cosmological application) is unclear - Somewhat abstract and technically demanding, which limits its adoption
Prominent advocates: Robert Griffiths (originator), Murray Gell-Mann and James Hartle (independent development), Roland Omnès.
24.10.6 Other Approaches (Brief Mentions)
Several other interpretive approaches deserve mention:
Objective collapse theories (GRW, Penrose): The wave function is real and undergoes spontaneous, random collapse at a rate that is negligible for individual particles but becomes rapid for macroscopic objects. These are not interpretations of standard QM — they are modifications that make different predictions, in principle testable. Current experiments constrain GRW parameters but have not yet confirmed or ruled out the theory.
Relational quantum mechanics (Rovelli): Physical quantities are defined only relative to other systems. There is no absolute state of a system — only states relative to an observer. Similar in spirit to QBism but with a more realist flavor.
Superdeterminism: The measurement settings are not free — they are correlated with the hidden variables through the initial conditions of the universe. This closes the "freedom of choice" loophole in Bell's theorem. Almost all physicists reject this because it undermines the possibility of doing science (if measurement settings are predetermined, no experiment is a genuine test).
Comparison Table
| Feature | Copenhagen | Many-Worlds | Bohmian | QBism | Consistent Histories |
|---|---|---|---|---|---|
| Wave function is... | Epistemic (knowledge) | Ontic (real) | Ontic (real field) | Epistemic (beliefs) | Tool for history probabilities |
| Collapse? | Yes (postulated) | No | No (apparent only) | No (belief update) | No |
| Deterministic? | No | Yes (universal $\Psi$) | Yes (given positions) | Not applicable | Not applicable |
| Non-local? | Ambiguous | No (all local in each branch) | Yes (explicitly) | No (beliefs are local) | No (within consistent sets) |
| Measurement problem | Unsolved | Dissolved (no collapse) | Solved (definite positions) | Dissolved (no objective collapse) | Dissolved (no collapse) |
| Ontological cost | Vague | Many worlds | Non-locality | No objective reality | Multiple consistent sets |
| Empirically distinct? | No | No | No | No | No (GRW is distinct) |
⚠️ Common Misconception: "The interpretation question will never be settled because interpretations make the same predictions." This is too pessimistic. Some interpretations (or modifications, like GRW) do make slightly different predictions that could in principle be tested. Moreover, different interpretations suggest different research programs: Bohmian mechanics leads to trajectory-based simulation methods; many-worlds motivates certain approaches to quantum gravity; QBism connects to quantum information theory. Interpretations matter even when they are empirically equivalent, because they shape how physicists think about and extend the theory.
⚖️ Interpretation: This textbook does not endorse an interpretation. The honest truth is that the foundations of quantum mechanics remain an open problem. The formalism works — spectacularly — but what it tells us about the nature of reality is genuinely uncertain. Anyone who tells you the interpretation question is settled is either selling you something or has not thought about it carefully enough. The discomfort you feel is appropriate. The greatest physicists who ever lived felt it too.
24.11 Summary and Progressive Project
What We Have Learned
This chapter covered some of the most profound results in all of physics:
-
EPR (Section 24.1): If locality and realism both hold, quantum mechanics is incomplete. Einstein was right that local realism implies the existence of hidden variables.
-
Bell's theorem (Sections 24.3–24.4): But local hidden variable theories are bounded by the CHSH inequality $|S| \leq 2$, while quantum mechanics predicts $|S|_{\max} = 2\sqrt{2}$. Bell transformed a philosophical question into a quantitative, testable prediction.
-
Experiments (Section 24.5): Loophole-free tests confirm $|S| > 2$. Local realism is experimentally falsified. This earned the 2022 Nobel Prize.
-
Entanglement as resource (Sections 24.6–24.8): Entanglement enables teleportation (transmitting a qubit using 1 Bell pair + 2 classical bits) and superdense coding (transmitting 2 classical bits using 1 Bell pair + 1 qubit).
-
No-cloning (Section 24.9): Arbitrary quantum states cannot be copied. This is a theorem, not a technological limitation, and it underpins both the power and the constraints of quantum information.
-
Interpretations (Section 24.10): Quantum mechanics' predictions are not in dispute. What they mean about reality is. Copenhagen, many-worlds, Bohmian, QBism, and consistent histories each offer coherent but mutually exclusive pictures. The question remains open.
Threshold Concept Check
The threshold concept for this chapter is:
Entanglement is not classical correlation.
Classical correlations can always be explained by shared information (hidden variables) at the source. Entangled quantum states produce correlations that cannot be explained this way — Bell's theorem proves it, and experiments confirm it. This distinction is not a matter of degree; it is a qualitative, fundamental difference between classical and quantum physics.
If you have truly absorbed this, you will never again think of entanglement as "just correlated random variables." You will understand why EPR is not trivially resolved, why Bell's theorem is a genuine theorem with testable consequences, and why entanglement is a resource rather than just a property.
Progressive Project: Quantum Simulation Toolkit
Module: bell_test.py — Bell inequality tester, CHSH calculator, teleportation simulator.
Your project checkpoint for this chapter (see code/project-checkpoint.py) asks you to build three tools:
-
chsh_value(rho, a1, a2, b1, b2)— Computes the CHSH parameter $S$ for a given two-qubit density matrix $\rho$ and measurement settings $\hat{a}_1, \hat{a}_2, \hat{b}_1, \hat{b}_2$. -
lhv_simulation(n_trials, settings)— Simulates a local hidden variable model for $n$ trials and computes $S$, demonstrating $|S| \leq 2$. -
teleportation_protocol(chi, bell_pair)— Simulates the full quantum teleportation protocol, performing a Bell measurement and returning the teleported state.
🔗 Connection: The tools you build here will feed directly into the capstone Bell test simulator (Chapter 39). The CHSH calculator will be used to analyze simulated experimental data, and the teleportation protocol will be extended to a full quantum network simulation in Chapter 40.
Looking Ahead
Chapter 25 takes the entanglement resource perspective to its logical conclusion: quantum information and computation. We will build quantum circuits from gates (Hadamard, CNOT, phase), implement quantum algorithms (Deutsch-Jozsa, Grover, Shor), and confront the question of what quantum computers can do that classical computers cannot. Everything in Chapter 25 depends on the ideas developed here — Bell states, entanglement as resource, no-cloning, and the CHSH framework.
Chapter 28 returns to the measurement problem and the interpretation question in full depth, armed with the additional tools of decoherence theory (Chapter 33) and the Wigner's friend paradox.
"The paradox is only a conflict between reality and your feeling of what reality ought to be." — Richard Feynman
"Is it not good to know what follows from what, even if it is not necessary FAPP?" — John Stewart Bell
Key Equations Summary
| Equation | Name | Reference |
|---|---|---|
| $\|\Psi^-\rangle = \frac{1}{\sqrt{2}}(\|{\uparrow\downarrow}\rangle - \|{\downarrow\uparrow}\rangle)$ | Singlet (Bell) state | Sec. 24.1 |
| $E(\hat{a}, \hat{b}) = -\hat{a} \cdot \hat{b}$ | Singlet correlation function | Sec. 24.4 |
| $\|S_{\text{LHV}}\| \leq 2$ | CHSH inequality | Sec. 24.4 |
| $\|S_{\text{QM}}\|_{\max} = 2\sqrt{2}$ | Tsirelson bound | Sec. 24.4 |
| $E(|\psi\rangle) = S(\rho_A) = -\text{Tr}(\rho_A \log_2 \rho_A)$ | Entanglement entropy | Sec. 24.6 |
| $\langle\psi\|\phi\rangle = \langle\psi\|\phi\rangle^2 \Rightarrow$ no cloning | No-cloning theorem | Sec. 24.9 |
Related Reading
Explore this topic in other books
Quantum Mechanics Quantum Information Quantum Mechanics Quantum Technologies Pattern Recognition Information as Universal Currency