Case Study 2: The Threshold Theorem — Hope for Quantum Computing
The Central Question
Is large-scale quantum computing physically possible? This question, which seemed almost metaphysical in the early 1990s, was given a precise mathematical answer by the threshold theorem — the most important result in the theory of quantum computation after Shor's factoring algorithm.
The theorem says: yes, quantum computing is possible, provided the error rate per gate is below a critical threshold. This case study traces the history, proof structure, and practical implications of this remarkable result.
The Crisis of the 1990s
The Skeptics' Argument
After Shor's 1994 factoring algorithm electrified the physics and computer science communities, a vigorous debate erupted about whether quantum computers could ever be built. The skeptics — including some distinguished physicists — marshaled powerful arguments:
Argument 1 (Unruh, 1995): Decoherence times are far too short. A useful quantum computation requires maintaining coherence across thousands of qubits for millions of gate operations. With decoherence times of microseconds and gate times of nanoseconds, you can perform perhaps $10^3$ gates before the quantum state is destroyed. Shor's algorithm requires $10^{10}$ gates. The gap is seven orders of magnitude — and growing exponentially with problem size.
Argument 2 (Landauer, 1995): Error correction requires more resources than it saves. Correcting errors introduces new errors (through imperfect gates in the correction procedure). This creates a vicious cycle: you need error correction to correct errors, but error correction itself introduces errors that need correcting.
Argument 3 (philosophical): Quantum error correction fights decoherence — the process by which quantum states become classical. But decoherence is not a flaw; it is a feature of the quantum-to-classical transition. Fighting decoherence at scale means maintaining a macroscopic quantum superposition, which may be fundamentally impossible.
These were not fringe arguments. Rolf Landauer, a legendary figure in information theory, wrote explicitly: "Quantum computing: the dream may be doomed."
The Response: Fault Tolerance
The response came from multiple researchers nearly simultaneously. In 1996-1997, several groups — including Shor, Steane, Knill, Laflamme, and Zurek — proved versions of the threshold theorem. The essential answer to the skeptics was:
To Argument 1: Error correction can extend the effective coherence time indefinitely, at the cost of using more physical qubits.
To Argument 2: There exists a critical error rate below which the correction procedure improves the state faster than it introduces new errors. This is the threshold.
To Argument 3: The threshold theorem does not claim that maintaining macroscopic quantum coherence is easy. It claims that it is possible in principle, with polynomial (not exponential) resource overhead. Whether the engineering challenge can be met is a separate question from whether the physics allows it.
The Proof: Concatenation
The Key Idea
The proof of the threshold theorem relies on code concatenation — encoding each physical qubit of a quantum code inside another copy of the same code, recursively.
Consider a quantum code that uses $n$ physical qubits to encode 1 logical qubit, correcting single errors. If each physical gate has error probability $p$, and the error correction circuit has $c_{\text{loc}}$ "locations" (gate slots where an error can occur), then the probability that the code fails to correct is bounded by the probability of two or more errors occurring in the same correction round:
$$p_1 \leq \binom{c_{\text{loc}}}{2} p^2 \leq c \cdot p^2$$
where $c = c_{\text{loc}}(c_{\text{loc}} - 1)/2$ is a constant that depends on the specific code and circuit.
Level-by-Level Analysis
Level 0 (no encoding): Error rate $p_0 = p$.
Level 1: Each physical qubit is encoded in $n$ qubits with error correction. The effective error rate is: $$p_1 \leq c p^2 = \frac{(cp)^2}{c}$$
Level 2: Each of the $n$ physical qubits at level 1 is itself encoded in $n$ qubits: $$p_2 \leq c p_1^2 \leq c(cp^2)^2 = c^3 p^4 = \frac{(cp)^4}{c}$$
Level $k$: $$p_k \leq \frac{(cp)^{2^k}}{c}$$
If $cp < 1$ (i.e., $p < 1/c \equiv p_{\text{th}}$), then $p_k$ decreases doubly exponentially with $k$: $$p_k \leq \frac{1}{c}\left(cp\right)^{2^k} \to 0 \quad \text{as } k \to \infty$$
Numerical Example
Let $c = 10^4$ (a rough estimate for the number of fault locations in a single error correction round for a 7-qubit code). Then $p_{\text{th}} = 1/c = 10^{-4}$.
With $p = 10^{-4}$ (right at threshold): $cp = 1$, and concatenation does not help — $p_k = 1/c$ for all $k$.
With $p = 10^{-5}$ (one order of magnitude below threshold): $cp = 0.1$, and:
| Level $k$ | Physical qubits per logical | Logical error rate $p_k$ |
|---|---|---|
| 0 | 1 | $10^{-5}$ |
| 1 | 7 | $10^{-6}$ |
| 2 | 49 | $10^{-8}$ |
| 3 | 343 | $10^{-12}$ |
| 4 | 2401 | $10^{-20}$ |
| 5 | 16807 | $10^{-36}$ |
Each level of concatenation squares the effective error rate (in units of $p_{\text{th}}$) while multiplying the qubit count by 7. Five levels of concatenation reduce the error rate from $10^{-5}$ to $10^{-36}$ — using about 17,000 physical qubits per logical qubit.
Resource Scaling
To achieve a target logical error rate $\epsilon$, we need:
$$k = \left\lceil\log_2\left(\frac{\log(1/\epsilon)}{\log(1/(cp))}\right)\right\rceil$$
levels of concatenation, using:
$$n_{\text{phys}} = n^k = n^{O(\log\log(1/\epsilon))}$$
physical qubits per logical qubit. The key result: the overhead is polylogarithmic in $1/\epsilon$. This means that reducing the error rate by, say, a factor of $10^{6}$ requires only a modest increase in the number of physical qubits.
For a computation of length $L$ (number of logical gates), we need $\epsilon \ll 1/L$, so:
$$n_{\text{phys}} \sim \text{polylog}(L)$$
This is the content of the threshold theorem: the overhead scales polylogarithmically, not polynomially or exponentially, with the computation length.
Beyond Concatenation: The Surface Code
Why Concatenation Is Not the Final Answer
While concatenated codes provide a clean proof of the threshold theorem, they have practical drawbacks:
- Low threshold: Concatenated codes typically have thresholds $p_{\text{th}} \sim 10^{-5}$ to $10^{-4}$, requiring extremely high-fidelity gates.
- Non-local connections: Concatenation requires interactions between qubits that are not physically adjacent, which is difficult in most hardware architectures.
- Large overhead for modest improvements: The qubit count grows as $n^k$, which is fast even for moderate $k$.
The Surface Code: Higher Threshold, Local Gates
The surface code, introduced by Kitaev (1997) and developed by Bravyi, Kitaev, Dennis, Landahl, and others, overcomes these limitations:
- Threshold $\sim 1\%$: The surface code has the highest known threshold of any quantum error-correcting code, about two orders of magnitude higher than concatenated codes.
- Local connections only: Syndrome measurements involve only nearest-neighbor interactions on a 2D lattice — perfectly suited to planar chip architectures.
- Graceful scaling: The code distance $d$ can be increased smoothly by adding more physical qubits, with $\sim 2d^2$ physical qubits per logical qubit.
The logical error rate scales as:
$$p_L \sim \left(\frac{p}{p_{\text{th}}}\right)^{(d+1)/2}$$
For $p = 10^{-3}$ and $p_{\text{th}} = 10^{-2}$, each increase in $d$ by 2 reduces $p_L$ by roughly a factor of 10.
The Surface Code in Practice
The surface code is now the leading candidate for fault-tolerant quantum computing. Here is the current status:
2023 (Google): First demonstration of "below threshold" behavior — a distance-5 logical qubit outperforms a distance-3 logical qubit under repeated error correction. This was the first experimental evidence that increasing code size actually helps.
2024-2025 (Multiple groups): Demonstrations of distance-5 and distance-7 surface codes with multiple rounds of error correction. Logical qubit lifetimes exceeding physical qubit lifetimes. Preliminary logical gate operations.
Target milestones: - Distance 11–15: Sufficient for quantum advantage in quantum chemistry simulations ($\sim 10^3$–$10^4$ logical qubits) - Distance 25–30: Sufficient for cryptographically relevant quantum computing ($\sim 4{,}000$ logical qubits)
The Resource Reality
Factoring RSA-2048: A Detailed Estimate
Let us work through the resource requirements for the headline application: breaking RSA-2048 encryption.
Algorithm requirements: - Logical qubits: $\sim 4{,}000$ (for Shor's algorithm with windowed arithmetic) - Logical $T$ gates: $\sim 10^{9}$ (the dominant cost) - Logical Clifford gates: $\sim 10^{10}$ - Required logical error rate: $p_L < 10^{-15}$ (so that the total failure probability $\sim L \cdot p_L \ll 1$)
Surface code parameters (assuming $p = 10^{-3}$): - Required distance: $d \approx 27$ (to achieve $p_L \sim 10^{-15}$) - Physical qubits per logical data qubit: $2 \times 27^2 \approx 1{,}460$ - Logical data qubits: $\sim 4{,}000 \times 1{,}460 \approx 5.8 \times 10^6$
Magic state distillation: - Each logical $T$ gate requires $\sim 15$–$20$ ancilla qubits and multiple rounds of distillation - Magic state factories: $\sim 10^6$ additional physical qubits - Total with distillation overhead: $\sim 10^7$ physical qubits
Runtime: - Physical gate time: $\sim 1\,\mu$s (surface code cycle time) - Logical gate time: $\sim d \times 1\,\mu$s $\approx 27\,\mu$s - Total computation time: $\sim 10^{10} \times 27\,\mu$s $\approx 3$ days
So the bill for breaking RSA-2048 is: approximately 10 million physical qubits, running for about 3 days. Current quantum computers have about 1,000 qubits — a gap of 4 orders of magnitude.
Is the Gap Closeable?
The gap between current capabilities and the RSA-2048 target is large but not obviously permanent. Consider the historical analogy with classical computing:
| Year | Transistors per chip | Factor below target |
|---|---|---|
| 1971 (Intel 4004) | 2,300 | $\sim 10^6$ below modern CPUs |
| 1985 (Intel 386) | 275,000 | $\sim 10^4$ below |
| 2000 (Pentium 4) | 42,000,000 | $\sim 10^2$ below |
| 2020 (Apple M1) | 16,000,000,000 | (current) |
The transistor count grew by a factor of $\sim 10^7$ over 50 years. Quantum computing needs a factor of $\sim 10^4$ growth in qubit count.
Of course, adding qubits is harder than adding transistors (each qubit must be individually controllable, low-noise, and connected to its neighbors). But the threshold theorem guarantees that the physics does not prevent this growth — only engineering challenges remain.
Philosophical Implications
What the Threshold Theorem Really Says
At its core, the threshold theorem is a statement about the nature of quantum information:
Quantum information is not inherently fragile. It appears fragile because individual qubits decohere rapidly. But with enough resources (redundant encoding, continuous error correction), quantum information can be stored and processed for arbitrarily long times. The apparent fragility of quantum states is a practical challenge, not a fundamental limitation.
This has implications beyond computing:
-
The quantum-to-classical transition is not inevitable. Decoherence pushes quantum systems toward classical behavior, but error correction pushes back. In principle, a sufficiently well-engineered system can maintain quantum coherence indefinitely.
-
The universe may be fault-tolerant. Some physicists have speculated that the laws of physics themselves are a form of quantum error correction — that the stability of the physical world (atoms don't decohere into nothingness) reflects an underlying error-correcting structure.
-
Information is physical. The threshold theorem is a statement about information (error rates, coding theory) that has direct implications for physics (what computations are possible in the physical universe). This supports the growing view that information theory and physics are deeply intertwined.
What It Does Not Say
The threshold theorem does not guarantee that useful quantum computers will be built. It guarantees that they can be built, in principle, if certain conditions are met. The conditions — physical error rates below threshold, scalable qubit architectures, efficient classical control — are engineering challenges of formidable difficulty. The theorem provides the blueprint; building the machine is another matter.
A Timeline of Progress
| Year | Milestone |
|---|---|
| 1994 | Shor's factoring algorithm |
| 1995 | Shor's 9-qubit error-correcting code |
| 1996 | Steane's 7-qubit code; CSS code construction |
| 1996 | Threshold theorem (Knill, Laflamme, Zurek; Aharonov, Ben-Or) |
| 1997 | Kitaev's toric code (precursor to surface code) |
| 1998 | First experimental QEC (NMR, 3-qubit code) |
| 2001 | Surface code threshold calculated ($\sim 1\%$) |
| 2004 | Ion trap QEC demonstration (NIST) |
| 2011 | Repetition code demonstrated in superconducting qubits |
| 2015 | Logical qubit in trapped ions (Monroe group) |
| 2021 | Repeated QEC cycles in superconducting qubits (Google, IBM) |
| 2023 | First "below threshold" surface code demonstration (Google) |
| 2024-25 | Logical qubit lifetimes exceeding physical lifetimes (multiple groups) |
| Target: 2030s | Cryptographically relevant quantum computing (?) |
Questions for Reflection
-
The threshold theorem requires the physical error rate to be below $p_{\text{th}}$. But the theorem assumes a specific noise model (independent, identically distributed errors). Real noise is correlated. How do correlated errors affect the threshold, and what is being done about this?
-
The surface code has a high threshold ($\sim 1\%$) but a low code rate ($1/(2d^2)$). Are there codes with both high thresholds and high code rates? What are the tradeoffs?
-
Landauer argued in 1995 that quantum error correction would never be practical because the overhead is too large. Thirty years later, we have experimental demonstrations but not yet useful computations. Was Landauer right about the difficulty, even if he was wrong about the impossibility?
-
The threshold theorem implies that quantum information can be protected indefinitely. Does this have implications for the black hole information paradox — the question of whether information falling into a black hole is truly lost?