42 min read

How quantum systems jump between energy levels — and how understanding these jumps gave us the laser, explained why the sky is blue, and revealed the deep structure of light-matter interaction

Chapter 21: Time-Dependent Perturbation Theory: Transitions and Radiation

How quantum systems jump between energy levels — and how understanding these jumps gave us the laser, explained why the sky is blue, and revealed the deep structure of light-matter interaction


Opening: The Universe Is Not Static

In Chapters 17 through 20, we developed a powerful arsenal of approximation techniques — time-independent perturbation theory, degenerate perturbation theory, the variational method, and WKB — but they all shared a common limitation: they described quantum systems in equilibrium. We found energy levels and stationary states — the notes on the page. We never heard the music.

Yet the real universe hums with transitions. Atoms absorb photons and jump to excited states. Excited nuclei decay. Electrons in molecules redistribute themselves when struck by light. Every color you see, every signal in a radio antenna, every photon detected by a telescope is testimony to a quantum system making a transition from one energy level to another.

The question is: how do we calculate the rate at which these transitions occur?

You already possess many of the ingredients. In Chapter 7, you learned the time-evolution operator $\hat{U}(t)$ and saw that the interaction picture splits dynamics into a "free" part and a "perturbation" part. In Chapter 17, you mastered the art of expanding in powers of a small parameter. Now we combine these ideas into one of the most useful tools in all of quantum mechanics: time-dependent perturbation theory.

The results of this chapter are not merely academic. Fermi's golden rule — perhaps the single most frequently invoked formula in all of physics after $E = mc^2$ and $F = ma$ — governs transition rates in atoms, nuclei, solids, and elementary particles alike. The Einstein $A$ and $B$ coefficients connect spontaneous and stimulated emission, providing the theoretical foundation for the laser. Selection rules, derived from the matrix elements we compute here, explain why some atomic transitions are strong and others vanishingly weak.

By the end of this chapter, you will be able to start from a time-dependent perturbation $\hat{V}(t)$, calculate the probability that a quantum system transitions from state $|i\rangle$ to state $|f\rangle$, and understand both the conceptual meaning and the practical power of that calculation.

Let us begin.


21.1 Time-Dependent Problems and the Interaction Picture

The Setup

Consider a quantum system whose Hamiltonian has two parts:

$$\hat{H}(t) = \hat{H}_0 + \hat{V}(t)$$

Here $\hat{H}_0$ is the unperturbed Hamiltonian, whose eigenstates and eigenvalues we know exactly:

$$\hat{H}_0 |n\rangle = E_n |n\rangle$$

The perturbation $\hat{V}(t)$ is "turned on" at some time — perhaps suddenly, perhaps gradually — and drives transitions between the eigenstates of $\hat{H}_0$. Crucially, $\hat{V}(t)$ is assumed small compared to $\hat{H}_0$ in a sense we will make precise.

Physical examples are everywhere:

  • An atom in an oscillating electromagnetic field ($\hat{V}(t) = -e\hat{\vec{r}} \cdot \vec{E}_0 \cos\omega t$)
  • A nucleus subjected to a time-varying magnetic field (NMR/MRI)
  • A quantum dot irradiated by a laser pulse
  • A particle scattered by a transient potential

Three Pictures of Quantum Mechanics

Recall from Chapter 7 that quantum mechanics can be formulated in three equivalent pictures, all giving the same measurable predictions:

Picture States Operators Generator of evolution
Schrödinger Evolve via full $\hat{H}$ Fixed $\hat{H} = \hat{H}_0 + \hat{V}$
Heisenberg Fixed Evolve via full $\hat{H}$ $\hat{H} = \hat{H}_0 + \hat{V}$
Interaction Evolve via $\hat{V}_I$ only Evolve via $\hat{H}_0$ Split: $\hat{H}_0$ moves operators, $\hat{V}_I$ moves states

For time-dependent perturbation theory, the interaction picture (also called the Dirac picture) is the natural choice. It strips away the "trivial" time dependence due to $\hat{H}_0$ and isolates the interesting dynamics caused by $\hat{V}(t)$.

Formal Construction

In the Schrödinger picture, the state evolves as:

$$i\hbar \frac{d}{dt}|\psi_S(t)\rangle = (\hat{H}_0 + \hat{V}(t))|\psi_S(t)\rangle$$

Define the interaction-picture state by stripping off the free evolution:

$$|\psi_I(t)\rangle = e^{i\hat{H}_0 t/\hbar} |\psi_S(t)\rangle$$

Substitute into the Schrödinger equation. After straightforward algebra (using the product rule on $d/dt[e^{i\hat{H}_0 t/\hbar}|\psi_S\rangle]$), we obtain:

$$\boxed{i\hbar \frac{d}{dt}|\psi_I(t)\rangle = \hat{V}_I(t)|\psi_I(t)\rangle}$$

where the interaction-picture perturbation is:

$$\hat{V}_I(t) = e^{i\hat{H}_0 t/\hbar}\, \hat{V}(t)\, e^{-i\hat{H}_0 t/\hbar}$$

This is the payoff: in the interaction picture, the state evolves due to $\hat{V}_I$ alone. If $\hat{V} = 0$, the interaction-picture state is constant — all the time dependence is carried by the operators (which rotate at their natural frequencies $\omega_{mn} = (E_m - E_n)/\hbar$).

💡 Key Insight: The interaction picture is the ideal framework for perturbation theory because the "smallness" of $\hat{V}$ directly controls the rate of change of $|\psi_I(t)\rangle$. If $\hat{V}$ is truly small, $|\psi_I(t)\rangle$ barely changes — exactly the condition for a perturbative expansion.

Matrix Elements in the Interaction Picture

Expand the interaction-picture state in the eigenbasis of $\hat{H}_0$:

$$|\psi_I(t)\rangle = \sum_n c_n(t) |n\rangle$$

The expansion coefficients $c_n(t)$ represent the probability amplitudes for finding the system in state $|n\rangle$ at time $t$. The probability of being in state $|n\rangle$ is $|c_n(t)|^2$.

Substituting into the interaction-picture equation of motion:

$$i\hbar \dot{c}_f(t) = \sum_n \langle f|\hat{V}_I(t)|n\rangle \, c_n(t)$$

Using $\hat{V}_I(t) = e^{i\hat{H}_0 t/\hbar}\hat{V}(t)e^{-i\hat{H}_0 t/\hbar}$ and $\hat{H}_0|n\rangle = E_n|n\rangle$:

$$\langle f|\hat{V}_I(t)|n\rangle = e^{i(E_f - E_n)t/\hbar}\langle f|\hat{V}(t)|n\rangle = e^{i\omega_{fn}t} V_{fn}(t)$$

where $\omega_{fn} = (E_f - E_n)/\hbar$ is the Bohr frequency between states $|f\rangle$ and $|n\rangle$, and $V_{fn}(t) = \langle f|\hat{V}(t)|n\rangle$ is the matrix element of the perturbation. The exact equation of motion is:

$$\boxed{i\hbar \dot{c}_f(t) = \sum_n V_{fn}(t)\, e^{i\omega_{fn}t}\, c_n(t)}$$

This is a set of coupled first-order ODEs — exact but, in general, impossible to solve analytically. We need approximations.

Checkpoint: Before proceeding, verify that you understand why the interaction picture is preferable to the Schrödinger picture for perturbation theory. In the Schrödinger picture, the coefficients $c_n(t)$ oscillate rapidly even when $\hat{V} = 0$ (because each acquires a phase $e^{-iE_n t/\hbar}$). In the interaction picture, these phases have been absorbed into the operators, so $c_n(t)$ changes only when the perturbation acts.

The Dyson Series

The formal solution to the interaction-picture equation of motion is obtained by iteration. Integrating $i\hbar\,d|\psi_I\rangle/dt = \hat{V}_I(t)|\psi_I\rangle$:

$$|\psi_I(t)\rangle = |\psi_I(0)\rangle + \left(-\frac{i}{\hbar}\right)\int_0^t \hat{V}_I(t_1)|\psi_I(t_1)\rangle\,dt_1$$

Substituting this expression back into itself repeatedly generates the Dyson series:

$$|\psi_I(t)\rangle = \left[\hat{I} + \left(-\frac{i}{\hbar}\right)\int_0^t dt_1\,\hat{V}_I(t_1) + \left(-\frac{i}{\hbar}\right)^2\int_0^t dt_1\int_0^{t_1} dt_2\,\hat{V}_I(t_1)\hat{V}_I(t_2) + \cdots\right]|\psi_I(0)\rangle$$

This can be written compactly as $|\psi_I(t)\rangle = \hat{U}_I(t, 0)|\psi_I(0)\rangle$ where

$$\hat{U}_I(t, 0) = \mathcal{T}\exp\left(-\frac{i}{\hbar}\int_0^t \hat{V}_I(t')\,dt'\right)$$

is the time-ordered exponential encountered in Chapter 7. The Dyson series is the perturbative expansion of this time-ordered exponential: each successive term involves one more power of $\hat{V}_I$ and one more time integration. In practice, the first-order term is usually sufficient — which is what we compute next.

🔵 Historical Note: Freeman Dyson derived this series in 1949 while developing the mathematical foundations of quantum electrodynamics, connecting Feynman's diagrammatic approach with Schwinger's and Tomonaga's operator methods. The Dyson series remains the starting point for essentially all perturbative calculations in quantum field theory.

Worked Example: Interaction Picture for a Spin in a Rotating Field

To make the interaction picture concrete, consider a spin-1/2 particle in a constant magnetic field $B_0\hat{z}$ with a small rotating transverse field $B_1(\cos\omega t\,\hat{x} + \sin\omega t\,\hat{y})$. This is precisely the setup of NMR (Chapter 13 connection).

The Hamiltonian is $\hat{H} = \hat{H}_0 + \hat{V}(t)$ with:

$$\hat{H}_0 = -\frac{\gamma\hbar B_0}{2}\hat{\sigma}_z = \frac{\hbar\omega_0}{2}\hat{\sigma}_z, \qquad \hat{V}(t) = -\frac{\gamma\hbar B_1}{2}(\cos\omega t\,\hat{\sigma}_x + \sin\omega t\,\hat{\sigma}_y)$$

where $\omega_0 = -\gamma B_0$ is the Larmor frequency. In the interaction picture:

$$\hat{V}_I(t) = e^{i\hat{H}_0 t/\hbar}\hat{V}(t)e^{-i\hat{H}_0 t/\hbar} = e^{i\omega_0\hat{\sigma}_z t/2}\hat{V}(t)e^{-i\omega_0\hat{\sigma}_z t/2}$$

Using the rotation properties of Pauli matrices ($e^{i\alpha\hat{\sigma}_z/2}\hat{\sigma}_\pm e^{-i\alpha\hat{\sigma}_z/2} = e^{\pm i\alpha}\hat{\sigma}_\pm$), the exponentials combine with the time dependence in $\hat{V}(t)$ to give:

$$\hat{V}_I(t) = -\frac{\gamma\hbar B_1}{2}\left[e^{i(\omega_0-\omega)t}\hat{\sigma}_+ + e^{-i(\omega_0-\omega)t}\hat{\sigma}_-\right]$$

At resonance ($\omega = \omega_0$), the interaction-picture perturbation becomes time-independent: $\hat{V}_I = -(\gamma\hbar B_1/2)(\hat{\sigma}_+ + \hat{\sigma}_-)$. This is why the interaction picture is so powerful — it has transformed a time-dependent problem into a time-independent one. The resulting dynamics are the Rabi oscillations you studied in Chapter 7.


21.2 First-Order Transition Probability

The Perturbative Expansion

We now exploit the smallness of $\hat{V}$ systematically. Write:

$$c_n(t) = c_n^{(0)} + c_n^{(1)} + c_n^{(2)} + \cdots$$

where the superscript denotes the order in $\hat{V}$. The initial condition is that the system starts in eigenstate $|i\rangle$ at $t = 0$:

$$c_n^{(0)} = \delta_{ni}$$

Zeroth order: $c_f^{(0)} = \delta_{fi}$. The system remains in its initial state.

First order: Substitute $c_n^{(0)}$ on the right-hand side:

$$i\hbar \dot{c}_f^{(1)}(t) = V_{fi}(t)\, e^{i\omega_{fi}t}$$

Integrate from $0$ to $t$:

$$\boxed{c_f^{(1)}(t) = -\frac{i}{\hbar}\int_0^t V_{fi}(t')\, e^{i\omega_{fi}t'}\, dt'}$$

This is the first-order transition amplitude from state $|i\rangle$ to state $|f\rangle$. The transition probability is:

$$P_{i \to f}(t) = |c_f^{(1)}(t)|^2 = \frac{1}{\hbar^2}\left|\int_0^t V_{fi}(t')\, e^{i\omega_{fi}t'}\, dt'\right|^2$$

Constant Perturbation Turned On at $t = 0$

The simplest case: $\hat{V}(t) = \hat{V}\theta(t)$, where $\theta(t)$ is the Heaviside step function and $\hat{V}$ is time-independent. Then:

$$c_f^{(1)}(t) = -\frac{i}{\hbar}V_{fi}\int_0^t e^{i\omega_{fi}t'}\, dt' = -\frac{V_{fi}}{\hbar\omega_{fi}}\left(e^{i\omega_{fi}t} - 1\right)$$

The transition probability is:

$$P_{i \to f}(t) = \frac{|V_{fi}|^2}{\hbar^2\omega_{fi}^2}\left|e^{i\omega_{fi}t} - 1\right|^2 = \frac{4|V_{fi}|^2}{\hbar^2\omega_{fi}^2}\sin^2\!\left(\frac{\omega_{fi}t}{2}\right)$$

Using the identity $|e^{i\theta} - 1|^2 = 4\sin^2(\theta/2)$, we can write this more elegantly:

$$\boxed{P_{i \to f}(t) = \frac{|V_{fi}|^2}{\hbar^2}\,\frac{\sin^2(\omega_{fi}t/2)}{(\omega_{fi}/2)^2}}$$

This result deserves careful analysis:

1. Resonance structure. The function $\sin^2(\omega_{fi}t/2)/(\omega_{fi}/2)^2$ is sharply peaked around $\omega_{fi} = 0$, i.e., when $E_f \approx E_i$. The transition probability is largest when the initial and final states have nearly the same energy. This is not yet energy conservation (that requires the continuum limit) but it is the precursor.

2. Time dependence. At exact resonance ($\omega_{fi} = 0$, meaning $E_f = E_i$), the probability grows as $t^2$:

$$P_{i \to f}(t)\big|_{\omega_{fi}=0} = \frac{|V_{fi}|^2 t^2}{\hbar^2}$$

This quadratic growth is an artifact of first-order perturbation theory — it eventually violates the requirement $P \leq 1$ — but it is valid for short times $t \ll \hbar/|V_{fi}|$.

3. Width of the peak. The central peak of $\sin^2(\omega_{fi}t/2)/(\omega_{fi}/2)^2$ has a width $\Delta\omega_{fi} \sim 2\pi/t$, or equivalently $\Delta E \sim 2\pi\hbar/t$. As time increases, the "energy window" narrows: the perturbation becomes more and more selective, coupling $|i\rangle$ only to states very close in energy.

📊 By the Numbers: This energy-time narrowing is the mathematical origin of the energy-time uncertainty relation $\Delta E \cdot \Delta t \gtrsim \hbar$. A brief perturbation ($\Delta t$ small) can induce transitions across a wide energy range ($\Delta E$ large). A long-duration perturbation ($\Delta t$ large) produces transitions only to states with $E_f \approx E_i$ ($\Delta E$ small).

Harmonic (Sinusoidal) Perturbation

Now consider the physically crucial case of a harmonic perturbation:

$$\hat{V}(t) = \hat{V}_0 e^{-i\omega t} + \hat{V}_0^\dagger e^{+i\omega t}$$

This describes, for instance, an atom driven by monochromatic light of angular frequency $\omega$. The first-order amplitude becomes:

$$c_f^{(1)}(t) = -\frac{i}{\hbar}\int_0^t \left[(V_0)_{fi}\, e^{i(\omega_{fi} - \omega)t'} + (V_0^\dagger)_{fi}\, e^{i(\omega_{fi} + \omega)t'}\right] dt'$$

Performing the integrations:

$$c_f^{(1)}(t) = -\frac{(V_0)_{fi}}{\hbar}\,\frac{e^{i(\omega_{fi}-\omega)t} - 1}{\omega_{fi} - \omega} - \frac{(V_0^\dagger)_{fi}}{\hbar}\,\frac{e^{i(\omega_{fi}+\omega)t} - 1}{\omega_{fi} + \omega}$$

Each term is large (resonant) under different conditions:

  • First term: resonant when $\omega_{fi} \approx \omega$, i.e., $E_f - E_i \approx \hbar\omega$. This corresponds to absorption: the system gains energy $\hbar\omega$ from the driving field.

  • Second term: resonant when $\omega_{fi} \approx -\omega$, i.e., $E_i - E_f \approx \hbar\omega$. This corresponds to stimulated emission: the system loses energy $\hbar\omega$ to the field.

Near resonance, only one term dominates (the other is "off-resonant" by $\sim 2\omega$). This is the rotating wave approximation (RWA), already encountered in the Rabi problem of Chapter 7. For absorption ($E_f > E_i$, $\omega \approx \omega_{fi}$):

$$P_{i \to f}^{\text{abs}}(t) \approx \frac{|(V_0)_{fi}|^2}{\hbar^2}\,\frac{\sin^2[(\omega_{fi} - \omega)t/2]}{[(\omega_{fi} - \omega)/2]^2}$$

This has the same $\sin^2$ structure as the constant perturbation, but now the peak is at $\omega = \omega_{fi}$ — the driving frequency matching the transition frequency. This is resonance absorption.

🔗 Connection: Compare with the Rabi formula from Chapter 7: $P_{1\to 2}(t) = (\Omega_R^2/\Omega^2)\sin^2(\Omega t/2)$. For weak fields, the Rabi formula reduces to our first-order result here. Time-dependent perturbation theory is the weak-field limit of the exact Rabi solution.

⚠️ Common Misconception: Students sometimes believe that time-dependent perturbation theory and the Rabi model are different physical theories. They are not. The Rabi model is the exact solution for a two-level system. Perturbation theory gives a systematic approximation for any number of levels, valid when the driving field is weak ($\Omega_R \ll \omega_{fi}$). In the weak-field limit, they agree exactly.

Worked Example: Transition Probability for a Gaussian Pulse

A pulse of radiation with a Gaussian temporal envelope provides a physically important example that bridges the ideal (infinite-duration monochromatic) and realistic (finite-duration) cases. Let:

$$V_{fi}(t) = V_0\, e^{-t^2/(2\tau^2)}\cos\omega t$$

where $\tau$ is the pulse duration and $\omega$ is the carrier frequency. The first-order amplitude (keeping only the near-resonant term) is:

$$c_f^{(1)}(t \to \infty) = -\frac{iV_0}{2\hbar}\int_{-\infty}^{\infty} e^{-t'^2/(2\tau^2)} e^{i(\omega_{fi}-\omega)t'}\,dt' = -\frac{iV_0\tau\sqrt{2\pi}}{2\hbar}\,\exp\left(-\frac{(\omega_{fi}-\omega)^2\tau^2}{2}\right)$$

The transition probability is:

$$P_{i \to f} = \frac{\pi V_0^2\tau^2}{2\hbar^2}\,\exp\left(-(\omega_{fi}-\omega)^2\tau^2\right)$$

This is a Gaussian in the detuning $\delta = \omega_{fi} - \omega$, with width $\Delta\omega \sim 1/\tau$. Two important limits:

  • Long pulse ($\tau \to \infty$): The Gaussian becomes a delta function $\sqrt{\pi}\tau\,\delta(\omega_{fi} - \omega)$, recovering exact energy conservation and the Fermi golden rule limit.

  • Short pulse ($\tau \to 0$ with $V_0\tau$ fixed): The Gaussian becomes broad — the pulse can drive transitions across a wide range of $\omega_{fi}$. This is the bandwidth-limited pulse familiar from ultrafast laser physics.

The product $\Delta\omega \cdot \tau \sim 1$ is a manifestation of the energy-time uncertainty relation. You cannot have both precise energy selection and a short interaction time. This trade-off governs the design of spectroscopy experiments: narrow-linewidth lasers (long coherence time $\tau$) give high spectral resolution ($\Delta\omega$ small), while ultrafast pulses (short $\tau$) sacrifice resolution for temporal resolution.


21.3 Fermi's Golden Rule

From Discrete to Continuum

The first-order result for a constant perturbation gave us $P_{i\to f}(t) \propto \sin^2(\omega_{fi}t/2)/(\omega_{fi}/2)^2$. For a single discrete final state, this oscillates — the transition probability goes up and down, never settling to a steady value. Real-world transitions, however, exhibit an irreversible decay from the initial state. What is missing?

The answer is the density of final states. In realistic situations, the final state is not a single discrete level but a member of a continuum (or near-continuum) of states: photon emission produces photons of slightly different directions and polarizations; nuclear decay products carry off various amounts of kinetic energy; and so on.

When we sum over a continuum of final states, the oscillating $\sin^2$ functions conspire to produce a constant transition rate.

Derivation

Suppose the final states form a quasi-continuum characterized by a density of states $\rho(E_f)$, defined so that $\rho(E_f)\,dE_f$ is the number of final states with energies between $E_f$ and $E_f + dE_f$. The total transition probability to all final states near energy $E_f \approx E_i$ is:

$$P_{i \to \text{cont}}(t) = \int P_{i \to f}(t)\, \rho(E_f)\, dE_f = \frac{1}{\hbar^2}\int |V_{fi}|^2 \frac{\sin^2(\omega_{fi}t/2)}{(\omega_{fi}/2)^2}\, \rho(E_f)\, dE_f$$

Change variables to $\omega = \omega_{fi} = (E_f - E_i)/\hbar$:

$$P_{i \to \text{cont}}(t) = \frac{1}{\hbar}\int |V_{fi}|^2 \frac{\sin^2(\omega t/2)}{(\omega/2)^2}\, \rho(E_i + \hbar\omega)\, d\omega$$

Now we invoke the key mathematical fact. For large $t$, the function

$$\frac{\sin^2(\omega t/2)}{(\omega/2)^2} \xrightarrow{t \to \infty} 2\pi t\, \delta(\omega)$$

This is a nascent delta function: as $t$ grows, the peak at $\omega = 0$ becomes taller and narrower, with the area under the curve always equal to $2\pi t$. (You can verify the area: $\int_{-\infty}^{\infty}\sin^2(x)/x^2\,dx = \pi$.)

Assuming that $|V_{fi}|^2$ and $\rho(E_f)$ vary slowly over the narrow energy window selected by the $\sin^2$ peak (width $\sim 2\pi\hbar/t$), we can evaluate them at $\omega = 0$ (i.e., $E_f = E_i$) and pull them out of the integral:

$$P_{i \to \text{cont}}(t) \approx \frac{|V_{fi}|^2}{\hbar}\, \rho(E_i)\, 2\pi t = \frac{2\pi}{\hbar}\, |V_{fi}|^2\, \rho(E_f)\bigg|_{E_f = E_i} \cdot t$$

The transition probability grows linearly in time, which means the transition rate $\Gamma = dP/dt$ is constant:

$$\boxed{\Gamma_{i \to f} = \frac{2\pi}{\hbar}\, |\langle f|\hat{V}|i\rangle|^2\, \rho(E_f)}$$

This is Fermi's golden rule. Despite its name, it was first derived by Dirac in 1927; Fermi named it "golden rule number two" in his 1950 nuclear physics lectures at the University of Chicago, and the name stuck.

🔵 Historical Note: Enrico Fermi's 1950 lecture notes at Chicago introduced the phrase "Golden Rule No. 2" with characteristic pragmatism: he called it "golden" because of its extraordinary usefulness, not its mathematical depth. Dirac had derived the essential result in 1927 using what he called the "theory of transitions" — the same framework we are developing here. The rule is sometimes called the "Dirac-Fermi golden rule" in recognition of this history.

Conditions of Validity

Fermi's golden rule requires several conditions to hold simultaneously:

  1. First-order perturbation theory is valid: $|V_{fi}| \ll \hbar/t_{\text{relevant}}$, or equivalently, $P_{i \to f} \ll 1$ for any individual final state.

  2. The continuum approximation holds: The density of states must be smooth over the energy range $\Delta E \sim 2\pi\hbar/t$. If the level spacing is $\delta E$, this requires $t \gg \hbar/\delta E$.

  3. Long enough time: The $\sin^2$ function must have had time to narrow into a nascent delta function. Combined with condition 2, this gives a "Goldilocks" window:

$$\frac{\hbar}{\delta E} \ll t \ll \frac{\hbar}{|V_{fi}|}$$

The time must be long enough for the energy selectivity to emerge, but short enough that perturbation theory has not broken down.

🔴 Warning: Fermi's golden rule gives a constant rate, implying exponential decay of the initial state: $P_i(t) = e^{-\Gamma t}$. But first-order perturbation theory cannot derive exponential decay — it gives $P_i(t) \approx 1 - \Gamma t$, which is just the first term in the Taylor expansion. The full exponential requires summing all orders, or equivalently, using the Wigner-Weisskopf approximation (covered in advanced courses). Fermi's golden rule gives the correct rate, but the correct time dependence requires going beyond first order.

Worked Example: Decay Rate of a Particle in a Box into a Continuum

Consider a particle initially in the ground state of a one-dimensional box of width $a$, suddenly coupled to a semi-infinite region (a continuum of free-particle states) by a matrix element $V_{fi} = V_0$ (constant, independent of $f$ for simplicity). This models, for example, an electron in a quantum dot that tunnels into a metallic electrode.

The density of states for free particles in one dimension (box normalization, length $L$) is:

$$\rho(E) = \frac{L}{2\pi\hbar}\sqrt{\frac{m}{2E}}$$

For the ground state energy $E_1 = \pi^2\hbar^2/(2ma^2)$, and taking $L = 1$ cm, $a = 1$ nm, $m = m_e$:

$$\rho(E_1) = \frac{L}{2\pi\hbar}\sqrt{\frac{m}{2E_1}} = \frac{L}{2\pi\hbar}\cdot\frac{ma}{\pi\hbar} = \frac{Lma}{2\pi^2\hbar^2}$$

Plugging in numbers:

$$\rho(E_1) \approx \frac{(10^{-2})(9.1\times10^{-31})(10^{-9})}{2\pi^2(1.05\times10^{-34})^2} \approx 4.2\times10^{40}\,\text{states/J}$$

With $V_0 \sim 10^{-22}$ J ($\sim 0.6$ meV), Fermi's golden rule gives:

$$\Gamma = \frac{2\pi}{\hbar}\cdot(10^{-22})^2\cdot 4.2\times10^{40} \approx 5.0\times10^{11}\,\text{s}^{-1}$$

This corresponds to a tunneling lifetime $\tau = 1/\Gamma \approx 2$ ps — consistent with known tunneling times in quantum dot devices. The calculation illustrates all three ingredients of the golden rule: the matrix element sets the coupling, the density of states counts the available final states, and the factor $2\pi/\hbar$ does the dimensional bookkeeping.

Fermi's Golden Rule for a Harmonic Perturbation

For a harmonic perturbation $\hat{V}(t) = \hat{V}_0 e^{-i\omega t} + \hat{V}_0^\dagger e^{+i\omega t}$, the same analysis applies with $\omega_{fi}$ replaced by $\omega_{fi} - \omega$ (for absorption) or $\omega_{fi} + \omega$ (for emission). The golden rule becomes:

$$\Gamma_{i \to f}^{\text{abs}} = \frac{2\pi}{\hbar}\, |(V_0)_{fi}|^2\, \rho(E_f)\bigg|_{E_f = E_i + \hbar\omega}$$

$$\Gamma_{i \to f}^{\text{em}} = \frac{2\pi}{\hbar}\, |(V_0^\dagger)_{fi}|^2\, \rho(E_f)\bigg|_{E_f = E_i - \hbar\omega}$$

Energy conservation is now explicit: The delta function in the derivation enforces $E_f = E_i + \hbar\omega$ (absorption) or $E_f = E_i - \hbar\omega$ (emission). The system absorbs or emits one quantum of energy $\hbar\omega$ from the driving field.

Checkpoint: Make sure you can explain why Fermi's golden rule produces a constant rate even though the underlying first-order probability oscillates. The key is the sum over a continuum: the oscillations of different final states destructively interfere, leaving only the linearly growing total probability. This is conceptually similar to the dephasing that causes irreversible decoherence (Chapter 33).


21.4 Electromagnetic Radiation as a Perturbation

The Minimal Coupling Hamiltonian

To apply time-dependent perturbation theory to the interaction of atoms with light, we need the correct form of $\hat{V}(t)$. An electromagnetic wave is described (in the Coulomb gauge) by a vector potential:

$$\vec{A}(\vec{r}, t) = A_0\, \hat{\epsilon}\, \left(e^{i(\vec{k}\cdot\vec{r} - \omega t)} + e^{-i(\vec{k}\cdot\vec{r} - \omega t)}\right)$$

where $\hat{\epsilon}$ is the polarization vector (perpendicular to the wave vector $\vec{k}$) and $A_0$ is the amplitude. The electric and magnetic fields are:

$$\vec{E} = -\frac{\partial \vec{A}}{\partial t}, \qquad \vec{B} = \nabla \times \vec{A}$$

The Hamiltonian for a charged particle (charge $q$, mass $m$) in an electromagnetic field is obtained by the minimal coupling prescription $\hat{\vec{p}} \to \hat{\vec{p}} - q\vec{A}/c$:

$$\hat{H} = \frac{1}{2m}\left(\hat{\vec{p}} - \frac{q}{c}\vec{A}(\hat{\vec{r}}, t)\right)^2 + V(\hat{\vec{r}})$$

Expanding:

$$\hat{H} = \underbrace{\frac{\hat{p}^2}{2m} + V(\hat{\vec{r}})}_{\hat{H}_0} \underbrace{- \frac{q}{2mc}\left(\hat{\vec{p}}\cdot\vec{A} + \vec{A}\cdot\hat{\vec{p}}\right)}_{\hat{V}^{(1)}} + \underbrace{\frac{q^2}{2mc^2}\vec{A}\cdot\vec{A}}_{\hat{V}^{(2)}}$$

In the Coulomb gauge, $\nabla\cdot\vec{A} = 0$, which means $\hat{\vec{p}}\cdot\vec{A} = \vec{A}\cdot\hat{\vec{p}}$ (the operators commute when $\vec{A}$ is transverse). The first-order perturbation simplifies to:

$$\hat{V}^{(1)} = -\frac{q}{mc}\,\vec{A}\cdot\hat{\vec{p}}$$

The $\vec{A}\cdot\vec{A}$ term ($\hat{V}^{(2)}$) is quadratic in the field amplitude and can be neglected for weak fields (it contributes to two-photon processes like Raman scattering and nonlinear optics). For our purposes:

$$\hat{V}(t) = -\frac{q}{mc}\,\vec{A}(\hat{\vec{r}}, t)\cdot\hat{\vec{p}}$$

The Electric Dipole Approximation

For optical transitions in atoms, the wavelength of light ($\lambda \sim 500$ nm) is much larger than the size of the atom ($a_0 \sim 0.05$ nm). The ratio $a_0/\lambda \sim 10^{-4}$ is tiny, which means the electromagnetic field is essentially uniform over the extent of the atom.

Mathematically, in the exponential $e^{i\vec{k}\cdot\vec{r}}$, we have $|\vec{k}\cdot\vec{r}| \sim |\vec{k}| \cdot a_0 = (2\pi/\lambda) \cdot a_0 \sim 10^{-3} \ll 1$. We can therefore expand:

$$e^{i\vec{k}\cdot\vec{r}} \approx 1 + i\vec{k}\cdot\vec{r} + \cdots \approx 1$$

Keeping only the leading term (the electric dipole approximation, or E1), the perturbation becomes:

$$\hat{V}(t) \approx -\frac{qA_0}{mc}\,\hat{\epsilon}\cdot\hat{\vec{p}}\,\left(e^{-i\omega t} + e^{+i\omega t}\right)$$

Using the identity (derivable from $[\hat{x}_j, \hat{H}_0] = i\hbar\hat{p}_j/m$):

$$\langle f|\hat{\vec{p}}|i\rangle = \frac{im\omega_{fi}}{\hbar}\,\langle f|\hat{\vec{r}}|i\rangle \cdot \hbar = im\omega_{fi}\langle f|\hat{\vec{r}}|i\rangle$$

we can rewrite the matrix element as:

$$\langle f|\hat{V}|i\rangle \propto \omega_{fi}\,\hat{\epsilon}\cdot\langle f|q\hat{\vec{r}}|i\rangle = \omega_{fi}\,\hat{\epsilon}\cdot\vec{d}_{fi}$$

where $\vec{d}_{fi} = \langle f|q\hat{\vec{r}}|i\rangle$ is the transition dipole moment — the single most important quantity in atomic spectroscopy. The perturbation in the dipole approximation can be written:

$$\hat{V}(t) = -\hat{\vec{d}} \cdot \vec{E}(t)$$

where $\hat{\vec{d}} = q\hat{\vec{r}}$ is the electric dipole operator and $\vec{E}(t) = \vec{E}_0\cos\omega t$ is the electric field of the light wave.

💡 Key Insight: The electric dipole approximation is the workhorse of atomic physics. It works because atoms are tiny compared to optical wavelengths. When it fails — for transitions where the dipole matrix element vanishes — we must go to higher orders: the magnetic dipole (M1) and electric quadrupole (E2) terms, which arise from the next terms in the expansion $e^{i\vec{k}\cdot\vec{r}} \approx 1 + i\vec{k}\cdot\vec{r} + \cdots$. These higher-order transitions are typically $10^{3}$–$10^{5}$ times weaker than E1 transitions.

📊 By the Numbers: For the hydrogen $1s \to 2p$ transition: - Wavelength: $\lambda = 121.6$ nm (Lyman-$\alpha$, ultraviolet) - Atom size: $a_0 = 0.0529$ nm - Ratio: $a_0/\lambda = 4.35 \times 10^{-4}$ - Dipole matrix element: $|\langle 2p|e\hat{r}|1s\rangle| = (2^7/3^5)^{1/2}\, ea_0 \approx 0.745\, ea_0$ - Transition rate (spontaneous emission): $A_{2p \to 1s} = 6.27 \times 10^8\,\text{s}^{-1}$ - Lifetime: $\tau = 1/A = 1.60$ ns

Worked Example: Calculating the Hydrogen $1s \to 2p$ Dipole Matrix Element

Let us compute the transition dipole moment explicitly. The wavefunctions are:

$$\psi_{100} = \frac{1}{\sqrt{\pi}}\left(\frac{1}{a_0}\right)^{3/2} e^{-r/a_0}, \qquad \psi_{210} = \frac{1}{4\sqrt{2\pi}}\left(\frac{1}{a_0}\right)^{3/2}\frac{r}{a_0}e^{-r/(2a_0)}\cos\theta$$

For the $z$-component of the dipole operator ($q = 0$, corresponding to $\pi$-polarization), we need:

$$\langle 210|ez|100\rangle = e\int_0^\infty R_{21}(r)\,r\,R_{10}(r)\,r^2\,dr \times \int_0^\pi\int_0^{2\pi} Y_1^0(\theta,\phi)\,\cos\theta\,Y_0^0(\theta,\phi)\,\sin\theta\,d\theta\,d\phi$$

The angular integral (using $z = r\cos\theta$ and $Y_1^0 = \sqrt{3/4\pi}\cos\theta$, $Y_0^0 = 1/\sqrt{4\pi}$):

$$\int Y_1^{0*}\cos\theta\,Y_0^0\,d\Omega = \sqrt{\frac{3}{4\pi}}\cdot\frac{1}{\sqrt{4\pi}}\int_0^\pi\cos^2\theta\sin\theta\,d\theta\cdot 2\pi = \frac{1}{\sqrt{3}}$$

The radial integral:

$$\int_0^\infty R_{21}(r)\,r\,R_{10}(r)\,r^2\,dr = \frac{1}{2\sqrt{6}}\cdot 2\cdot\frac{1}{a_0^3}\int_0^\infty \frac{r}{a_0}\,r\,e^{-3r/(2a_0)}\,r^2\,dr$$

$$= \frac{1}{\sqrt{6}\,a_0^4}\int_0^\infty r^4\,e^{-3r/(2a_0)}\,dr = \frac{1}{\sqrt{6}\,a_0^4}\cdot 4!\left(\frac{2a_0}{3}\right)^5 = \frac{2^7\sqrt{2}}{3^5}\,a_0$$

This gives $|\langle 210|er|100\rangle| = ea_0 \cdot 2^7\sqrt{2}/(3^5\sqrt{3}) = 0.745\,ea_0/\sqrt{3}$ for the $m = 0$ component. Summing over all three $m$-substates ($m = -1, 0, +1$) and using the fact that $\sum_m|\langle 21m|\hat{r}_q|100\rangle|^2$ is the same for each value of the polarization index $q$, we obtain the total:

$$|\vec{d}_{2p \to 1s}|^2 = e^2a_0^2\left(\frac{2^7\sqrt{2}}{3^5}\right)^2\frac{1}{3} = \frac{2^{15}}{3^{11}}\,e^2a_0^2$$

Inserting this into the Einstein $A$ formula with $\omega_0 = 3E_1/(4\hbar)$ where $E_1 = 13.6$ eV yields $A = 6.27 \times 10^8\,\text{s}^{-1}$, confirming the result.


21.5 Absorption and Stimulated Emission

Transition Rates for a Monochromatic Field

Combining the results of Sections 21.2–21.4, the first-order transition rate for an atom in a monochromatic field (in the electric dipole approximation) is given by Fermi's golden rule:

Absorption ($E_f > E_i$, atom goes up):

$$\Gamma_{i \to f}^{\text{abs}} = \frac{\pi q^2 E_0^2}{2m^2\hbar\omega^2}\,|\hat{\epsilon}\cdot\langle f|\hat{\vec{p}}|i\rangle|^2\,\delta(E_f - E_i - \hbar\omega)$$

Or equivalently, in the electric dipole form using $\vec{d}_{fi}$:

$$\Gamma_{i \to f}^{\text{abs}} = \frac{\pi E_0^2}{2\hbar^2}\,|\hat{\epsilon}\cdot\vec{d}_{fi}|^2\,\delta(\omega_{fi} - \omega)$$

Stimulated emission ($E_f < E_i$, atom goes down):

$$\Gamma_{i \to f}^{\text{stim}} = \frac{\pi E_0^2}{2\hbar^2}\,|\hat{\epsilon}\cdot\vec{d}_{fi}|^2\,\delta(\omega_{if} - \omega)$$

A remarkable fact: the rates for absorption and stimulated emission are identical (for the same pair of states and the same driving field). This is because $|\vec{d}_{fi}|^2 = |\vec{d}_{if}|^2$ — the matrix element squared is symmetric under interchange of initial and final states. Einstein noticed this in 1917, as we will see in the next section.

Broadband Radiation: Connection to Spectral Energy Density

In practice, radiation is rarely perfectly monochromatic. For a thermal (broadband) radiation field, we must integrate over the frequency spectrum. The spectral energy density $u(\omega)$ gives the electromagnetic energy per unit volume per unit angular frequency:

$$u(\omega) = \frac{\epsilon_0 E_0^2(\omega)}{2}$$

(in SI units). After integrating the transition rate over the spectrum of the radiation field and averaging over the random polarization directions of thermal radiation (which introduces a factor of $1/3$ because $\langle|\hat{\epsilon}\cdot\hat{r}|^2\rangle_{\text{avg}} = 1/3$ for isotropic light), we obtain:

$$\Gamma_{i \to f}^{\text{abs}} = \frac{\pi}{3\epsilon_0\hbar^2}\,|\vec{d}_{fi}|^2\,u(\omega_{fi})$$

This is the Einstein B coefficient (times the energy density):

$$\Gamma_{i \to f}^{\text{abs}} = B_{i \to f}\, u(\omega_{fi})$$

with

$$\boxed{B_{i \to f} = \frac{\pi}{3\epsilon_0\hbar^2}\,|\vec{d}_{fi}|^2 = B_{f \to i}}$$

The equality $B_{i \to f} = B_{f \to i}$ — the absorption and stimulated emission coefficients are identical — is one of Einstein's great insights from 1917.

🧪 Experiment: The equality of absorption and stimulated emission rates was verified experimentally by Willis Lamb and colleagues in the 1940s–50s using microwave spectroscopy of hydrogen. Their measurements of the Lamb shift (Chapter 18) relied critically on stimulated emission to depopulate excited states selectively.


21.6 Spontaneous Emission and the Einstein A/B Coefficients

The Problem: What Makes Atoms Glow?

There is a glaring gap in our treatment so far. Time-dependent perturbation theory, as we have developed it, describes transitions induced by an external electromagnetic field — absorption and stimulated emission. But atoms in excited states decay spontaneously, emitting photons even in the absence of any external field.

Spontaneous emission cannot be explained within the semiclassical framework (classical electromagnetic field + quantum atom) that we have been using. It requires the quantization of the electromagnetic field — treating photons as quantum objects with creation and annihilation operators (Chapter 34). The full derivation belongs to quantum electrodynamics (QED).

However, Einstein, in a stroke of brilliance in 1917, derived the spontaneous emission rate without quantizing the field, using a thermodynamic argument that remains one of the most elegant pieces of reasoning in all of physics.

Einstein's 1917 Argument

Consider a collection of atoms with two energy levels $|1\rangle$ and $|2\rangle$ ($E_2 > E_1$), bathed in thermal radiation at temperature $T$. In thermal equilibrium, the rate of upward transitions (absorption) must equal the rate of downward transitions (stimulated + spontaneous emission).

Let $N_1$ and $N_2$ be the populations of the two levels. Einstein postulated three processes:

  1. Absorption: rate = $B_{12}\, u(\omega_0)\, N_1$
  2. Stimulated emission: rate = $B_{21}\, u(\omega_0)\, N_2$
  3. Spontaneous emission: rate = $A_{21}\, N_2$

where $\omega_0 = (E_2 - E_1)/\hbar$. In equilibrium, rates balance:

$$B_{12}\, u(\omega_0)\, N_1 = B_{21}\, u(\omega_0)\, N_2 + A_{21}\, N_2$$

The populations in thermal equilibrium are given by the Boltzmann distribution:

$$\frac{N_2}{N_1} = e^{-\hbar\omega_0/k_B T}$$

Substituting and solving for $u(\omega_0)$:

$$u(\omega_0) = \frac{A_{21}/B_{21}}{(B_{12}/B_{21})\,e^{\hbar\omega_0/k_B T} - 1}$$

But we know from Planck's radiation law (Chapter 1!) that the thermal energy density is:

$$u(\omega_0) = \frac{\hbar\omega_0^3}{\pi^2 c^3}\,\frac{1}{e^{\hbar\omega_0/k_B T} - 1}$$

Comparing the two expressions term by term:

$$\boxed{B_{12} = B_{21}} \qquad \text{(already known from perturbation theory)}$$

$$\boxed{A_{21} = \frac{\hbar\omega_0^3}{\pi^2 c^3}\, B_{21}}$$

This is the Einstein A-B relation. It connects the spontaneous emission rate $A$ (a quantum phenomenon requiring field quantization to derive directly) to the stimulated emission/absorption rate $B$ (calculable from semiclassical perturbation theory). Using our expression for $B$:

$$\boxed{A_{21} = \frac{\omega_0^3}{3\pi\epsilon_0\hbar c^3}\,|\vec{d}_{21}|^2}$$

💡 Key Insight: Einstein derived a fundamentally quantum result (spontaneous emission rate) using only thermodynamics and Planck's law. This is possible because in thermal equilibrium, the detailed balance condition constrains the relationship between the $A$ and $B$ coefficients. The physics content is that the vacuum itself fluctuates, and these vacuum fluctuations stimulate emission — but Einstein did not need to know this to get the right answer.

🔵 Historical Note: Einstein's 1917 paper "Zur Quantentheorie der Strahlung" (On the Quantum Theory of Radiation) is one of the most consequential papers in 20th-century physics. Not only did it establish the A and B coefficients, it also predicted stimulated emission — the phenomenon that would eventually make lasers possible, 43 years later.

Properties of Spontaneous Emission

The $A$ coefficient has several important features:

1. The $\omega^3$ scaling. The spontaneous emission rate scales as the cube of the transition frequency. This has profound consequences:

  • Optical transitions ($\omega \sim 10^{15}$ Hz): $A \sim 10^{7}$–$10^{9}\,\text{s}^{-1}$, lifetimes of nanoseconds
  • Microwave transitions ($\omega \sim 10^{10}$ Hz): $A \sim 10^{-3}\,\text{s}^{-1}$, lifetimes of minutes to hours
  • Radio-frequency transitions ($\omega \sim 10^{6}$ Hz): $A \sim 10^{-15}\,\text{s}^{-1}$, essentially infinite lifetimes

This is why the 21-cm hydrogen line (a radio-frequency hyperfine transition) has a lifetime of approximately 11 million years — and yet it is the most important spectral line in radio astronomy, because there is so much hydrogen in the universe.

2. Ratio of spontaneous to stimulated emission. In thermal equilibrium at temperature $T$:

$$\frac{A_{21}}{B_{21}\,u(\omega_0)} = e^{\hbar\omega_0/k_B T} - 1$$

At room temperature ($k_B T \approx 0.026$ eV) and optical frequencies ($\hbar\omega \approx 2$ eV), this ratio is $\sim e^{77} \approx 10^{33}$. Spontaneous emission completely dominates. Stimulated emission becomes important only when $\hbar\omega \lesssim k_B T$ (microwave frequencies at room temperature) or when the radiation density is enormously enhanced above the thermal value — as inside a laser cavity.

📊 By the Numbers: At what temperature does stimulated emission equal spontaneous emission for the hydrogen Lyman-$\alpha$ line ($\lambda = 121.6$ nm, $\hbar\omega = 10.2$ eV)? Setting $A = Bu$, we need $k_BT \approx \hbar\omega/\ln 2 \approx 10.2/0.693 \approx 14.7$ eV, corresponding to $T \approx 170{,}000$ K — the temperature of the early universe about 400,000 years after the Big Bang, when recombination occurred.

Spontaneous Emission as Stimulation by Vacuum Fluctuations

Why does an excited atom radiate even when no external field is present? Within our semiclassical treatment (classical field, quantum atom), spontaneous emission is simply postulated — Einstein's $A$ coefficient is determined by consistency with thermodynamics, not derived from first principles.

The deeper explanation, which we will develop fully in Chapter 27 (Quantum Optics) and Chapter 34 (Second Quantization), is that the electromagnetic field is itself quantized. Each mode of the field is a quantum harmonic oscillator with zero-point energy $\hbar\omega/2$. Even in the vacuum state $|0\rangle$ (no photons), the electric field fluctuates:

$$\langle 0|\hat{E}^2|0\rangle = \frac{\hbar\omega}{2\epsilon_0 V} \neq 0$$

These vacuum fluctuations provide the "stimulation" that triggers spontaneous emission. In the fully quantized theory, the matrix element for emission involves the photon creation operator $\hat{a}^\dagger$, and the vacuum state contributes through $\langle 1|\hat{a}^\dagger|0\rangle = 1$. The spontaneous emission rate computed this way agrees exactly with Einstein's thermodynamic result — a beautiful consistency check between two very different approaches.

There is a complementary viewpoint: the radiation reaction of the atom's own electromagnetic field acts back on the atom and causes it to radiate, just as a classical accelerating charge radiates. Dalibard, Dupont-Roc, and Cohen-Tannoudji showed in 1982 that in a symmetric ordering of field operators, vacuum fluctuations and radiation reaction each contribute exactly half the spontaneous emission rate. The two contributions are individually ordering-dependent, but their sum is physical and ordering-independent.

⚖️ Interpretation: The question "Does vacuum fluctuation or radiation reaction cause spontaneous emission?" does not have a gauge-invariant answer. What is unambiguous is that spontaneous emission requires field quantization — it cannot be explained by classical electromagnetism coupled to quantum mechanics. This was one of the earliest triumphs of quantum electrodynamics.

The Purcell Effect: Modifying Spontaneous Emission

A remarkable consequence of the vacuum-fluctuation picture is that spontaneous emission rates can be modified by changing the electromagnetic environment. If you place an atom inside a cavity whose resonant frequency matches the atomic transition, the density of electromagnetic modes is enhanced relative to free space, and the spontaneous emission rate increases. Conversely, if the cavity has no mode at the transition frequency, spontaneous emission is suppressed.

This is the Purcell effect, predicted by Edward Purcell in 1946. The enhancement factor for a cavity with quality factor $Q$ and mode volume $V_{\text{mode}}$ is:

$$F_P = \frac{3Q\lambda^3}{4\pi^2 V_{\text{mode}}}$$

For a microwave cavity with $Q \sim 10^{10}$ (achievable with superconducting resonators), the Purcell factor can exceed $10^6$, dramatically accelerating the emission rate. This is the foundation of cavity quantum electrodynamics (cavity QED), which Serge Haroche and David Wineland pioneered (2012 Nobel Prize).


21.7 Selection Rules from Matrix Elements

Why Do Some Transitions Not Occur?

The transition rate is proportional to $|\hat{\epsilon}\cdot\vec{d}_{fi}|^2 = |\langle f|q\hat{\epsilon}\cdot\hat{\vec{r}}|i\rangle|^2$. If this matrix element vanishes, the transition is forbidden (in the electric dipole approximation). The conditions under which it is nonzero define the selection rules.

For hydrogen-like atoms, the states are labeled by quantum numbers $(n, l, m)$ (or $(n, l, m, m_s)$ including spin). The selection rules follow from the angular integration in the matrix element $\langle n'l'm'|\hat{r}_q|nlm\rangle$, where $\hat{r}_q$ are the spherical components of the position operator.

The position vector $\hat{\vec{r}} = r(\hat{r}_x, \hat{r}_y, \hat{r}_z)$ can be decomposed in terms of spherical harmonics $Y_1^q(\theta, \phi)$ (since $x, y, z$ are proportional to $Y_1^{0,\pm 1}$). The angular integral becomes:

$$\int Y_{l'}^{m'*}(\theta, \phi)\, Y_1^q(\theta, \phi)\, Y_l^m(\theta, \phi)\, \sin\theta\, d\theta\, d\phi$$

Using the properties of Clebsch-Gordan coefficients (Chapter 14), this integral vanishes unless:

The Electric Dipole (E1) Selection Rules

$$\boxed{\Delta l = \pm 1} \qquad \text{(parity must change)}$$

$$\boxed{\Delta m = 0, \pm 1} \qquad \text{(angular momentum conservation with photon)}$$

$$\boxed{\Delta m_s = 0} \qquad \text{(electric dipole does not flip spin)}$$

Additionally, for multi-electron atoms in LS coupling:

$$\Delta L = 0, \pm 1 \quad (L = 0 \not\to L' = 0)$$ $$\Delta S = 0$$ $$\Delta J = 0, \pm 1 \quad (J = 0 \not\to J' = 0)$$

Physical Interpretation

$\Delta l = \pm 1$: The photon carries one unit of angular momentum ($\ell_\gamma = 1$). Conservation of angular momentum requires the atom's orbital angular momentum to change by exactly one unit. This is why $s \to s$ transitions are forbidden (would require $\Delta l = 0$), while $s \to p$, $p \to d$, $p \to s$ transitions are allowed.

$\Delta m = 0, \pm 1$: This corresponds to the three polarizations of the photon: $\Delta m = 0$ for light polarized along $\hat{z}$ ($\pi$-polarization), and $\Delta m = \pm 1$ for right/left circular polarization ($\sigma^{\pm}$-polarization).

Parity: The position operator $\hat{\vec{r}}$ is an odd-parity operator. The matrix element $\langle f|\hat{\vec{r}}|i\rangle$ vanishes unless $|i\rangle$ and $|f\rangle$ have opposite parity. Since the parity of $Y_l^m$ is $(-1)^l$, this requires $\Delta l$ to be odd — consistent with $\Delta l = \pm 1$.

Forbidden Transitions and Higher Multipoles

When the E1 matrix element vanishes, the transition is "forbidden" — but not truly impossible. The next terms in the multipole expansion give:

  • Magnetic dipole (M1): Selection rules $\Delta l = 0$, $\Delta m = 0, \pm 1$. Rates suppressed by $\sim (\alpha)^2 \sim 10^{-5}$ relative to E1.
  • Electric quadrupole (E2): Selection rules $\Delta l = 0, \pm 2$, $\Delta m = 0, \pm 1, \pm 2$. Rates suppressed by $\sim (a_0/\lambda)^2 \sim 10^{-7}$ relative to E1.

⚠️ Common Misconception: "Forbidden" transitions do occur — they are just much slower. The famous "forbidden" green line of doubly-ionized oxygen ([O III] $\lambda 5007$) is one of the strongest features in many nebulae. In the extremely low-density interstellar medium, atoms can survive in metastable states long enough for these slow transitions to proceed, because collisional de-excitation (which would be dominant in a laboratory) is rare.

Worked Example: Hydrogen Selection Rules in Action

Consider transitions from the $n = 3$ shell of hydrogen:

Initial state Final state $\Delta l$ Allowed? Type
$3s$ ($l=0$) $2p$ ($l=1$) $+1$ Yes (E1) Balmer-$\alpha$ component
$3s$ ($l=0$) $1s$ ($l=0$) $0$ No (E1 forbidden) Two-photon or M1
$3p$ ($l=1$) $2s$ ($l=0$) $-1$ Yes (E1) Balmer-$\alpha$ component
$3p$ ($l=1$) $1s$ ($l=0$) $-1$ Yes (E1) Lyman-$\beta$
$3d$ ($l=2$) $2p$ ($l=1$) $-1$ Yes (E1) Balmer-$\alpha$ component
$3d$ ($l=2$) $2s$ ($l=0$) $-2$ No (E1 forbidden) E2 allowed
$3d$ ($l=2$) $1s$ ($l=0$) $-2$ No (E1 forbidden) E2 allowed

The $2s$ state of hydrogen is metastable: it cannot decay to $1s$ by E1 emission (because $\Delta l = 0$). It decays instead by two-photon emission with a lifetime of about $1/7$ of a second — eight orders of magnitude longer than the $2p$ state's 1.6 ns lifetime. This enormous difference in lifetimes, arising from a simple selection rule, has major consequences in astrophysics and laboratory spectroscopy.

🔗 Connection: Selection rules are ultimately consequences of symmetry (Chapter 10). The rule $\Delta l = \pm 1$ follows from the rotational symmetry of the atom and the fact that the photon carries spin 1. The rule $\Delta m = 0, \pm 1$ follows from the axial symmetry of the particular polarization component. Every selection rule in physics traces back to a conservation law, which traces back to a symmetry (Noether's theorem).


21.8 The Laser: Light Amplification by Stimulated Emission of Radiation

The Key Idea

Einstein's 1917 analysis identified three radiation processes: absorption, stimulated emission, and spontaneous emission. The crucial property of stimulated emission is that the emitted photon is identical to the stimulating photon — same frequency, same direction, same polarization, same phase. The emitted photon is a perfect copy of the incoming one.

If we could arrange for stimulated emission to dominate over absorption, we could amplify a beam of light: every photon passing through the medium would stimulate the emission of an additional identical photon, producing coherent amplification. This is the laser.

The Population Inversion Problem

Under normal (thermal) conditions, $N_1 > N_2$ (more atoms in the ground state, by the Boltzmann factor). Since $B_{12} = B_{21}$, absorption dominates over stimulated emission: any photon passing through the medium is more likely to be absorbed than to trigger emission. The medium is opaque.

For laser action, we need population inversion: $N_2 > N_1$. This cannot be achieved in a two-level system in thermal equilibrium (it would require a "negative temperature"). It requires a pumping mechanism to drive the system out of equilibrium.

The Three-Level and Four-Level Laser

Three-level laser (e.g., ruby laser, the first laser ever built, by Theodore Maiman in 1960):

  1. Pump: Excite atoms from level 1 (ground) to level 3 (short-lived upper pump level) using a bright flash lamp.
  2. Fast decay: Atoms in level 3 rapidly decay (non-radiatively) to level 2 (metastable upper laser level). This step must be fast.
  3. Laser transition: Stimulated emission from level 2 to level 1 produces the laser light.

The challenge: you must pump more than half the atoms out of the ground state to achieve $N_2 > N_1$. This requires enormous pump power.

Four-level laser (e.g., Nd:YAG laser):

  1. Pump: Excite atoms from level 0 (ground) to level 3 (pump band).
  2. Fast decay: Atoms rapidly decay non-radiatively from level 3 to level 2 (upper laser level, metastable).
  3. Laser transition: Stimulated emission from level 2 to level 1 (lower laser level).
  4. Fast decay: Atoms rapidly decay from level 1 back to level 0.

The advantage: since level 1 is rapidly depopulated, $N_1 \approx 0$, and population inversion ($N_2 > N_1$) is achieved as soon as any atoms reach level 2. Four-level lasers have a much lower pumping threshold.

The Optical Cavity

Population inversion alone is necessary but not sufficient for lasing. The medium must be placed inside an optical cavity — typically two mirrors facing each other — that provides:

  1. Feedback: Photons bounce back and forth, making multiple passes through the gain medium and stimulating more emission on each pass.

  2. Mode selection: The cavity supports only certain standing-wave modes with frequencies $\omega_q = q\pi c/L$ (where $L$ is the cavity length and $q$ is a large integer). This narrows the output spectrum dramatically.

  3. Output coupling: One mirror is slightly transmitting (the "output coupler"), allowing a fraction of the light to escape as the laser beam.

Laser threshold condition: The gain per round trip (from stimulated emission) must exceed the loss per round trip (from mirror transmission, scattering, and absorption):

$$G \geq L$$

Above threshold, the intracavity intensity builds up until the gain is saturated (reduced by stimulated emission depleting the population inversion) and exactly equals the loss — a self-regulating steady state.

Gain and the Threshold Condition

Let us quantify the amplification process. Consider a beam of photons at frequency $\omega_0$ propagating through a medium with population inversion $\Delta N = N_2 - N_1 > 0$ (per unit volume). The net stimulated emission rate per unit length is:

$$\frac{dI}{dz} = \sigma_{\text{stim}}(\omega)\,\Delta N\, I$$

where $\sigma_{\text{stim}}(\omega)$ is the stimulated emission cross section, related to the Einstein $B$ coefficient:

$$\sigma_{\text{stim}}(\omega) = \frac{A_{21}\lambda^2}{8\pi}\,g(\omega - \omega_0)$$

Here $g(\omega - \omega_0)$ is the normalized lineshape function (Lorentzian for natural broadening, Gaussian for Doppler broadening). The solution is exponential amplification:

$$I(z) = I(0)\,e^{\alpha z}, \qquad \alpha = \sigma_{\text{stim}}(\omega_0)\,\Delta N$$

The coefficient $\alpha$ is the gain coefficient. For a laser cavity of length $L$ with mirror reflectivities $R_1$ and $R_2$, a photon makes a round trip and must satisfy:

$$R_1 R_2\, e^{2\alpha L} \geq 1$$

This is the laser threshold condition. Rewriting:

$$\Delta N_{\text{threshold}} = \frac{-\ln(R_1 R_2)}{2\sigma_{\text{stim}}(\omega_0)\,L}$$

For a HeNe laser ($L = 30$ cm, $R_1 = 1.0$, $R_2 = 0.98$, $\sigma_{\text{stim}} \approx 3 \times 10^{-13}\,\text{cm}^2$):

$$\Delta N_{\text{threshold}} \approx \frac{-\ln(0.98)}{2 \times 3\times10^{-13} \times 30} \approx 1.1 \times 10^{9}\,\text{cm}^{-3}$$

This is a tiny fraction of the total neon atom density ($\sim 10^{16}\,\text{cm}^{-3}$), which is why HeNe lasers can operate at very low pump powers (a few milliwatts of electrical discharge).

Properties of Laser Light

Laser light differs fundamentally from thermal (incandescent) light:

Property Thermal light Laser light
Coherence length $\sim \mu$m–mm cm–km
Spectral width THz (broad) Hz–MHz (narrow)
Directionality Isotropic Highly collimated ($\sim$mrad divergence)
Intensity mW/cm$^2$ MW/cm$^2$ or higher (pulsed)
Photon statistics Bose-Einstein (bunched) Poissonian (coherent state)

Every one of these properties follows from the physics of stimulated emission in an optical cavity. The temporal coherence arises because all photons share the same frequency (set by the cavity mode). The spatial coherence and directionality arise because all photons share the same wave vector (the cavity mode). The high intensity is possible because the coherent amplification process has no thermodynamic upper limit (there is no "temperature" associated with a coherent state).

🧪 Experiment: The first laser was built by Theodore Maiman on May 16, 1960, at Hughes Research Laboratories in Malibu, California. It was a ruby laser (three-level, $\lambda = 694.3$ nm, red). When it was announced, the laser was famously described as "a solution looking for a problem." Today, the global laser market exceeds \$20 billion annually, with applications from surgery to telecommunications to gravitational-wave detection (LIGO's mirrors are positioned to sub-attometer precision using laser interferometry).

📊 By the Numbers: Some laser specifications that illustrate the extraordinary range: - HeNe laser (common lab laser): $\lambda = 632.8$ nm, power $\sim 1$ mW, linewidth $\sim 1$ MHz - Nd:YAG (industrial/medical): $\lambda = 1064$ nm, CW power $\sim 1$–100 W; pulsed peak power $\sim 10^{9}$ W - CO$_2$ laser (industrial cutting): $\lambda = 10.6$ $\mu$m, CW power up to 100 kW - National Ignition Facility (fusion research): 192 beams, total energy 2.15 MJ in 20 ns pulses, peak power $\sim 500$ TW (briefly, the most powerful "device" on Earth) - LIGO laser: $\lambda = 1064$ nm, power $\sim 200$ W, frequency stability $\sim 10^{-21}$


21.9 Summary and Progressive Project

What We Accomplished

This chapter has been a journey from the abstract (the interaction picture, perturbative expansions) to the concrete (laser physics, atomic lifetimes, selection rules). Let us take stock of what we have built and where each piece fits.

The central intellectual achievement is the connection between quantum amplitudes (calculated from Schrödinger's equation) and measurable rates (observed in the laboratory). This connection, embodied in Fermi's golden rule, is not obvious. Quantum mechanics is fundamentally reversible and unitary, yet the transitions we observe — radioactive decay, photon emission, atomic absorption — are irreversible. The resolution lies in the density of final states: when a quantum system can transition to a continuum of final states, the reversibility is effectively lost because the probability of returning to the initial state involves a destructive interference of an infinite number of oscillating amplitudes.

This is not merely a mathematical trick. It is the physical origin of irreversibility in quantum mechanics, and it connects directly to the arrow of time and the approach to thermal equilibrium. The machinery we developed here — Fermi's golden rule, Einstein's coefficients, selection rules — is the toolbox with which essentially all of atomic, molecular, nuclear, and particle physics computes observable quantities.

Summary of Key Results

This chapter developed time-dependent perturbation theory and applied it to light-matter interaction. The central results are:

1. Interaction picture. The state $|\psi_I(t)\rangle = e^{i\hat{H}_0 t/\hbar}|\psi_S(t)\rangle$ evolves only under the perturbation:

$$i\hbar \frac{d}{dt}|\psi_I(t)\rangle = \hat{V}_I(t)|\psi_I(t)\rangle$$

This separates "trivial" free evolution from the interesting perturbation-induced dynamics.

2. First-order transition amplitude:

$$c_f^{(1)}(t) = -\frac{i}{\hbar}\int_0^t V_{fi}(t')\, e^{i\omega_{fi}t'}\, dt'$$

For a constant perturbation: $P_{i\to f}(t) = (|V_{fi}|^2/\hbar^2)\sin^2(\omega_{fi}t/2)/(\omega_{fi}/2)^2$.

3. Fermi's golden rule:

$$\Gamma_{i \to f} = \frac{2\pi}{\hbar}\,|\langle f|\hat{V}|i\rangle|^2\,\rho(E_f)$$

Constant transition rate into a continuum of final states. Energy conservation enforced by the delta function hidden in $\rho$.

4. Einstein coefficients. For electric dipole transitions in thermal radiation:

$$B_{12} = B_{21} = \frac{\pi}{3\epsilon_0\hbar^2}\,|\vec{d}_{12}|^2, \qquad A_{21} = \frac{\omega_0^3}{\pi^2 c^3}\cdot\hbar\, B_{21} = \frac{\omega_0^3}{3\pi\epsilon_0\hbar c^3}\,|\vec{d}_{21}|^2$$

5. Selection rules. Electric dipole transitions require $\Delta l = \pm 1$, $\Delta m = 0, \pm 1$, $\Delta m_s = 0$.

6. The laser. Population inversion + stimulated emission + optical cavity = coherent light amplification.

Concept Map

    Time-Dependent Schrödinger Eq.
              │
              ▼
    Interaction Picture (Sec 21.1)
              │
     ┌────────┴────────┐
     ▼                 ▼
 First-Order       Higher-Order
 Amplitude         (Dyson series)
 (Sec 21.2)
     │
     ├──── Constant perturbation → sin² resonance
     │
     ├──── Harmonic perturbation → absorption/emission
     │
     ▼
 Fermi's Golden Rule (Sec 21.3)
     │
     ├──── + E&M field (Sec 21.4)
     │         │
     │         ▼
     │     Dipole approximation
     │         │
     │    ┌────┴────┐
     │    ▼         ▼
     │  Absorption  Stimulated
     │  (Sec 21.5)  Emission
     │    │         │
     │    ▼         ▼
     │  Einstein B  Einstein A (Sec 21.6)
     │              │
     │              ▼
     │         Spontaneous Emission
     │
     ├──── Selection Rules (Sec 21.7)
     │         Δl = ±1, Δm = 0,±1
     │
     └──── The Laser (Sec 21.8)
              Population inversion
              + optical cavity

Progressive Project: Time-Dependent Perturbation Theory Module

Toolkit version 0.21 additions:

Add the following to your quantum simulation toolkit:

  1. transition_prob(H0, V, psi_i, psi_f, t_array) — Calculate first-order transition probability $P_{i\to f}(t)$ for a general time-dependent perturbation by numerical integration of $c_f^{(1)}(t)$.

  2. fermi_golden_rule(V_fi, rho_E) — Compute the Fermi golden rule transition rate $\Gamma = (2\pi/\hbar)|V_{fi}|^2\rho(E_f)$.

  3. dipole_matrix_element(psi_f, psi_i, r_grid) — Evaluate the transition dipole moment $\vec{d}_{fi} = \langle f|q\hat{\vec{r}}|i\rangle$ for hydrogen-like wavefunctions.

  4. einstein_coefficients(d_fi, omega) — Compute the Einstein $A$ and $B$ coefficients from the transition dipole moment and transition frequency.

  5. check_selection_rules(l_i, m_i, l_f, m_f) — Return whether a given $(l_i, m_i) \to (l_f, m_f)$ transition is E1-allowed.

These functions connect to your existing toolkit: - Uses hydrogen_Rnl() from Chapter 5 for radial wavefunctions - Uses time_evolve() from Chapter 7 for comparison with exact evolution - Uses perturbation_1st() from Chapter 17 for consistency checks - Uses clebsch_gordan() from Chapter 14 for angular integrals

Test cases:

  • Verify that the hydrogen $1s \to 2p$ transition rate matches the known value $A = 6.27 \times 10^8\,\text{s}^{-1}$.
  • Confirm that $1s \to 2s$ gives $|\vec{d}_{fi}| = 0$ (E1 forbidden).
  • Check that the ratio $A/B$ agrees with Einstein's formula.
  • Compare perturbative transition probability with exact Rabi oscillation for a two-level system in a weak field.

The Four Techniques

This chapter introduced four key techniques that you should now be able to deploy:

# Technique When to use it Key formula
1 Interaction-picture transformation Whenever $\hat{H} = \hat{H}_0 + \hat{V}(t)$ and you want to isolate the effect of $\hat{V}$ $|\psi_I\rangle = e^{i\hat{H}_0 t/\hbar}|\psi_S\rangle$
2 First-order transition amplitude Calculating $P_{i\to f}$ for weak perturbations and discrete final states $c_f^{(1)} = -(i/\hbar)\int_0^t V_{fi}e^{i\omega_{fi}t'}dt'$
3 Fermi's golden rule Transition rates into a continuum of final states $\Gamma = (2\pi/\hbar)|V_{fi}|^2\rho(E_f)$
4 Selection rule analysis Determining which transitions are allowed before computing rates $\Delta l = \pm 1$, $\Delta m = 0, \pm 1$ for E1

Looking Forward

🔗 Connection: The machinery of this chapter — Fermi's golden rule, selection rules, transition rates — reappears throughout the rest of the book:

  • Chapter 22 (Scattering): The Born approximation is essentially Fermi's golden rule applied to scattering states. The density of states becomes the density of plane-wave final states.

  • Chapter 27 (Quantum Optics): Quantizing the electromagnetic field replaces our classical $\vec{A}$ with creation/annihilation operators, giving a fully quantum derivation of spontaneous emission.

  • Chapter 29 (Dirac Equation): Relativistic corrections to transition rates, spin-flip transitions, and the anomalous magnetic moment modify the selection rules we derived here.

  • Chapter 34 (Second Quantization): Fermi's golden rule in the language of field operators — the natural framework for nuclear and particle physics.


Key Vocabulary

Term Definition
Interaction picture Representation where states evolve under $\hat{V}_I$ and operators evolve under $\hat{H}_0$
Bohr frequency $\omega_{fi} = (E_f - E_i)/\hbar$; the natural oscillation frequency between two energy levels
Transition amplitude $c_f^{(1)}(t)$; the first-order probability amplitude for $|i\rangle \to |f\rangle$
Transition probability $P_{i\to f}(t) = |c_f^{(1)}(t)|^2$
Fermi's golden rule $\Gamma = (2\pi/\hbar)|V_{fi}|^2\rho(E_f)$; constant rate for transitions into a continuum
Density of states $\rho(E)$; number of quantum states per unit energy interval
Minimal coupling $\hat{\vec{p}} \to \hat{\vec{p}} - q\vec{A}/c$; prescription for coupling charged particles to E&M fields
Electric dipole approximation Setting $e^{i\vec{k}\cdot\vec{r}} \approx 1$; valid when $\lambda \gg a_0$
Transition dipole moment $\vec{d}_{fi} = \langle f|q\hat{\vec{r}}|i\rangle$; determines E1 transition strength
Einstein A coefficient Spontaneous emission rate; $A_{21} = (\omega_0^3/3\pi\epsilon_0\hbar c^3)|\vec{d}_{21}|^2$
Einstein B coefficient Stimulated emission/absorption rate per unit spectral energy density
Selection rule Condition on quantum numbers for a nonzero matrix element
Population inversion $N_2 > N_1$; necessary condition for laser action
Stimulated emission Photon-induced emission of an identical photon; basis of laser amplification
Metastable state Excited state with no E1 decay channel; long lifetime

Chapter 22 will apply the same mathematical framework — perturbation theory and Fermi's golden rule — to the scattering problem, where a particle interacts briefly with a potential and emerges into a continuum of free-particle final states.