Case Study 2: The 2020 Election Needle and the Annotation That Explained Itself
On election night in November 2020, tens of millions of Americans watched a single chart on the New York Times website as it updated live with incoming votes. The chart was strange, unfamiliar, and technically complex. It worked anyway, and it worked because of its annotations.
The Situation
At roughly 8:00 PM Eastern on November 3, 2020, polls closed in much of the United States and the first presidential election returns started flowing in. The New York Times, as it had done since 2016, published a "needle" — a forecasting chart that combined live vote counts with a statistical model of the still-uncounted votes to produce a probabilistic estimate of the eventual winner. The needle was not a simple bar chart. It was a curved dial, like an automotive speedometer, with a pointer that moved between "Likely Trump" and "Likely Biden" as the model's estimate updated. The dial had probability zones marked on it. The pointer oscillated visibly as new results came in. Millions of people watched it for hours.
The needle had been controversial since its debut in 2016, when it had famously shifted from "Likely Clinton" to "Likely Trump" over the course of the evening, delivering a psychologically brutal live-news experience to anyone who had treated the initial forecast as a prediction. Many viewers had asked the NYT to remove the needle entirely. The NYT team had kept it, partly because they believed the needle communicated uncertainty in a way traditional forecasts did not, and partly because its audience watched it despite the discomfort.
For the 2020 election, the NYT graphics team faced a specific communication challenge. The 2020 election was going to involve an unprecedented volume of mail-in and early voting, much of which would be counted after Election Day. In many swing states, the early-counted votes (Election Day, in person) were expected to lean Republican, while the later-counted votes (mail-in, absentee, early voting) were expected to lean Democratic. This "red mirage" effect — a state appearing to favor Trump on election night and then shifting toward Biden as mail votes were counted — was known to the team in advance. The needle's model accounted for it. The viewer's intuition did not.
The team's problem was communication, not modeling. How do you show a live-updating chart, based on a model that knows about the red mirage, to an audience that does not know about the red mirage? How do you prevent the audience from panicking when the raw count shows Trump leading in Pennsylvania by 700,000 votes, when the model knows that 2 million mail ballots from Philadelphia and Pittsburgh are still uncounted and are heavily Democratic?
The answer, for the NYT team, was annotation — aggressive, explicit, specific annotation on the chart itself, explaining to every viewer what the model was assuming, why the dial was pointing where it was pointing, and what the viewer should expect in the hours ahead.
The Data
The underlying data for the needle was a continuously updating combination of several sources:
- Live vote counts from the Associated Press and individual state election boards, updated as counties reported their tallies
- A statistical model developed by the NYT team that estimated how the uncounted votes in each state would likely split, based on historical precinct-level patterns, mail-ballot versus in-person voting patterns, and demographic factors
- Uncertainty estimates around the model's predictions, expressed as a probabilistic range
- Pre-election polling used to establish priors for the model
The challenge for visualization was that the raw data (live vote counts) and the modeled data (projected final outcome) were telling different stories at different moments in the evening. At 10:00 PM, the raw vote in Pennsylvania might show Trump leading. The model, accounting for the 1.5 million mail ballots yet to be counted, would show Biden favored to win the state. Both were correct. The reader had to understand that both were correct, and the chart had to explain why.
A standard bar chart showing "current vote total" would mislead the reader by hiding the model. A standard probability gauge showing the model's estimate would mislead the reader by hiding the fact that the current count was heading in a different direction. The needle, combined with live-updating vote counts and projection annotations, tried to show both simultaneously.
The Visualization
The 2020 needle, as deployed on election night, had several distinct elements that worked together:
Element 1: The dial itself. A curved gauge running from left (Democrat) to right (Republican), with a pointer indicating the model's current probability estimate. Color-coded zones indicated the confidence level: deep red for "Very Likely Republican," light red for "Lean Republican," a neutral zone for "Too Close to Call," light blue for "Lean Democrat," and deep blue for "Very Likely Democrat."
Element 2: Live vote totals. A small table below the dial showing the current vote count in the state, with the Democratic and Republican totals and the margin. The numbers updated as new counties reported.
Element 3: "Estimated votes remaining" indicator. A count of how many votes the team estimated were still uncounted in the state. This was the critical piece of information the raw vote count did not convey: the dial might be pointing toward Biden because the model knew that 2 million Democrat-heavy mail ballots were still uncounted, even though the current lead was Trump's.
Element 4: Annotations explaining the model. This is where the NYT team went beyond previous needles. The chart included explicit, prominent annotations explaining what the viewer was seeing:
- "We expect Biden to gain ground as mail ballots are counted in Philadelphia and Pittsburgh."
- "The needle is based on our estimate of the final outcome, not the current vote."
- "There are still 1.8 million ballots to count in this state, mostly mail votes that tend to favor Democrats."
- "If the current trend in mail votes continues, we expect Biden to win this state by about 1 point."
These annotations appeared next to the dial, updating as the state of the race changed. They were written in plain language, not statistical jargon. They explained why the dial was pointing where it was pointing. They told the viewer what to expect next.
Element 5: A live-updating timestamp. "Last updated 11:37 PM ET." This small detail mattered — the viewer knew how recent the chart was, which anchored it in the ongoing news flow.
Element 6: A source attribution and methodology link. "Source: The New York Times statistical model. Read about our methodology." The link took the reader to a detailed explanation of how the model worked, which satisfied readers who wanted the full technical story.
The Impact
The 2020 needle was, by several measures, one of the most-watched charts in the history of data visualization. The NYT reported that tens of millions of viewers loaded the election-night live results page, and the needle was a prominent feature of that page for most of them. The chart was discussed on cable news, referenced in Twitter threads, criticized by some and defended by others, and screenshotted countless times. It became part of the vocabulary of election night 2020 in a way no other chart had been in any previous election.
More importantly for this case study, the needle's annotations were a significant part of the conversation. Viewers who had experienced the 2016 needle — where the chart had seemed to "change its mind" mid-evening — were primed to distrust it in 2020. The annotations addressed this distrust directly. When a viewer saw the needle pointing toward Biden while Trump was leading in Pennsylvania by 700,000 votes, the annotation explained: "We expect Biden to gain ground as mail ballots are counted." The explanation did not eliminate the uncertainty, but it gave the viewer a framework for interpreting what they were seeing. The needle was not wrong; it was accounting for information the raw vote total did not show.
In the days following the election, as mail ballots were counted and the states gradually resolved, the needle's projections turned out to be broadly accurate. The team's model had correctly anticipated the red mirage and the subsequent shift. The annotations had correctly warned viewers what to expect. Some viewers who had initially panicked during the Election Day evening (when Trump was leading in the raw count) later reported that the annotations had helped them stay calm — they had understood what the dial was doing and why, and they had not been surprised when the final count confirmed the model's prediction.
The 2020 needle also generated a different kind of impact: a public education in how probabilistic forecasts work. Many viewers had never encountered a live statistical model before 2020. The needle, combined with its annotations, gave them a crash course in probability interpretation. "This chart says Biden has a 70% chance of winning Pennsylvania, which means the model estimates Trump has a 30% chance — that's not negligible." "The needle is not a prediction; it is a current estimate that updates as more information comes in." "The dial can shift because the model gets new information, not because the chart is broken." These lessons were embedded in the chart through its annotations, and viewers absorbed them by watching the chart behave as it updated.
Not everyone was persuaded. Some viewers continued to hate the needle — they found the oscillation stressful, the probability framing confusing, and the whole concept of a live-updating forecast inappropriate for election night. The NYT team responded to this feedback by offering a "static" view of the live results that did not include the needle, letting viewers choose their own level of statistical engagement. But the needle remained for those who wanted it, and the team defended its existence on the grounds that a well-annotated probabilistic forecast was more honest than a simple vote count that would mislead viewers about the state of the race.
Why It Worked: A Theoretical Analysis
The 2020 needle, and specifically its annotations, succeeded for reasons that connect directly to the principles of this chapter.
1. The annotations carried the specific context the raw data did not. A live vote total without context was misleading — it looked like Trump was winning Pennsylvania when in fact he was not. An annotation that said "1.8 million mail ballots still to count, mostly Democratic" gave the viewer the specific information they needed to interpret the raw total correctly. The annotation did the work of a paragraph of explanation in fifteen words. This is the "text that does the most work per word" principle in action.
2. The annotations were placed where the data was. The NYT team did not put the methodology explanation in a separate article that viewers would have to find. They put it directly on the chart, next to the needle, updating as the state of the race changed. Viewers could not miss it, because it was in the same visual field as the dial they were watching. This is the "annotations should be near the data they describe" principle applied to a live, updating context.
3. The annotations were in plain language. A statistical model's assumptions can be explained in technical jargon (priors, posteriors, confidence intervals, Bayesian updating) or in plain language ("we expect Biden to gain ground as mail ballots are counted"). The NYT team chose plain language, and the choice was essential. Statistical jargon would have alienated the general audience and left them no better off than before. Plain language gave everyone the same access to the model's reasoning. This is a matching of the chart text to the audience — one of the typographic principles from Section 7.2.
4. The annotations acknowledged uncertainty. Rather than claiming certainty ("Biden will win"), the annotations used the language of probability and expectation ("We expect Biden to gain ground"). This was honest about the model's limits and gave viewers a vocabulary for thinking about uncertainty. The alternative — claiming certainty — would have been more persuasive in the short term but would have collapsed if the model had been wrong. The uncertainty-acknowledging language is also ethically aligned with Chapter 4's discussion of context omission: hiding uncertainty is a form of distortion, and the annotations were designed to prevent that.
5. The chart was self-explanatory for its moment. A viewer landing on the NYT election page could read the needle, the vote totals, the ballot counts, and the annotations, and understand what they were seeing without consulting any external source. The chart passed an extreme version of the self-explanatory test: it worked in a live, high-stress context where viewers had no patience for external documentation. This is the self-explanatory standard at its most demanding, and the chart met it because every piece of essential context was built into the chart itself.
6. The annotations scaled with the chart. As the state of the race changed, the annotations changed. "We expect Biden to gain ground" in the early hours gave way to "Biden has overtaken Trump in this state" once the mail ballots were counted. The annotations were not static; they tracked the live evolution of the situation. This is an advanced application of annotation design, and it is only possible in live-updating charts, but it shows what is achievable when annotation becomes a first-class part of the chart rather than an afterthought.
Complications and Counter-Arguments
The 2020 needle was praised but also criticized, and the critiques are worth understanding.
The needle itself is an unusual chart type. A speedometer gauge for a probability is not a standard chart in any selection matrix. It is a custom design, and custom designs require the audience to learn a new visual vocabulary. Critics argue that this learning cost is unjustified — a simpler chart (a probability bar, or a pair of candidate names with percentage labels) would have communicated the same information with less novelty. The NYT team defends the needle on the grounds that its oscillation makes uncertainty visible in a way a static bar does not, but the debate is legitimate.
The oscillation creates false drama. When the pointer moves visibly every few minutes as new votes come in, the viewer experiences the race as more dramatic than the underlying statistical situation warrants. Small shifts in the model's estimate produce visible motion, and the motion itself becomes emotionally engaging — arguably more engaging than the statistical content deserves. This is a general problem with animated charts: motion is attention-grabbing in ways that can distort interpretation.
The annotations require reading time. A viewer who glances at the needle for three seconds sees the dial and the pointer but may not read the annotations. The annotations are text, which takes longer to process than the dial itself. For very fast glances, the annotations may not reach the viewer. This is a limitation of text-based annotations in a high-attention, low-time context.
The methodology is opaque even with annotations. The plain-language annotations explain what the model is doing but not how. A viewer who wants to verify that the model's assumptions are correct has to go to the methodology page and read the technical details. Most viewers will not do this. This means the annotations are a form of "trust me" communication — they report the model's output in plain language without exposing the model itself to scrutiny. This is a necessary compromise in a live context, but it is a compromise.
The needle's critics were not persuaded. Many viewers continued to hate the needle even with the improved annotations. Some viewers argued that the chart was inherently inappropriate for election night because it treated a live-democratic event as a statistical gambling problem. The NYT team's response was to offer a choice (static view or live needle), which was a reasonable design response, but it did not resolve the underlying disagreement about whether the needle should exist at all.
Lessons for Modern Practice
You will probably not build an election-night needle. But the lessons of the 2020 needle apply to any chart that has to carry statistical context to a non-technical audience.
Put the explanation where the data is. If your chart requires interpretation — a model, an adjustment, a caveat — put the explanation directly on the chart as an annotation, not in a separate document. Viewers will not follow links. They will look at what is in front of them.
Write annotations in plain language. If you find yourself using statistical jargon in an annotation, rewrite it. "We expect Biden to gain ground" is better than "The posterior distribution assigns 0.72 probability to Biden winning conditional on the remaining ballot inventory." The plain version does not lose accuracy; it gains accessibility.
Acknowledge uncertainty in the words, not just the visual. A probability band or an error bar shows uncertainty visually, but most viewers do not read error bars correctly. An annotation that states the uncertainty in words ("There's still a chance Trump wins if mail votes are more Republican than we expect") makes the uncertainty unavoidable.
Let annotations change as the context changes. Static annotations are fine for static charts. Live or interactive charts can and should have annotations that update with the data. The 2020 needle's annotations changed hour by hour; your dashboard's annotations might change day by day. The principle is the same: the annotation should always reflect the current state of the data.
Recognize when you need a methodology note. Some charts need a small "how this is calculated" link or footnote, even if the primary annotations are in plain language. The link is not for most viewers — it is for the minority who want to verify or dig deeper. Providing it is a trust mechanism; not providing it leaves viewers who want to verify with no path to do so.
Accept that some viewers will resist, and give them an alternative. The NYT team offered a static view for viewers who hated the needle. This is a design pattern: when a chart type generates strong negative reactions from a portion of the audience, offer an alternative view that satisfies them. Do not try to force everyone through the same visualization.
The bar for annotation is lower than you think. Many charts that would benefit from annotation have none. The friction of adding annotations in matplotlib is real but surmountable, and the benefit to the reader is often large. The 2020 needle shows what annotations can do in the most extreme context; your charts can borrow a fraction of that benefit with a fraction of the effort.
Discussion Questions
-
On the needle as a chart type. The needle is an unusual chart — a curved gauge used for a statistical probability. Is the unfamiliarity of the chart type worth the benefit of making uncertainty visible? Would a simpler chart (a probability bar, a pair of numbers) have communicated the same information without the learning cost?
-
On annotations and live context. The 2020 needle's annotations updated in real time as the state of the race changed. How does this change the practice of writing annotations? What are the challenges of designing annotations that must work at multiple points in an evolving story?
-
On plain language vs. statistical rigor. The NYT team chose plain-language annotations ("we expect Biden to gain ground") over statistically rigorous ones ("posterior probability of 0.72 for Biden"). Was this the right call? What is lost and what is gained in the translation?
-
On the ethics of visible uncertainty. Chapter 4 argued that hiding uncertainty is a form of visualization dishonesty. The needle makes uncertainty visible by oscillating as new data comes in. Does this visible uncertainty serve the reader, or does it mislead them by treating random noise as meaningful motion?
-
On the self-explanatory standard in live contexts. The 2020 needle met an extreme version of the self-explanatory standard: it had to be readable by a viewer who had no external context, who was watching in real time, and who had no patience for methodology documents. What lessons from this extreme context apply to less demanding contexts — for example, a dashboard that is updated daily rather than every few seconds?
-
On audience resistance. Many viewers hated the needle despite the improved annotations. Should a chart maker defer to audience preferences when the preferences conflict with what the maker believes is the more honest visualization? Or should the maker defend the chart that communicates most accurately, even at the cost of audience discomfort?
The 2020 election needle is an extreme case study — a live, unfamiliar chart, for a high-stakes moment, with an audience of tens of millions. But the principles the NYT team relied on are the same principles you can apply to a business dashboard, a policy report, or a single chart in a slide deck. Annotations that explain the chart's logic, placed where the data is, written in plain language, acknowledging uncertainty — these are the tools that turn a technically correct chart into a chart that communicates its meaning. The needle proves they work under pressure. Your charts will rarely be tested under that pressure, which means the tools will work for you with room to spare.