The clock strikes 7:00 PM Eastern, and the first polls close. Within seconds, decision desks at major networks begin processing streams of data: precinct-level returns, exit poll cross-tabs, absentee ballot counts, real-time social media sentiment...
Learning Objectives
- Describe the major sources and types of political data in the modern landscape
- Explain how campaigns, media, and citizens use political data differently
- Identify the key tensions and ethical questions that arise from data-driven politics
- Articulate why measurement choices shape political reality
- Recognize the analytical stakes of a competitive election
In This Chapter
- Opening Scene: Election Night, November 2024
- 1.1 The Data Explosion
- 1.2 A Global Phenomenon with Local Roots
- 1.2b Three Worlds of Political Data
- 1.3 Meet the Garza-Whitfield Race
- 1.4 Meet Meridian Research Group
- 1.5 Meet OpenDemocracy Analytics
- 1.6 What Is Political Analytics?
- 1.7 Data in Democracy: Tool or Weapon?
- 1.8 The Analytical Stakes of the Garza-Whitfield Race
- 1.9 How This Book Is Organized
- 1.10 What You Will Need
- 1.11 The Stakes of Getting It Right---and Getting It Wrong
- 1.12 Returning to Your List
- Chapter Summary
Chapter 1: The Age of Political Data
Opening Scene: Election Night, November 2024
The clock strikes 7:00 PM Eastern, and the first polls close. Within seconds, decision desks at major networks begin processing streams of data: precinct-level returns, exit poll cross-tabs, absentee ballot counts, real-time social media sentiment. In a cramped campaign headquarters two thousand miles away, a young analytics director stares at a dashboard displaying turnout models, persuasion scores, and vote-share projections that update every ninety seconds. Across town, a veteran campaign manager watches the same returns on cable news and refreshes a browser tab showing his pollster's final forecast. In a university basement, a political scientist feeds county-level data into a regression model she has been refining for three years.
All of them are doing the same thing: trying to understand what millions of people have decided to do, and why.
Welcome to political analytics.
This book is about how we measure, model, predict, and explain political behavior using data. It is about the tools, the techniques, the traditions, and the traps. It is about the people who do this work---pollsters and data scientists, campaign operatives and civic technologists, journalists and academics---and the citizens whose lives are shaped by the numbers those professionals produce. Most of all, it is about learning to think critically in an era when political data is everywhere, and when the difference between good analysis and bad analysis can determine who governs, which policies pass, and whether democracy itself functions as intended.
You do not need to be a statistics expert to read this book. You do not need to know Python, though you will learn some. You do not need to have worked on a campaign or in a newsroom. What you do need is curiosity about how politics really works beneath the surface of cable news chyrons and social media arguments, and a willingness to think carefully about numbers that are often presented with more confidence than they deserve.
Let us begin.
1.1 The Data Explosion
The Scale of Modern Political Data
Consider a single registered voter in a competitive state---let us call her Elena, a 34-year-old Latina nurse living in a suburb of Phoenix. Before she casts a single ballot, here is a partial list of what is already known about her, scattered across various databases:
- Voter registration file: Her name, address, date of birth, party registration (if any), and her voting history in every election since she turned eighteen. Did she vote in the 2018 midterm? The 2020 presidential? The 2022 primary? The file records it all.
- Consumer data: Her purchasing habits, magazine subscriptions, car type, estimated household income, homeownership status, and hundreds of other commercial data points, available from data brokers who aggregate information from loyalty programs, credit bureaus, and public records.
- Census data: The demographic profile of her neighborhood---median income, racial composition, educational attainment, housing density---down to her Census block.
- Social media footprint: Her public posts, likes, shares, and the pages she follows, all potentially scrapeable (and often scraped) by campaigns and researchers.
- Donation records: If she has given more than $200 to a federal candidate, her name, employer, and donation amount are publicly available through Federal Election Commission filings.
- Contact history: If a campaign has knocked on her door or called her phone, the result of that interaction---was she persuadable? Already committed? Hostile?---is logged in a campaign's voter contact database.
Now multiply Elena by roughly 160 million registered voters in the United States. Then add the data generated by those voters' representatives: roll-call votes, committee hearing transcripts, floor speeches, press releases, campaign finance filings, lobbying disclosures. Then add the data generated about politics: news articles, television broadcasts, podcasts, social media posts, fact-checks, opinion polls. Then add the data generated by governments: economic indicators, crime statistics, public health records, environmental reports, all of which shape the political environment in which elections occur.
The result is a data ecosystem of staggering size and complexity. And it is growing every day.
💡 Intuition: Political data is not just polls. It encompasses everything from individual voter registration records to macroeconomic indicators to the text of every tweet mentioning a candidate's name. One of the first skills you will develop in this course is learning to see the political data that surrounds you.
Why Now?
The explosion of political data is not an accident. It results from the convergence of several forces:
Digitization of public records. Voter files, campaign finance reports, legislative records, and government statistics that once existed only on paper or in proprietary databases are now available electronically, often for free. The Federal Election Commission's website lets anyone download every federal campaign contribution. State voter files, while varying in accessibility, are available in most states for a fee.
The internet and social media. The rise of digital communication has created an enormous corpus of political expression. Every tweet, Facebook post, Reddit thread, and YouTube comment about politics is, in principle, data. The 2008 Obama campaign is often credited as the first truly "data-driven" presidential campaign, but by 2024, even local school board races routinely use digital targeting.
Advances in computing. The statistical techniques that underpin modern polling, election forecasting, and voter targeting are not new---many date to the mid-twentieth century. What is new is the ability to apply those techniques to massive datasets in real time. A regression model that once required hours on a mainframe now runs in milliseconds on a laptop.
Commercial data infrastructure. The same data brokerage industry that helps retailers target consumers has been adapted for political use. Firms like L2, TargetSmart, and Aristotle maintain enriched voter files that merge public voter registration data with consumer data, creating detailed portraits of individual voters that campaigns use for outreach.
Cultural demand for data. The rise of data journalism, election forecasting sites, and poll aggregators has created a public appetite for political data analysis. FiveThirtyEight, The Economist's election model, and similar projects have made "probability of winning" a standard feature of election coverage. Voters themselves now expect to see data, not just opinions.
📊 Real-World Application: The 2020 U.S. presidential election generated an estimated 750 million data points in commercial voter files alone, not counting social media, polling, or government statistics. By contrast, George Washington's first election in 1789 involved roughly 43,000 voters across ten states, with results transmitted by horseback.
A Word of Caution
Before we go further, a caution: more data does not automatically mean better understanding. The history of political analytics---which you will explore in detail in Chapter 2---is littered with cases where more data led to more confident predictions that turned out to be spectacularly wrong. The 1936 Literary Digest poll surveyed 2.4 million people and predicted Alf Landon would defeat Franklin Roosevelt in a landslide. Roosevelt won 46 of 48 states. In 2016, sophisticated election models gave Hillary Clinton a 70-to-99 percent probability of winning the presidency. She lost the Electoral College.
Data is powerful, but it is not self-interpreting. The numbers do not speak for themselves. Every dataset reflects the choices of the people who created it: what to measure, whom to ask, how to categorize, what to include, what to leave out. These choices are not merely technical. They are political, in the deepest sense of the word.
⚠️ Common Pitfall: Beginners in political analytics often assume that "the data" is a neutral, objective representation of political reality. In fact, every dataset is a construction---shaped by decisions about what to count, how to count it, and who does the counting. Being aware of this is not a weakness; it is the foundation of rigorous analysis.
This is one of the central themes of this book: measurement shapes reality. How we define "likely voter" determines which voices count in a poll. How we draw district lines determines who represents whom. How we categorize race and ethnicity on a Census form determines how resources are allocated. The numbers are never just numbers. They are always the product of human choices, and those choices have consequences.
1.2 A Global Phenomenon with Local Roots
Before we map the landscape of political data in the United States, it is worth acknowledging that data-driven politics is a global phenomenon. In India, the Bharatiya Janata Party built a massive digital outreach operation that reached hundreds of millions of voters through WhatsApp and social media during the 2014 and 2019 elections. In the United Kingdom, the Brexit referendum of 2016 was shaped by targeted digital advertising campaigns that exploited voter data in ways that are still being investigated. In Brazil, Kenya, the Philippines, and dozens of other countries, political campaigns have embraced data analytics with varying degrees of sophistication and varying levels of ethical constraint.
The tools and techniques you will learn in this book are rooted in the American context---American elections, American data infrastructure, American legal frameworks---but the underlying principles are universal. Every democracy faces the same fundamental questions: How do you measure what citizens want? How do you translate those measurements into governance? And who controls the data that mediates between citizens and their government?
🌍 Global Perspective: In 2019, India's general election involved approximately 900 million eligible voters across 29 states and 7 union territories. The BJP's digital campaign reached voters through a network of more than 12 million WhatsApp groups. This scale dwarfs anything in American politics, but the underlying analytical challenge---reaching diverse voters with targeted messages---is the same. Political analytics is not an American invention; it is a democratic necessity that takes different forms in different political systems.
What makes the American case particularly useful for study is the combination of three factors: an extraordinarily rich public data infrastructure (the Census, FEC filings, state voter files), a fiercely competitive two-party system that incentivizes analytical investment, and a tradition of academic research on political behavior that stretches back more than seventy years. These conditions have produced the most developed political data ecosystem in the world---and also the most studied, which means we have a rich body of knowledge about what works, what fails, and why.
As you read this book, keep the global context in mind. The specific datasets and institutions are American, but the analytical principles---how to design a survey, how to model voter behavior, how to assess the quality of a forecast---apply wherever elections are held and wherever citizens seek to understand the political world around them.
1.2b Three Worlds of Political Data
Political data does not exist in a vacuum. It is produced, consumed, and contested by people and organizations with different goals, different resources, and different stakes. To understand the landscape, it helps to think about three overlapping worlds.
The Campaign World
For political campaigns, data is a weapon. The goal is simple: win. Everything else---policy, messaging, mobilization---is in service of that objective. In the campaign world, data is used for:
- Voter targeting: Identifying which voters to contact, what message to deliver, and through which channel (door knock, phone call, digital ad, direct mail).
- Fundraising optimization: Modeling which potential donors are most likely to give, how much, and in response to which appeals.
- Resource allocation: Deciding where to open field offices, where to spend advertising dollars, and where to send the candidate.
- Persuasion modeling: Estimating which voters are persuadable and which are already committed, so campaigns can focus scarce resources where they will have the most impact.
- Turnout modeling: Predicting which supporters will actually vote and targeting get-out-the-vote efforts accordingly.
The campaign world is intensely competitive and secretive. Campaigns treat their data operations as closely guarded advantages. A sophisticated voter targeting model can mean the difference between winning and losing a close race, and campaigns invest millions in building and protecting these tools.
The scale of this investment has grown dramatically. In 2004, the total spending on data and analytics across all federal campaigns was estimated at roughly $100 million. By 2020, that figure had grown to well over $1 billion, driven by the explosion of digital advertising, the sophistication of voter modeling, and the arms race between the two major parties' data operations. This growth reflects a fundamental shift in how campaigns understand their task: not as an exercise in mass communication---broadcasting the same message to everyone---but as an exercise in precision targeting, delivering the right message to the right voter at the right time through the right channel.
For Nadia Osei on the Garza campaign, this means her analytics team is not a support function. It is the nervous system of the entire operation. Every decision the campaign makes---from where to send the candidate for a rally to how to allocate its television advertising budget to which doors to knock on Saturday morning---is informed, at least in part, by the models and analyses her team produces. The pressure is immense. A model that is wrong by two percentage points in a county-level turnout prediction could send volunteer canvassers to the wrong neighborhoods, wasting thousands of hours of effort in the final week before the election.
The Media and Research World
For journalists, academics, and independent analysts, data is a lens. The goal is understanding---or, more precisely, explanation. In this world, data is used for:
- Polling: Measuring public opinion on candidates, issues, and institutions.
- Election forecasting: Building models that predict election outcomes.
- Data journalism: Using quantitative analysis to tell stories about politics that cannot be told through traditional reporting alone.
- Academic research: Testing theories about political behavior, public opinion formation, polarization, and democratic functioning.
This world values transparency, reproducibility, and methodological rigor (at least in principle). Academic researchers publish their methods and data. Reputable pollsters disclose their methodology. Election forecasters explain their models in detail.
But this world also faces pressures. Media organizations need clicks and viewers, which can incentivize sensationalism. Academic publishing rewards novelty over replication. And the line between "analyst" and "advocate" is not always clear, especially in an era of partisan media.
The media and research world has also been transformed by the democratization of analytical tools. A generation ago, election forecasting was the province of a handful of academics with access to mainframe computers and proprietary datasets. Today, anyone with a laptop, an internet connection, and basic programming skills can download public polling data, build a forecasting model, and publish the results on a blog or social media account. This democratization has produced remarkable work---some of the best election analysis in recent years has come from independent analysts working outside traditional institutions---but it has also created a crowded and sometimes confusing information environment, where rigorous analysis sits alongside sloppy work and outright charlatanism with no easy way for non-experts to tell the difference.
At Meridian Research Group, Vivian Park navigates this environment daily. Her polls compete for attention not just with other polling firms but with aggregators, forecasters, political betting markets, social media pundits, and anyone with a spreadsheet and a Twitter account. The challenge is not just producing accurate data but cutting through the noise to ensure that accurate data reaches the people who need it.
The Civic World
For civic technologists, advocacy organizations, and engaged citizens, data is a tool for accountability. The goal is democratic participation---making government more responsive, making information more accessible, making power more visible. In this world, data is used for:
- Government transparency: Publishing campaign finance records, lobbying data, and legislative voting records in accessible formats.
- Redistricting analysis: Using mapping tools and demographic data to identify gerrymandering.
- Voter empowerment: Building tools that help citizens register to vote, find their polling places, and understand what is on their ballot.
- Advocacy: Using data to support arguments about policy, representation, and resource allocation.
The civic world often operates with fewer resources than campaigns or major media organizations. Its practitioners tend to be motivated by ideals of democratic openness and may be suspicious of how campaigns and media use data. But the civic world also faces its own challenges: sustainability (how do you fund open-source tools?), impact (does making data available actually change anything?), and co-optation (what happens when partisan actors use "civic" tools for campaign purposes?).
🔗 Connection: These three worlds---campaign, media/research, and civic---will appear throughout this book. In Chapter 28, you will learn how modern campaigns build data infrastructure. In Chapter 23, you will explore the media ecosystem in detail. And in Chapter 38, you will grapple with the ethical questions that arise when these worlds intersect.
1.3 Meet the Garza-Whitfield Race
To make the ideas in this book concrete, we will follow a fictional but realistic U.S. Senate race throughout the text. Think of it as a laboratory---a controlled environment where you can see analytical concepts play out in a setting that captures the messiness, stakes, and human drama of real politics.
The State
The race takes place in a purple Sun Belt state---let us call it a state that looks demographically like a composite of Arizona, Nevada, Georgia, and parts of Texas. The state's population is approximately 38 percent white non-Hispanic, 32 percent Hispanic/Latino, 18 percent Black, and 12 percent Asian, Pacific Islander, Native American, and multiracial residents. It has a booming metropolitan core, a ring of fast-growing suburbs that have shifted from red to purple in recent cycles, a handful of midsize cities with mixed politics, and a vast rural interior that leans solidly Republican.
The incumbent senator, a Republican, is retiring after two terms. The seat is wide open, and both parties see it as a top pickup opportunity. National money is pouring in. The fundamentals---presidential approval, economic indicators, the national environment---suggest a competitive race.
Maria Garza (D)
Maria Garza is 47 years old, the daughter of Mexican immigrants who settled in the state's capital city when she was three. Her father worked construction; her mother cleaned offices at night and attended community college during the day, eventually earning an accounting degree. Garza went to the state's flagship public university on a scholarship, then to law school, then to the state attorney general's office, where she rose to become the state's first Latina attorney general at age 39.
As AG, Garza built a reputation as a pragmatic progressive---tough on consumer fraud and corporate malfeasance, cautious on criminal justice reform, and effective at building bipartisan coalitions on issues like opioid enforcement and elder abuse. She is disciplined, data-oriented, and not naturally charismatic on the stump, though she can be magnetic in small groups. Her Senate campaign is running on a platform of economic opportunity, healthcare access, and "practical solutions over partisan games."
Her challenge: in a state where 38 percent of the electorate is white non-Hispanic and the rural interior is deeply conservative, she needs to run up enormous margins in the metros, win the suburbs convincingly, and turn out a diverse coalition that includes many infrequent voters.
Tom Whitfield (R)
Tom Whitfield is 54, the owner of a chain of fourteen hardware stores spread across the state's smaller cities and rural towns. He grew up in one of those towns, the son of a schoolteacher and a farm equipment dealer. He played football at a Division II college, got a business degree, and came home to take over his father's single hardware store, which he expanded into a regional brand known for community involvement and old-school customer service.
Whitfield is a first-time candidate who won a crowded primary by running as a populist outsider. He is affable, plainspoken, and skilled at retail politics. His campaign message is built around economic populism---attacking "corporate elites" and "Washington insiders" who have forgotten working families---combined with cultural conservatism on immigration, education, and what he calls "common-sense values." He is not a polished debater, but he has an instinct for connecting with voters who feel overlooked by both parties.
His challenge: he needs to hold the rural base, win back suburban voters who have drifted toward Democrats in recent cycles, and make inroads with working-class Hispanic voters who may be economically sympathetic to his populist message but culturally wary of a Republican Party they associate with anti-immigrant rhetoric.
Nadia Osei: The Analytics Director
Nadia Osei is 31 years old, Ghanaian-American, and the analytics director for the Garza campaign. She grew up in the suburbs of Columbus, Ohio, the daughter of a nurse and a civil engineer who emigrated from Accra in the 1990s. She went to the University of Michigan for political science and statistics, then entered a PhD program at Stanford studying political behavior and computational methods.
She left the PhD program after three years---not because she could not do the work, but because she was frustrated by the pace of academic research and drawn to the immediacy of campaigns. She joined a Democratic data firm, worked on two congressional races and a gubernatorial campaign, and developed a reputation for building voter contact models that were both statistically rigorous and operationally useful. The Garza campaign recruited her away from the firm with the promise of a leadership role and the autonomy to build the analytics operation from scratch.
Nadia is brilliant, intense, and prone to overwork. She thinks in probabilities and confidence intervals, which sometimes makes her difficult to work with for campaign staffers who want simple yes-or-no answers. She is acutely aware that the voters she is modeling---many of them people of color, many of them working-class---are not abstractions but people like her parents and their friends. This awareness shapes her approach to analytics in ways that will become apparent as the book progresses.
🔴 Critical Thinking: Nadia left a PhD program to work in campaign analytics. What does this career choice suggest about the relationship between academic political science and applied political data work? What are the trade-offs between the two? We will explore this tension further in Chapter 41, when we discuss careers in political analytics.
Jake Rourke: The Campaign Manager
Jake Rourke is 48, a veteran Republican campaign manager who has worked on races up and down the ballot for two decades. He grew up in a working-class Irish-Catholic family in a Rust Belt city, worked his way through a state university, and fell into politics almost by accident when he volunteered for a city council campaign after college. He discovered he had a talent for the operational side of campaigns---logistics, scheduling, message discipline---and never left.
Rourke has managed three winning Senate campaigns and two losing ones. He is old school in some ways: he trusts his gut, values relationships, and believes that campaigns are ultimately about "candidate quality and shoe leather." But he is not anti-data. He has seen enough campaigns to know that good data can sharpen strategy, and he has also seen data-obsessed campaigns lose because they forgot that politics is about people, not spreadsheets.
Whitfield hired Rourke because the candidate knew he needed a seasoned hand to professionalize his operation. Rourke signed on because he saw in Whitfield a candidate with genuine populist appeal who could win a race that most Washington Republicans had already written off.
The tension between Rourke's experience-driven intuition and Nadia Osei's model-driven analysis will be one of the through-lines of this book. They represent two ways of knowing in politics, and neither is entirely right or entirely wrong.
🧪 Try This: Before reading further, take five minutes to think about the Garza-Whitfield race from a data perspective. What data would you want to have if you were advising either campaign? What would you want to measure? What questions would you want to answer? Write down at least five specific data needs. We will return to your list at the end of the chapter.
1.4 Meet Meridian Research Group
The second running example in this book is a nonpartisan polling firm called Meridian Research Group. You will meet Meridian's founder and chief methodologist, Dr. Vivian Park, in detail in Chapter 2. For now, here is what you need to know.
Meridian is a mid-sized polling firm based in the same metropolitan area where the Garza-Whitfield race is taking place. It was founded fifteen years ago by Vivian Park, a former political science professor who left academia to build a firm that would combine scholarly rigor with practical relevance. The firm conducts public polls for media clients, private polls for campaigns and advocacy organizations, and occasional pro bono work for nonprofits and academic researchers.
Meridian employs about thirty people, including pollsters, survey methodologists, field directors, data analysts, and support staff. Two of them will appear frequently in this book:
Carlos Mendez is a 24-year-old junior analyst who joined Meridian six months ago after completing a master's degree in survey methodology. He is your proxy in this book---smart, motivated, still learning, and full of the questions that an intelligent newcomer would ask. Carlos grew up in a predominantly Latino neighborhood, went to a state university, and came to polling because he was frustrated by the way polls seemed to consistently misunderstand or misrepresent his community. He is eager to do better.
Trish McGovern is 42, Meridian's senior field director, responsible for managing the logistics of survey data collection---hiring interviewers, managing call centers, overseeing online panel recruitment, and ensuring that fieldwork meets the firm's quality standards. She has been in the polling industry for eighteen years and has seen it transform from a world of live telephone interviews to the current hybrid landscape of online panels, text-to-web surveys, and mixed-mode approaches.
Meridian has been hired by a consortium of media organizations to conduct a series of public polls on the Garza-Whitfield race. This means the firm's work will intersect directly with the campaigns, creating opportunities to explore how polling data shapes campaign strategy and media coverage---and how campaigns, in turn, try to shape polling.
1.5 Meet OpenDemocracy Analytics
The third running example is OpenDemocracy Analytics (ODA), a civic technology nonprofit. You will learn about ODA's founding and data infrastructure in detail in Chapter 3. For now, the essentials:
ODA is a small nonprofit dedicated to making political data accessible and understandable to ordinary citizens, journalists, and researchers. It builds open-source tools for data visualization, voter information, and government transparency. The organization was founded five years ago by Adaeze Nwosu, a former data journalist who became convinced that the biggest barrier to democratic accountability was not a lack of data but a lack of accessible infrastructure for using it.
Adaeze Nwosu is 39, Nigerian-American, and the kind of person who inspires both admiration and exhaustion in those around her. She grew up in Houston, studied journalism and computer science at Northwestern, and spent a decade at major news organizations building data projects---interactive maps, searchable databases, accountability tools. She left journalism to start ODA because she wanted to build tools that would outlast any single news cycle.
Sam Harding is ODA's lead data journalist, 35, non-binary (they/them), and the organization's most visible public-facing analyst. Sam has a gift for translating complex data into accessible narratives and visualizations. They came to ODA from a data journalism fellowship and have become the organization's de facto spokesperson, appearing on panels and podcasts to explain political data to general audiences.
ODA will serve as our window into the civic technology world---the idealists and pragmatists who believe that democracy works better when data is open, tools are free, and citizens have the information they need to hold power accountable.
🔗 Connection: Throughout this book, the Garza-Whitfield race, Meridian Research Group, and OpenDemocracy Analytics will intersect in ways that illuminate the relationships between campaigns, media, and civic organizations. In Chapter 5, you will use ODA's open-source tools to analyze your first political dataset. In Chapter 10, you will evaluate one of Meridian's polls using the techniques you learn in Chapters 6 through 9. And in Chapter 28, you will see the Garza-Whitfield race from the inside, examining how both campaigns use data to make strategic decisions.
1.6 What Is Political Analytics?
Now that you have met the people and the race that will anchor this book, let us step back and define the field.
Political analytics is the application of quantitative and computational methods to the study of political behavior, public opinion, elections, governance, and political communication. It draws on techniques from statistics, data science, computer science, survey methodology, and social science to answer questions like:
- What does the public think about a given issue, and how is that opinion distributed across demographic groups?
- Who is likely to vote in a given election, and for whom?
- What factors predict election outcomes: the economy, candidate quality, campaign spending, something else?
- How do political messages spread through media ecosystems, and what effects do they have on public opinion?
- Are political institutions---parties, legislatures, courts---becoming more polarized, and if so, why?
- How can campaigns most efficiently allocate their resources to maximize their probability of winning?
Political analytics is not a single discipline. It is a meeting ground where political science, statistics, journalism, computer science, and practical campaign work converge. The best political analysts combine deep substantive knowledge of politics with technical skill in data analysis and a healthy skepticism about their own conclusions.
Analytics vs. Punditry
One of the most important distinctions you will learn in this book is the difference between analytics and punditry. Both involve making claims about politics, but they do so in fundamentally different ways.
Punditry deals in narratives, impressions, and predictions offered with confidence. "Garza is in trouble because she is not connecting with suburban voters." "Whitfield's populist message is resonating." "This race is going to be close." These statements may be true, but they are typically offered without rigorous evidence, without quantified uncertainty, and without clear criteria for being proven wrong.
Analytics deals in evidence, models, and estimates offered with quantified uncertainty. "Our model estimates Garza's vote share at 48.3 percent, with a 95 percent confidence interval of 45.1 to 51.5 percent." "Whitfield's favorability among non-college white voters is 12 points higher than among college-educated white voters, based on a sample of 800 likely voters with a margin of error of plus or minus 3.5 points." "Our turnout model predicts 2.1 million votes cast, with an uncertainty range of 1.9 to 2.3 million."
Notice the difference. Analytics does not claim certainty. It quantifies uncertainty. It tells you not just what it thinks will happen, but how confident it is, and why. This is harder, less dramatic, and far more useful.
💡 Intuition: A pundit says "Garza is going to win." An analyst says "Our model gives Garza a 62 percent chance of winning, which means Whitfield wins in roughly four out of every ten simulations." The pundit sounds more confident; the analyst is more honest. Learning to think like the analyst is one of the main goals of this book.
The Analyst's Toolkit
What does a political analyst actually do day to day? The specific tasks vary by context---a campaign analyst does different work from an academic researcher or a data journalist---but the core toolkit is remarkably consistent:
Data acquisition and management. Before you can analyze anything, you need data. This means knowing where to find it, how to access it, how to clean and merge it, and how to store it responsibly. You will develop these skills starting in Chapter 3, when we map the political data ecosystem, and Chapter 5, when you work with your first political dataset in Python.
Descriptive analysis. The most underrated skill in analytics is simply describing what the data shows, clearly and accurately. What is the racial composition of registered voters in a given district? How has turnout changed over the last four election cycles? What percentage of campaign spending goes to digital advertising versus television? Descriptive analysis sounds simple, but doing it well requires precision, honesty, and an awareness of what the data can and cannot tell you.
Statistical modeling. This is what most people think of when they hear "analytics": building models that capture relationships in data and use those relationships to estimate or predict outcomes. You will learn about survey weighting (Chapter 8), regression models (Chapter 18), probabilistic forecasting (Chapter 19), and more.
Visualization and communication. An analysis that no one can understand is an analysis that does not matter. Political analysts must be able to present their findings in clear, compelling ways---through charts, maps, dashboards, and written narratives. You will practice these skills throughout the book, especially in Chapters 16 and 33.
Critical evaluation. Perhaps the most important skill of all: the ability to look at someone else's analysis---a poll, a forecast, a campaign's claims about its data operation---and assess its quality. Is the methodology sound? Are the assumptions reasonable? What are the limitations? What questions should you ask before trusting the results? This is what separates a consumer of political data from a critic of political data, and it is what this book will teach you to be.
1.7 Data in Democracy: Tool or Weapon?
Let us return to our running examples and consider a question that will shadow every chapter of this book: is political data a tool for democracy or a weapon against it?
The Optimistic Case
Consider what political data makes possible. Campaigns can identify voters who have never been contacted by a political organization and reach them with information about registration and voting. Journalists can use data to hold politicians accountable---tracking their votes, their donors, their promises versus their actions. Civic organizations can map disparities in government services and give communities the evidence they need to demand change. Academic researchers can study political behavior with a rigor and scale that was impossible a generation ago.
In the Garza-Whitfield race, Nadia Osei's analytics operation could help the Garza campaign reach voters---particularly low-propensity voters of color---who might otherwise be ignored. At Meridian Research Group, Vivian Park's polls provide the public with independent measurements of the race that do not depend on what either campaign wants people to believe. At ODA, Adaeze Nwosu's open-source tools give ordinary citizens the ability to look up campaign finance records, explore demographic data, and make informed choices.
This is the optimistic case: political data as a tool for inclusion, accountability, and informed citizenship.
The Pessimistic Case
Now consider the darker possibilities. Campaigns can use data to manipulate voters---targeting them with messages designed to exploit their fears, suppress their turnout, or mislead them about candidates and policies. The same voter file that Nadia uses to reach underserved communities can be used to identify and target vulnerable voters with disinformation. Polling data can be weaponized---leaked selectively to shape media narratives, used to create bandwagon effects, or cited misleadingly by partisan actors.
The Cambridge Analytica scandal of 2018, in which a political consulting firm harvested Facebook data from millions of users without their consent, illustrated the potential for abuse. So did the proliferation of "push polls"---fake surveys designed not to measure opinion but to change it by asking leading questions. So did the use of microtargeted political advertising to send different, sometimes contradictory, messages to different voter segments, making it difficult for journalists or opponents to hold candidates accountable for their claims.
⚖️ Ethical Analysis: The same dataset can be used to empower voters or to manipulate them. The same model can be used to identify underserved communities or to exploit vulnerable ones. The ethics of political analytics are not determined by the data or the technique but by the intent, the transparency, and the accountability of the people who use them. We will explore these questions in depth in Chapters 38 and 39.
The Complicated Truth
The truth, as you will discover throughout this book, is that political data is neither inherently good nor inherently bad. It is a tool whose effects depend on who uses it, how, and for what purpose. But it is not a neutral tool, either. Data infrastructures have built-in biases. They amplify some voices and silence others. They make some things visible and other things invisible. And the people who build and control those infrastructures wield a kind of power that is rarely acknowledged and even more rarely regulated.
This is one of the reasons we emphasize the theme Who Gets Counted, Who Gets Heard throughout this book. Political data does not come from nowhere. It is produced by institutions---the Census Bureau, state election offices, commercial data brokers, social media platforms---that make choices about categories, definitions, and access. Those choices have political consequences.
For example: if a state's voter file categorizes ethnicity using broad labels like "Hispanic," it obscures the significant political differences between Cuban Americans in Miami, Mexican Americans in Phoenix, and Puerto Rican voters in Philadelphia. If a poll's sample underrepresents young voters or voters without landline telephones, its results will skew toward older, whiter, more conservative respondents. If a civic data tool is available only in English, it excludes the very communities it claims to serve.
Seeing these choices---making them visible, questioning them, and thinking about alternatives---is a core skill of political analytics. It is also, we would argue, a civic responsibility.
1.8 The Analytical Stakes of the Garza-Whitfield Race
Let us bring this discussion back to earth by examining the specific analytical challenges that the Garza-Whitfield race presents. These challenges will recur throughout the book, and understanding them now will give you a framework for everything that follows.
Challenge 1: Measuring a Changing Electorate
The state where the Garza-Whitfield race takes place is demographically diverse and rapidly changing. Its Hispanic/Latino population has grown by 15 percent in the last decade, driven by both immigration and natural increase. Its Asian American population has grown even faster. Its white non-Hispanic population has shrunk as a share of the total, even as it remains the single largest group.
These demographic shifts create analytical challenges. Voter files and polling samples that were reasonably representative four years ago may not be representative today. Models trained on past elections may not capture the behavior of newly registered voters or newly naturalized citizens. And the political assumptions embedded in old models---about how Hispanic voters behave, for example, or about the relationship between education and partisanship---may no longer hold.
Nadia Osei is acutely aware of these challenges. In one of her first briefings to the Garza campaign, she puts it bluntly: "Every model we build is a bet on who shows up. If the electorate looks like 2020, we win by three. If it looks like 2018, we win by one. If it looks like something we have never seen before---and it might---then all our models are guesses."
Challenge 2: The Polling Problem
Meridian Research Group faces its own version of this challenge. Dr. Vivian Park knows that public polls of competitive races serve an important democratic function: they give voters, journalists, and candidates independent information about where the race stands. But she also knows that polls have been struggling.
Response rates to telephone surveys have plummeted from roughly 35 percent in the late 1990s to less than 5 percent by the mid-2020s. This means that the people who answer polls are increasingly unrepresentative of the broader electorate. Pollsters use statistical weighting to correct for this, but weighting works only if you know the right variables to weight on and the right targets to weight to. In recent elections, some polls have underestimated support for Republican candidates, possibly because conservative voters were less likely to participate in surveys. Other polls have overestimated Democratic strength among young voters and voters of color by assuming turnout levels that did not materialize.
Vivian's challenge is methodological: how do you produce an accurate measurement of public opinion when the people you can reach are systematically different from the people you cannot? This is a question that will occupy much of Part II of this book (Chapters 6 through 10).
Challenge 3: The Transparency Deficit
For OpenDemocracy Analytics, the challenge is different. Adaeze Nwosu and Sam Harding want to make the data surrounding the Garza-Whitfield race accessible to ordinary citizens. But they face a transparency deficit: much of the most important data is either proprietary (campaign voter files, internal polls), legally restricted (certain government records), or technically inaccessible (requiring programming skills or specialized software to use).
ODA's response is to build tools that bridge this gap. They maintain a publicly accessible database of campaign finance records, a voter information lookup tool, and a series of data visualizations that track the race. They also publish explainers that help non-specialists understand polling methodology, election forecasting, and campaign strategy.
But Adaeze worries about reach. "We can build the best tools in the world," she tells Sam during a staff meeting, "but if the only people who use them are already engaged, already informed, already empowered, then we are just reinforcing the information gap instead of closing it."
🔵 Debate: Consider Adaeze's concern. Is it the responsibility of civic tech organizations to reach disengaged citizens, or is it enough to make tools available for those who seek them out? What are the practical and ethical implications of each position? There is no easy answer, and reasonable people disagree.
Challenge 4: Prediction vs. Explanation
There is a deep tension in political analytics between two goals: prediction and explanation. Prediction asks, "What will happen?" Explanation asks, "Why does it happen?"
These goals are not always aligned. A model that is excellent at predicting election outcomes might be terrible at explaining why voters behave the way they do. A model that beautifully explains the relationship between economic conditions and incumbent party performance might fail to predict a specific election because of factors the model does not include.
In the Garza-Whitfield race, this tension plays out concretely. Nadia Osei needs predictive models: she needs to know which precincts to target, which voters to contact, where to spend advertising dollars. She does not necessarily need to know why a particular voter is persuadable; she just needs to know that the voter is persuadable.
But for Vivian Park at Meridian, explanation matters more. When she publishes a poll showing Garza ahead by two points, she wants to be able to tell the public why---which voter groups are driving the lead, what issues are most salient, how the race has shifted and for what reasons. Explanation gives her polls meaning beyond a horse-race number.
And for Adaeze Nwosu at ODA, both matter. She wants to predict which communities are at risk of low voter turnout (so ODA can target its outreach), but she also wants to explain the systemic factors---registration barriers, polling place closures, information deserts---that produce low turnout in the first place.
We will return to this tension throughout the book, especially in Part IV (Chapters 17 through 22), where you will build your own election forecasting model.
Challenge 5: The Map and the Territory
There is a famous saying in statistics, borrowed from the philosopher Alfred Korzybski: "The map is not the territory." A map is a simplified representation of reality, useful precisely because it leaves things out. A map that included every detail of the territory would be as large and unwieldy as the territory itself, and therefore useless.
Political data is a map of political reality. Polls are maps of public opinion. Voter files are maps of the electorate. Election models are maps of the electoral landscape. Each of these maps is useful, but each leaves things out. The poll does not capture the voter who refused to answer the phone. The voter file does not capture the eligible citizen who never registered. The model does not capture the October surprise, the last-minute scandal, the snowstorm on Election Day that suppresses turnout in one region.
The danger arises when people mistake the map for the territory---when they treat the poll as a perfect reflection of public opinion, the model as a certain prediction, the data as a complete picture. This mistake is called reification: treating an abstraction as if it were a concrete, complete, and unambiguous reality.
Every character in our running examples faces this temptation. Nadia Osei must remind her team that voter contact models are estimates, not certainties---that a voter with a "72 percent probability of supporting Garza" might well vote for Whitfield. Vivian Park must ensure that Meridian's polls are reported with appropriate caveats, not as definitive statements of fact. Adaeze Nwosu must design ODA's tools in ways that communicate uncertainty rather than false precision.
And you, as a student of political analytics, must cultivate the habit of always asking: What is this map leaving out? What would I see if I looked at the territory directly? How confident should I be that the map is accurate?
🔗 Connection: The distinction between prediction and explanation is one of the deepest in social science. In Chapter 4, you will learn to think about it more formally. In Chapter 18, you will encounter "fundamentals models" that prioritize explanation. In Chapter 19, you will encounter probabilistic forecasting models that prioritize prediction. And in Chapter 20, you will study cases where models of both types failed---and what those failures teach us.
1.9 How This Book Is Organized
This book is divided into nine parts:
Part I: Foundations (Chapters 1-5) introduces the field, its history, the data ecosystem, the analytical mindset, and your first hands-on work with political data in Python.
Part II: Public Opinion and Polling (Chapters 6-10) covers the theory of public opinion, survey design, sampling, data collection, and how to read and evaluate polls critically.
Part III: Voter Behavior and the Electorate (Chapters 11-16) explores theories of voting behavior, partisanship, demographics, turnout, campaign effects, and tools for visualizing the electorate.
Part IV: Election Forecasting (Chapters 17-22) teaches you how to aggregate polls, build fundamentals models, think probabilistically, learn from model failures, and construct your own simple election forecast.
Part V: Media, Communication, and Information (Chapters 23-27) examines the media ecosystem, theories of persuasion, political advertising, misinformation, and computational text analysis.
Part VI: Campaigns and Strategy (Chapters 28-33) takes you inside the modern data-driven campaign, covering voter targeting, field experiments, digital campaigning, opposition research, and campaign dashboards.
Part VII: Populism, Movements, and Money (Chapters 34-37) broadens the lens to explore how we measure populism, study social movements, track political money, and analyze political rhetoric computationally.
Part VIII: Ethics, Equity, and the Future (Chapters 38-41) addresses the ethical dimensions of political analytics, questions of race and data justice, the impact of artificial intelligence, and career paths in the field.
Part IX: Capstone Projects (Chapters 42-44) brings everything together in three comprehensive projects that integrate the skills and knowledge from the entire book.
Each chapter includes exercises at three difficulty levels, a quiz, two case studies, key takeaways, and an annotated further reading list. Python-based chapters include working code that you can run and modify.
The book is designed to be read sequentially, but it can also be used selectively. If you are primarily interested in polling and public opinion, focus on Parts I and II. If you are interested in campaigns and strategy, focus on Parts I, III, and VI. If you are interested in election forecasting, focus on Parts I, II, and IV. And if you are interested in the ethical and social implications of political data, Parts I, VII, and VIII will be most relevant.
Throughout the book, we will use a consistent set of conventions. Running examples are introduced in bold when they first appear. Key terms are also bolded at first use and defined in context. Cross-references to other chapters are provided in Connection callout boxes. Python code appears in code blocks with comments explaining each step. And callout boxes provide intuitions, real-world applications, common pitfalls, best practices, debates, ethical analyses, critical thinking prompts, global perspectives, and hands-on activities.
1.10 What You Will Need
Technical Requirements
You do not need any programming experience to begin this book. Chapters 1 through 4 are conceptual. Chapter 5 introduces Python for political data analysis, and subsequent Python chapters (10, 16, 21, 27, 33, 37) build on that foundation gradually.
When you reach the Python chapters, you will need:
- Python 3.9 or later
- Jupyter Notebook or JupyterLab
- The pandas, matplotlib, numpy, and statsmodels libraries (all free and open source)
- A willingness to make mistakes and debug your code
Intellectual Requirements
More importantly, you will need:
- Curiosity about how politics works beneath the surface
- Skepticism about claims made with data, including claims made in this book
- Patience with uncertainty, ambiguity, and the limits of what data can tell us
- Ethical awareness of the consequences of analytical choices
- Humility about what you know and do not know
Political analytics is a field where overconfidence is the most dangerous tendency. The analysts who make the worst mistakes are usually the ones who are most certain they are right. The best analysts are the ones who hold their conclusions loosely, update them frequently, and never forget that the people they are modeling are not data points but human beings with complex, irreducible lives.
✅ Best Practice: Throughout this book, we will emphasize a practice we call "calibrated confidence"---being as confident as the evidence warrants, and no more. If a model gives a candidate a 55 percent chance of winning, that means the candidate is a slight favorite, not a sure thing. If a poll shows a three-point lead with a four-point margin of error, that means the race is a toss-up, not a clear victory. Learning to resist the temptation to overinterpret is one of the most valuable skills you can develop.
1.11 The Stakes of Getting It Right---and Getting It Wrong
Before we close this chapter, let us be explicit about why political analytics matters beyond the intellectual satisfaction of understanding politics more deeply.
When political analytics works well, it can strengthen democracy. Accurate polls give citizens reliable information about where candidates stand and how the public feels about issues. Effective voter targeting can bring previously ignored communities into the political process. Transparent data tools can empower citizens to hold their representatives accountable. Rigorous election forecasting can help media organizations allocate their coverage resources wisely and set realistic expectations for election outcomes.
When political analytics fails, the consequences can be severe. Inaccurate polls can create false narratives about the state of a race, leading media organizations to devote disproportionate coverage to a "frontrunner" who is not actually ahead, or causing donors and volunteers to abandon a candidate who is more competitive than the data suggests. Flawed voter targeting can cause campaigns to ignore communities that should be part of their coalition, or to waste resources on voters who were never persuadable. Misleading forecasts can suppress voter turnout by creating a sense of inevitability (why vote if the outcome is predetermined?) or a sense of futility (why vote if your candidate has no chance?).
Consider a concrete example from the Garza-Whitfield race. If Meridian Research Group publishes a poll showing Garza with a comfortable seven-point lead---but the poll's sample underrepresents rural voters and does not weight by education---the consequences cascade through the system. Media coverage shifts to treat the race as less competitive than it is. Democratic donors redirect their money to other races. Garza volunteers become complacent. Meanwhile, Whitfield's team, which has access to better internal data, knows the race is closer than the public polls suggest and intensifies its efforts. The inaccurate poll has not just described reality incorrectly; it has changed reality, by altering the behavior of campaigns, donors, journalists, and voters.
This is the weight of political analytics. The numbers matter not just as descriptions of the world but as forces that shape it. This is why rigor matters. This is why methodology matters. This is why the difference between good analysis and bad analysis is not an academic distinction but a democratic one.
1.12 Returning to Your List
At the beginning of this chapter, we asked you to write down five data needs for the Garza-Whitfield race. Now that you have read about the three worlds of political data, the analytical challenges, and the characters involved, look at your list again.
- Did you think about voter files and registration data? (Chapter 3 will map this landscape.)
- Did you think about polling? (Chapters 6-10 will teach you how polls are designed, conducted, and evaluated.)
- Did you think about demographic data? (Chapter 13 will explore demographics and the electorate in depth.)
- Did you think about campaign finance? (Chapter 36 covers money in politics.)
- Did you think about media coverage? (Chapter 23 examines the media ecosystem.)
- Did you think about social media? (Chapters 26 and 27 address misinformation and computational text analysis.)
- Did you think about the data you cannot get? The private polls, the internal campaign analytics, the conversations that happen off the record? Being aware of the data you lack is as important as knowing how to use the data you have.
Keep your list. Add to it as you read. By the end of this book, you will know how to acquire, analyze, and interpret every type of data on it---and you will understand why each one matters, and what each one can and cannot tell you.
Chapter Summary
This chapter has introduced you to the age of political data---an era of unprecedented volume, velocity, and variety in the information available about political behavior, public opinion, and electoral competition. You have met the three running examples that will anchor this book: the Garza-Whitfield Senate race, with its diverse electorate and contrasting campaigns; Meridian Research Group, the nonpartisan polling firm navigating a crisis in survey methodology; and OpenDemocracy Analytics, the civic tech nonprofit trying to make political data accessible to all.
You have learned that political analytics sits at the intersection of political science, statistics, journalism, and practical campaign work. You have encountered the distinction between analytics and punditry, and the importance of quantifying uncertainty. You have begun to grapple with the ethical tensions inherent in data-driven politics---the ways that data can empower and manipulate, include and exclude, illuminate and obscure.
Most importantly, you have encountered the themes that will run through every chapter of this book: that measurement shapes reality, that who gets counted matters as much as what gets counted, that prediction and explanation serve different purposes, and that political data is never neutral---it is always shaped by the choices and interests of the people who produce and use it.
The age of political data is here. The question is not whether data will shape politics---it already does, profoundly and irreversibly. The question is whether you will be a critical, informed, ethical participant in that process, or a passive consumer of other people's numbers.
Let us make sure it is the former.
In Chapter 2, we will travel back in time to trace the history of polling and political measurement, from straw polls at county fairs to the sophisticated methodological debates of the twenty-first century. Along the way, you will meet Dr. Vivian Park and learn how her journey from academic researcher to polling entrepreneur shaped the philosophy of Meridian Research Group.