50 min read

> "I wasn't searching for different things. The algorithm inferred different things about me and started feeding me different realities."

Chapter 17: Algorithms, the Attention Economy, and Filter Bubbles

"I wasn't searching for different things. The algorithm inferred different things about me and started feeding me different realities."

— Ingrid Larsen, Hartwell University Seminar on Propaganda and Persuasion


Learning Objectives

By the end of this chapter, students will be able to:

  1. Define the attention economy and explain why its structure systematically advantages propaganda over accurate information.
  2. Describe the basic mechanics of recommendation algorithms and explain why engagement optimization produces outcomes that diverge from information health.
  3. Distinguish between filter bubbles (algorithmically created) and echo chambers (socially maintained) and evaluate the empirical evidence for each.
  4. Analyze the documented radicalization pathway on YouTube and explain the structural mechanism by which engagement optimization produces escalation.
  5. Summarize the key findings of the Frances Haugen disclosures and their implications for the claim that social media platforms are passive conduits of user-generated content.
  6. Evaluate the three major regulatory approaches to algorithmic design (no regulation, structural regulation, and the EU DSA model) using the analytical frameworks developed in earlier chapters.
  7. Conduct a basic audit of their own information environment for algorithmic influence.

Opening: Two Accounts, Two Realities

The seminar room was quieter than usual when Ingrid Larsen pulled up her laptop. It was a Tuesday morning in October, and the windows were streaked with the first real rain of the semester. Professor Marcus Webb had asked the class to come prepared with observations from their own digital environments — anything that struck them as relevant to the previous week's discussion of social media architecture.

Ingrid had something specific.

"I ran an experiment," she said, angling her screen so the students nearest her could see. "Over the summer, before I came here. I was thinking about filter bubbles — I'd read Pariser's book — and I wanted to test whether what he described was actually happening in practice, not just in theory."

She had created two Google accounts. Both were new, with no prior search history, no YouTube watch history, no Gmail data — completely clean. She accessed one account from her apartment in Copenhagen. The other she accessed through a U.S.-based VPN, which masked her actual location and assigned her an American IP address.

Over the course of two weeks, she ran identical searches on both accounts. The same political keywords: "immigration reform," "climate policy," "election fraud," "gun control." She watched the same categories of news videos on both — mainstream outlets, nothing extreme. She did not click on partisan opinion content. She made the behavior as parallel as she could make it.

"By the end of the first week," she said, "the recommendation panels were already diverging. Different news sources were being surfaced. Different political frames for the same stories. By the end of the second week, the two accounts were living in different information universes. The Denmark account was being fed a lot of European policy coverage, centrist to center-left in its framing. The U.S. VPN account had drifted pretty hard into partisan American content — specifically, quite a bit of conservative populist material."

She paused. "I hadn't searched for conservative populist material. I hadn't clicked on it. The algorithm inferred something from the combination of my apparent location and my search terms and started feeding me a constructed reality."

Professor Webb leaned back in his chair. He had been an investigative journalist for twenty years before moving into academia, and he had a habit of letting silences sit long enough to be useful.

"What does that tell you?" he said finally.

"That the algorithm is making editorial decisions," Ingrid said. "Constantly. And most users don't know it."

Tariq Hassan, seated two chairs down from Ingrid, tapped his pen on the table. "But it's not making those decisions randomly. It's optimizing for something. The question is: what is it optimizing for, and in whose interest?"

Webb nodded slowly. "That," he said, "is exactly the right question. And answering it is going to require us to understand something that most people who talk about filter bubbles and disinformation don't spend enough time on: the underlying economic architecture of the platforms themselves."

He picked up a marker and wrote two words on the whiteboard:

ATTENTION ECONOMY

"Everything else," he said, "follows from this."


Section 1: The Attention Economy

Herbert Simon's Foundational Insight

In 1971, the economist and cognitive psychologist Herbert Simon published a short paper that has become, in retrospect, one of the most prescient analytical observations of the twentieth century. The paper was about organizational decision-making, not media, but its core insight has shaped how scholars understand the information environment of the twenty-first century.

Simon observed that "a wealth of information creates a poverty of attention." He was identifying a structural shift: as information becomes more abundant, the bottleneck in human cognition and decision-making shifts from information access to attention allocation. When information is scarce, the problem is getting enough of it. When information is abundant — as it has been, with increasing intensity, since the printing press, through broadcasting, and into the internet age — the problem becomes managing the flood. Attention, not information, is the scarce resource.

This observation has profound implications that Simon himself only partially developed. If attention is the scarce resource, then any system designed to deliver information to people is really in the business of competing for attention. The design of that system — how it allocates, concentrates, and directs human attention — becomes enormously consequential for what people believe, what they act on, and ultimately for what is politically and socially possible.

Tim Wu and the History of Attention Merchants

Tim Wu's 2016 book The Attention Merchants traces the history of this competition across more than a century. Wu's argument is that advertising-supported media has always been, at its core, in the business of harvesting human attention and selling it to advertisers. The content — the news, the entertainment, the information — is the bait. The product being sold is the audience's time and mental bandwidth.

Wu traces this dynamic from the first mass-circulation newspapers of the 1830s (which discovered that cheap or free newspapers, funded by advertising, could build enormous audiences) through radio broadcasting, television, and finally to the internet. Each new medium represented a more efficient mechanism for capturing attention, and each brought with it a new wave of propaganda and advertising techniques calibrated to exploit whatever psychological vulnerabilities the medium enabled.

What changed with the internet, and particularly with social media platforms, was the granularity and dynamism of the attention-capture system. Earlier media captured attention in relatively coarse units: a reader's time with a newspaper, a viewer's evening in front of a television. The internet, with its capacity to track individual user behavior in real time, enabled attention capture at a completely different level of precision.

The Engagement Economy

The specific form the attention economy took on the internet was built around engagement metrics: measurable proxies for attention that could be tracked, aggregated, and sold. Time on site. Clicks. Shares. Comments. Reactions. Likes. These metrics became the currency of the platform economy because they could be measured continuously, attributed to specific pieces of content, and used to price advertising inventory.

The consequence of making engagement metrics the central measure of value is that the entire architecture of the platform — what content gets surfaced, what gets amplified, what gets buried — becomes oriented around maximizing engagement. Platforms are not, in any primary sense, in the information business. They are in the attention business. Information is only valuable to them insofar as it captures attention, as measured by engagement metrics.

This is not a conspiracy. It is a structural logic that emerges from the economic incentives of advertising-supported platforms. But the structural consequence is profound: the content that performs best under engagement optimization is not necessarily the content that is most accurate, most important, or most useful. It is the content that triggers engagement responses — and the most reliable triggers for engagement are emotional ones.

The Propaganda Advantage

Here is where the attention economy becomes directly relevant to the study of propaganda: emotional, outrage-generating content captures attention more reliably than accurate, nuanced information.

This is not a guess or an allegation. It is the finding of a substantial body of research. A 2018 MIT study by Soroush Vosoughi, Deb Roy, and Sinan Aral, published in Science, analyzed 126,000 rumor cascades on Twitter over eleven years and found that false news spread significantly faster, farther, and more broadly than true news. False stories were 70% more likely to be retweeted than true stories. The primary driver was not bots but human beings — and the mechanism was novelty and emotional engagement. False stories tended to be more novel and emotionally arousing than true stories, and people were more likely to share content that made them feel surprised, fearful, or disgusted.

The implication for propaganda analysis is stark: the attention economy does not simply provide a neutral channel through which propaganda and accurate information compete on equal terms. It is structurally biased toward the characteristics that propaganda has always exploited — emotional appeal, outrage, novelty, threat, and in-group versus out-group framing. Propaganda, by its nature, is engineered to trigger engagement. Careful, nuanced journalism is not.

Sophia Marin, who had been taking notes rapidly, looked up from her notebook. "So the platforms didn't design a disinformation machine. But they designed a machine that runs on emotional engagement, and propaganda is optimally engineered to exploit emotional engagement. So the outcome is the same."

"That's precisely right," Webb said. "The intent is irrelevant to the structural analysis. What matters is the architecture and what it selects for."


Section 2: How Recommendation Algorithms Work

The Basic Structure

To understand how algorithmic systems shape information environments, it is necessary to understand, at a functional level, how recommendation algorithms operate. The technical details are proprietary and complex, but the basic structural logic is publicly documented and can be described in terms that are analytically useful without requiring expertise in computer science.

Modern recommendation systems typically combine three broad approaches:

Collaborative filtering is the most fundamental approach. It works by identifying patterns across users: people who have engaged with content X have also tended to engage with content Y, so a user who has engaged with X is recommended Y. This creates a vast web of inferred associations based on aggregate user behavior. The recommendation is not based on an analysis of the content itself — the algorithm does not "understand" what either X or Y is about — but on the behavioral patterns of the user population.

Content-based filtering analyzes the characteristics of content a user has engaged with and recommends content with similar characteristics. On YouTube, this might involve analyzing the topics, channels, format, or linguistic features of videos a user has watched and recommending videos with similar attributes. On Facebook, it might involve surfacing posts from accounts whose past content has generated high engagement from a specific user.

Engagement optimization overlays both of these approaches with a core objective function: maximize the engagement metrics that are most valuable to the platform. On YouTube, the primary metric shifted around 2012 from view count to watch time — the total number of minutes users spent watching videos. This shift was consequential: it privileged content that kept users watching for longer over content that attracted clicks but was quickly abandoned. The system learned to recommend content that was compelling enough to hold attention for extended periods.

The Feedback Loop

These approaches combine to create a feedback loop that is central to understanding algorithmic amplification of propaganda.

When a user engages with a piece of content — watches a video, clicks a link, reacts to a post — the algorithm registers that engagement as a signal of interest. It then surfaces more content predicted to generate similar engagement. If the user engages with that content, the signal is reinforced. The algorithm does not ask whether the engagement is healthy, whether the user is enjoying the experience, or whether the content is accurate. It asks: is this user engaging? If yes, serve more like this.

The feedback loop means that the algorithm continuously calibrates its recommendations toward content that generates the strongest engagement signals from that specific user. For users who are susceptible to emotionally provocative content — which, given the MIT study's findings, is most users — the algorithm will tend to surface increasingly emotionally provocative content, because that is what generates the engagement signals it is optimized to maximize.

This is the mechanism behind what journalists and researchers call the "rabbit hole" phenomenon: the observation that users who begin watching relatively mainstream content on recommendation-driven platforms can find themselves, after a series of algorithm-driven recommendations, watching content that is substantially more extreme than what they started with. Each step in the recommendation chain makes sense as an incremental increment — from content A to content B that is marginally more engaging, then to content C that is marginally more engaging than B. But the cumulative drift can be substantial.

The Ribeiro et al. Finding

The rabbit hole phenomenon was systematically analyzed in a 2020 study by Ribeiro et al. that examined YouTube's recommendation algorithm. Using a large-scale dataset of YouTube videos and recommendation pathways, the researchers mapped the actual pathways through which recommendations led users from mainstream content to increasingly extreme content.

The study found systematic pathways from mainstream political content — mainstream news channels, mainstream conservative commentary — through what the researchers called the "alternative influence network" to explicitly white nationalist and far-right extremist content. These were not random or isolated pathways. They were consistent, recurring routes through the recommendation graph that the algorithm had effectively carved through the content landscape.

The mechanism was not that YouTube had intentionally designed a radicalization pathway. The mechanism was that content on the pathway from mainstream to extreme tended to generate higher engagement metrics than the mainstream content users started with — more comments, more watch time, more shares — and so the algorithm, optimizing for engagement, systematically surfaced it.

Why Engagement Optimization Diverges from User Interest

The gap between what engagement optimization maximizes and what users actually want or benefit from is one of the central analytical problems in understanding algorithmic media systems.

There is a well-documented phenomenon in behavioral economics and psychology called present bias: people systematically overweight immediate gratification relative to their own long-term preferences and interests. A person may genuinely prefer to exercise rather than watch television but nonetheless choose television in the immediate moment. Engagement metrics measure immediate behavioral responses — the click, the watch, the reaction — not considered preferences or long-term well-being.

When an algorithm is optimized to maximize immediate engagement responses, it is in effect exploiting present bias systematically. It surfaces content that triggers immediate emotional engagement responses, even when those responses run counter to what users would choose for themselves if they were reflecting carefully on their information diet. This is structurally analogous to what the tobacco industry discovered about nicotine addiction: the product that generates the most immediate repeat consumption is not necessarily the product that serves the consumer's interests, and the most efficient business model involves not asking that question.


Section 3: Filter Bubbles and Echo Chambers

Eli Pariser's Filter Bubble Concept

The term filter bubble was coined by Eli Pariser in his 2011 book The Filter Bubble: What the Internet Is Hiding from You. Pariser's core argument was that algorithmic personalization — the tendency of search engines, social media feeds, and news aggregators to surface content tailored to each individual user's past behavior — was creating information cocoons. Rather than encountering a shared information environment, users were encountering individually tailored environments that reflected and reinforced their existing interests, preferences, and beliefs.

Pariser illustrated the concept with a vivid observation: he had noticed that when he searched on Google, he got different results than his conservative friends searching for the same terms. Not because Google had indexed different content, but because its personalization algorithm was surfacing different results based on their prior search histories and behavioral data. They were, in effect, experiencing different versions of the same internet.

The filter bubble concept resonated powerfully because it provided a structural explanation for a widely observed phenomenon: the sense that people in different political communities were not merely disagreeing about interpretations of shared facts, but appeared to be inhabiting different factual universes. If algorithms were systematically curating each person's information environment to reinforce existing beliefs, this would produce exactly that outcome without requiring any deliberate effort by the individuals involved.

The Empirical Debate

The filter bubble hypothesis proved, on empirical investigation, to be both real and considerably more modest than Pariser's formulation suggested. A series of major academic studies through the 2010s produced a more nuanced picture.

Research by Eytan Bakshy, Solomon Messing, and Lada Adamic at Facebook (2015), published in Science, analyzed the actual distribution of cross-cutting political content in users' Facebook feeds. They found that while the algorithmic ranking did modestly reduce exposure to cross-cutting content, the larger factor was user behavior: people were more likely to click on content that confirmed their existing views even when cross-cutting content was present in their feed. The algorithm created a modest filter bubble, but human choice created a larger one.

Similar findings emerged from studies of Twitter, search engines, and online news consumption. The consistent pattern was that algorithmic curation effects were real but relatively modest — and that they were dwarfed by the selective exposure effects produced by users' own choices about whom to follow, which sources to trust, and which content to engage with.

The Echo Chamber Distinction

This empirical nuance requires a conceptual distinction that is often collapsed in popular discussion: the difference between filter bubbles and echo chambers.

A filter bubble is algorithmically created. It is the result of personalization systems that shape what content users are exposed to based on their behavioral data. Filter bubbles can be narrow or wide; they can be countered by user choices to actively seek out diverse sources; and their magnitude is an empirical question about specific algorithmic systems at specific times.

An echo chamber is socially maintained. It is the result of people's choices about whom to trust, which communities to participate in, and which sources to seek out. Echo chambers exist independently of algorithmic systems — they existed in print media, in radio, in television — and they are driven by deep psychological tendencies toward seeking confirmatory information (a phenomenon psychologists call confirmation bias) and toward forming communities of shared belief.

Both phenomena exist. Both contribute to the fragmentation of shared information environments. And they interact: algorithmic filter bubbles can reinforce echo chambers by reducing the friction involved in information cocoon construction; echo chambers can amplify filter bubble effects by producing strong behavioral signals (consistent engagement with partisan content) that drive algorithmic personalization.

The propaganda implication of this interaction is significant. Even when filter bubble effects are modest, they reduce the probability that propaganda content circulating within a targeted community will be encountered with counter-argument by members of that community. Propaganda is most effective when it operates in an information environment where counter-narratives are absent or low-salience — and the combination of algorithmic and social echo chamber dynamics tends to produce exactly that environment for targeted audiences.

The Bail et al. Counterintuitive Finding

The most counterintuitive and analytically important empirical finding in the filter bubble literature is the 2018 study by Christopher Bail and colleagues, published in the Proceedings of the National Academy of Sciences. Its findings challenge the simplest solution that might seem to follow from filter bubble analysis: just show people more cross-cutting content.

Bail et al. recruited a sample of Twitter users who identified as Republicans or Democrats and paid them to follow bots that retweeted content from elected officials, political journalists, and opinion leaders from the opposing party. The intervention was, by design, exactly what a naive filter bubble theory would prescribe: expose people to more content from the other side and see if their views become more moderate.

The results were the opposite of what the hypothesis predicted. Republicans who followed the liberal bot became more conservative on a series of political attitude measures, not less. Democrats who followed the conservative bot showed a small (statistically non-significant) shift in the liberal direction.

The finding suggests that cross-cutting exposure without relationship or context triggers defensive identity protection rather than persuasion or moderation. When people encounter political content from the opposing side without the relational context that might make it legible or trustworthy, they experience it as a threat to their group identity rather than as information that merits consideration. The response to threat is not open-mindedness; it is entrenchment.

This finding has major implications for both propaganda analysis and for policy responses to filter bubbles. It undermines the simple "more speech" intervention — the idea that exposing people to more diverse information will reduce polarization. It suggests that the problem is not merely one of information exposure but of the social and epistemic context in which information is received. And it raises hard questions about whether algorithmic interventions to increase cross-cutting content exposure would be helpful, neutral, or actively counterproductive.


Section 4: YouTube's Radicalization Pipeline

The Alternative Influence Network

The most extensively documented case study of algorithmic radicalization is the pathway from mainstream political content to far-right extremism on YouTube. The analytical foundation was laid by a 2018 report from Rebecca Lewis at Data & Society, titled "Alternative Influence: Broadcasting the Reactionary Right on YouTube." Lewis mapped a network of 65 YouTube channels and approximately 81 influencers who, despite ranging from mainstream conservatives to explicit white nationalists, formed a densely interconnected content ecosystem.

What Lewis documented was not simply the existence of extremist content on YouTube. It was the structure of a network in which mainstream and moderate-right content creators maintained relationships — through collaborations, cross-promotions, and shared audiences — with progressively more extreme figures. This created pathways through which audiences could move gradually from content that was unambiguously mainstream (political commentary, news analysis) to content that was explicitly white nationalist or accelerationist, with each step seeming like a small increment.

The algorithmic significance of this network structure was identified by Ribeiro et al. in their 2019 study "Auditing Radicalization Pathways on YouTube," published in the proceedings of the ACM Web Science conference. Using a dataset of over 330,000 videos from 360 channels spanning mainstream, alternative, and extreme right categories, Ribeiro et al. mapped the recommendation pathways between these categories.

Their findings documented a systematic tendency for YouTube's recommendation algorithm to recommend content from the alternative influence network to users who had been watching mainstream political content, and subsequently to recommend content from explicitly extremist channels to users who had been watching alternative influence network content. The pathway was not inevitable for every user — many users who encountered recommended alternative content did not follow it — but it was structural and consistent. The algorithm had effectively operationalized the Lewis network map as a set of recommendation pathways.

The Mechanism: What Extremism Offers the Algorithm

The mechanism behind this pathway is analytically important. Extremist and near-extremist content did not succeed on YouTube because YouTube wanted to radicalize its users. It succeeded because extremist content has characteristics that are strongly rewarded by engagement optimization.

Compared to mainstream political commentary, far-right alternative content on YouTube tended to exhibit several engagement-maximizing characteristics:

Higher emotional intensity. Content that frames politics in terms of existential threat — to the nation, to white people, to traditional culture, to individual rights — generates more intense emotional responses than measured policy analysis. Fear and outrage are among the most reliable engagement triggers.

Community and identity. Far-right YouTube content frequently built strong parasocial communities around specific creators, with viewers who watched long hours, commented extensively, and formed social identities around channel membership. This behavior generated exactly the engagement signals YouTube's algorithm was optimized to maximize.

Novelty and counter-narrative framing. Much far-right content explicitly positioned itself as forbidden truth that mainstream media was suppressing. This framing is both engaging (novelty, transgression) and resistant to counter-argument (any counter-argument is framed as evidence of the suppression).

Watch-time optimization. Many far-right creators produced long-form content — videos of one, two, or three hours in length — that was specifically calibrated to maximize watch time. YouTube's shift to watch time as a primary metric directly rewarded this format.

Caleb Cain's Documented Experience

The most publicly documented individual case of YouTube radicalization is that of Caleb Cain, whose experience was reported by Kevin Roose in The New York Times in 2019. Cain, who had experienced depression and social isolation in his early twenties, began watching YouTube videos on self-help and personal development. The algorithm recommended political content — initially mainstream conservative commentary, then, incrementally, content from figures in the alternative influence network, and finally explicitly white nationalist content.

Cain described the experience as gradual and almost imperceptible. Each recommended video seemed like a small step from the last. The content provided a sense of community, of shared identity, of possessing special knowledge that others lacked — precisely the psychosocial needs that Cain's isolation had left unmet. By the time he was watching explicitly white nationalist content and beginning to internalize its worldview, he had traversed a pathway that, viewed in retrospect, was obvious in its architecture but was experienced as a series of small, natural-feeling moves.

What Cain's case illustrates — and what makes it analytically valuable beyond individual anecdote — is the interaction between algorithmic recommendation and psychosocial vulnerability. The algorithm did not create his vulnerability; it exploited it. It surfaced content that addressed unmet needs with increasingly powerful emotional content, and it did so without any mechanism for asking whether that content was accurate, healthy, or in Cain's interest. It asked only whether Cain was watching.

YouTube's Responses and Their Limitations

YouTube implemented a series of policy changes between 2019 and 2022 in response to the documented radicalization pathway:

In January 2019, YouTube announced that it would reduce recommendations of what it called "borderline content" — content that did not violate community guidelines but that "could misinform users in harmful ways." In subsequent months and years, it removed thousands of channels associated with explicit white nationalism and far-right extremism.

These interventions reduced the most visible manifestations of the radicalization pipeline. The explicitly white nationalist channels documented by Ribeiro et al. and Lewis were largely removed from the platform. Studies of recommendation pathways conducted after the 2019 changes found that the pathways to extreme content were less prominent than they had been.

But what remained was the fundamental architecture. YouTube's recommendation system continued to be optimized primarily for engagement metrics. The specific content that had exploited those metrics was removed, but the structural incentive for content creators to produce emotionally intense, community-building, counter-narrative content remained. New channels and new content ecosystems emerged to fill the space left by removed channels, calibrated to the same engagement optimization logic.

The analogy to propaganda control that Webb drew for his class was telling: removing the most egregious propaganda content while leaving the economic and algorithmic architecture that rewards propaganda-like content is analogous to treating symptoms without diagnosing the underlying disease. The specific instances can be addressed; the structural incentive that produces them continuously is not.


Section 5: Facebook's Internal Research

The Frances Haugen Disclosures

In October 2021, Frances Haugen, a former product manager at Facebook who had worked on its civic integrity team, turned over tens of thousands of pages of internal Facebook documents to the Wall Street Journal, to Congress, and to a consortium of news organizations. The disclosures became known as the "Facebook Files" and constituted the most extensive documented evidence yet assembled about what Facebook's own researchers knew about the harms generated by the platform's design — and what decisions the company made in response.

Haugen's central argument, articulated in Senate testimony on October 5, 2021, was that Facebook's executives had repeatedly been presented with internal research documenting that the platform's algorithmic design generated significant harms — to individual mental health, to civic information quality, to political polarization — and had repeatedly chosen not to make changes to the platform's core design because those changes would reduce engagement and therefore advertising revenue.

"The company's leadership knows how to make Facebook and Instagram safer," Haugen testified, "but won't make the necessary changes because they have put their astronomical profits before people."

Key Findings from the Internal Documents

The internal documents disclosed by Haugen revealed several findings that are directly relevant to propaganda analysis:

The angry reaction and algorithmic amplification. Facebook introduced reaction buttons — Like, Love, Haha, Wow, Sad, Angry — in 2016, expanding beyond the original Like button. Internal research documented that the "angry" reaction generated approximately five times more algorithmic distribution than the "like" reaction. This was a direct result of Facebook's engagement optimization: the algorithm weighted reactions by their predictive value for continued engagement, and anger was a strong predictor of continued engagement. The practical consequence was that content that made users angry was systematically amplified relative to content that made users happy or sad.

This finding has a direct implication for propaganda analysis: Facebook's algorithm had, in effect, built an amplification premium for outrage-generating content. Political content that provoked anger — which, as documented in the propaganda literature, is characteristic of fear-threat framing, enemy construction, and calls to mobilization — was structurally advantaged in Facebook's distribution system over content that was informative but not emotionally provocative.

The 2018 algorithm change. In 2018, Facebook made a major change to its News Feed algorithm, ostensibly to promote "meaningful social interactions" — prioritizing content that generated comments and shares over content that generated passive consumption. Internal documents revealed that Facebook's own researchers had warned that this change would systematically reward "engagement bait" — content that provoked reactions and arguments without necessarily being accurate or valuable. The change was implemented anyway.

The Instagram mental health research. Separate from the civic integrity documents, Haugen disclosed internal Facebook research on Instagram that found the platform caused significant harm to the mental health of teenage girls, particularly around body image. The research found that 32% of teenage girls who reported feeling bad about their bodies said Instagram made them feel worse. When Facebook's own research identified this harm, the company chose not to change Instagram's core design.

The integrity team and business overrides. The documents revealed a consistent pattern: Facebook's integrity research teams would identify a harm generated by a platform design choice, propose a change to mitigate the harm, and be overruled by business teams who argued that the change would reduce engagement. The integrity research was not disputed — it was accepted as accurate and then set aside. The business logic was straightforward: the design choices that generated harm also generated engagement, and engagement generated revenue.

The Big Tobacco Analogy

Ingrid was the one who named the structural parallel directly. "This is what the Documents on tobacco looked like," she said, when Webb walked the seminar through the Haugen disclosures. "Internal researchers document harm. Executives are informed. Executives make a business decision not to change the product. And the external message remains: our product is safe."

The parallel is analytically illuminating. The tobacco industry's internal research in the 1950s through 1970s had established, to the industry's own scientists' satisfaction, that cigarettes caused cancer and were addictive. The internal documentation of this knowledge, combined with the industry's external denial of it, became the basis for major legal liability and regulatory action decades later.

The comparison does not require establishing that Facebook's harm was equivalent in scale or type to tobacco's. It requires only noting the structural similarity: internal research documenting harm, decisions at the executive level not to change the product, and external representations that did not reflect what the internal research showed. This is the pattern of knowledge-and-denial that has characterized institutional harm management across industries, and its emergence in the context of social media platforms is analytically significant.

The Digital Services Act Response

Ingrid, drawing on her European regulatory background, pointed to the legislative response that had proceeded most systematically from the Haugen disclosures: the European Union's Digital Services Act (DSA), which entered into force in 2022.

"This is what the DSA was designed for," she said. "Not to police individual pieces of content — that raises free speech issues that are genuinely complicated. But to require very large platforms to assess whether their design choices create systemic risks to democratic discourse, and to mitigate identified risks. The logic is that Facebook's own research showed the systemic risk. The problem was that there was no external requirement to act on it."

The DSA creates exactly such an external requirement for platforms operating in the EU. It requires very large online platforms to conduct systemic risk assessments, to implement mitigation measures for identified risks, and to allow auditors and researchers access to data necessary to conduct independent assessments. It represents, as of its passage, the most comprehensive regulatory response to the structural harms documented by Haugen and by years of academic research.


Section 6: Microtargeting and Behavioral Advertising

Cambridge Analytica: Claims and Reality

No discussion of algorithmic propaganda infrastructure is complete without addressing Cambridge Analytica, the political data firm whose activities became the focus of intense public and legislative scrutiny following a series of revelations in 2018. But an accurate analysis of Cambridge Analytica requires distinguishing carefully between what the firm claimed to be able to do, what it actually did, and what that actually means for propaganda analysis.

The claims. Cambridge Analytica and its parent company SCL Group marketed themselves on the basis of "psychographic" profiling: the claim that they could use Facebook data to construct detailed psychological profiles of voters based on the "OCEAN" model (Openness, Conscientiousness, Extraversion, Agreeableness, Neuroticism) and use those profiles to deliver customized persuasion messages that exploited specific psychological vulnerabilities. The founder Alexander Nix gave TED-style presentations claiming that this approach represented a revolution in political communication.

The documented reality. Academic researchers who subsequently analyzed Cambridge Analytica's actual methodologies found them to be substantially less revolutionary than claimed. The OCEAN-based psychographic targeting at scale relied on a 2013 academic study whose findings were subsequently found to be overstated and difficult to replicate. Christopher Wylie, the firm's former director of research who became a whistleblower, acknowledged that the psychographic claims were substantially inflated for marketing purposes.

What Cambridge Analytica actually did, and what was documented, was more conventional but not trivial: they improperly obtained Facebook data on approximately 87 million users through a third-party application, they built voter profiles using this data and commercial data, and they conducted conventional political micro-targeting on behalf of clients including the Trump 2016 campaign and the Leave.EU campaign in the Brexit referendum.

Why the "conventional" activities still matter. The gap between Cambridge Analytica's inflated claims and its documented activities should not produce complacency. What behavioral micro-targeting can actually do — even without exotic psychographic profiling — is analytically significant. A political campaign using behavioral data can show different political messages to different demographic and psychographic segments simultaneously, with no mechanism for journalists, opponents, or regulators to observe that different messages are being shown to different audiences.

This is what researchers call the "dark ad" problem: digital political advertising can target specific messages to specific micro-segments with no visibility to anyone outside the targeted segment. A campaign can tell one demographic group one thing and tell a different demographic group something that contradicts the first message. This is structurally impossible in broadcast media (where all viewers see the same advertisement) and was impossible in pre-digital direct mail (where it was too expensive to produce individualized messaging at scale). Digital behavioral advertising makes it not only possible but routine.

Dog-Whistle Targeting and Message Fragmentation

The political implications of micro-targeted messaging extend beyond simply showing different versions of the same message to different audiences. They enable what analysts call "dog-whistle targeting" at a much finer grain than was previously possible.

A dog whistle is a political message that carries one meaning to a general audience and a different, more specific meaning to a target audience. The classic broadcast version involves using code words or cultural references that are recognized by the target audience but appear innocuous to others. Digital micro-targeting enables a precise version of this: literally showing different messages to different audiences, so that the "whistle" is not even visible in the general media environment.

The opacity created by this structure is itself a propaganda mechanism. Political actors can make commitments to specific constituencies that are invisible to other constituencies. They can mobilize fear in one community while presenting a moderate face to a different community. The disaggregation of the political message — previously an organic limitation on political communication — becomes a tactical capability.

Sophia framed this in terms the class had developed earlier in the semester. "It makes the audience construction process algorithmic and invisible. You're not just targeting a message to an audience — you're simultaneously constructing different audiences for different versions of the message, and none of those audiences has access to what the other audiences are seeing. It's the most fragmented version of the medium-is-the-message dynamic we've discussed."


Section 7: Research Breakdown — Bail et al. (2018)

"Exposure to Opposing Views on Social Media Can Increase Political Polarization"

Citation: Bail, C. A., Argyle, L. P., Brown, T. W., Bumpus, J. P., Chen, H., Hunzaker, M. B. F., Lee, J., Mann, M., Merhout, F., & Volfovsky, A. (2018). Exposure to opposing views on social media can increase political polarization. Proceedings of the National Academy of Sciences, 115(37), 9216–9221.

Background and Motivation

By 2018, the filter bubble hypothesis had become widely accepted in public discourse as an explanation for political polarization, and its apparent policy implication — that platforms should expose users to more content from across the political spectrum — had gained traction among policymakers and platform executives. Bail et al. designed their study specifically to test whether this intervention logic was correct.

Study Design

The study recruited a sample of 1,652 Twitter users who described themselves as Republicans (901) or Democrats (751) and who used Twitter at least three times per week. Participants were randomly assigned to one of two conditions:

  • Experimental condition: Participants were paid $11 per month to follow a Twitter bot that retweeted content from elected officials, opinion leaders, and media figures associated with the opposing political party. The bots were designed to appear as real Twitter accounts.
  • Control condition: Participants were paid the same amount but did not follow any bot.

Before and after the one-month intervention, participants completed surveys measuring their political views on a range of issues. The difference in attitude change between experimental and control conditions was the measure of the intervention's effect.

Findings

The results contradicted the filter bubble hypothesis's policy implication:

  • Republicans who followed the liberal bot became more conservative on measured attitude scales, with the effect being statistically significant.
  • Democrats who followed the conservative bot showed a small shift in the more liberal direction, but the effect was not statistically significant.

The study also found that participants' engagement with bot-retweeted content was low — many experimental condition participants did not click on much of the content the bot retweeted. This suggests that mere exposure to cross-cutting content, even when delivered to the user's feed, is not sufficient to generate attitude change, and when it does generate change, the direction may be backfire rather than moderation.

Interpretation and Significance

Bail et al. interpreted their findings in terms of social identity theory. When people encounter political content from the opposing party without any relational or contextual scaffolding, they tend to experience it not as information to be evaluated but as a marker of out-group identity. The appropriate response, according to social identity dynamics, is not to evaluate the content on its merits but to signal in-group loyalty by opposing it. The result is entrenchment rather than persuasion.

This finding has three major implications for propaganda analysis and counter-propaganda strategy:

First, it shows that information exposure alone does not produce the outcomes that simple information deficit models predict. People are not passive receivers of information who update their beliefs when presented with new data; they are social actors who evaluate information in terms of its social identity implications.

Second, it suggests that counter-propaganda strategies built on simply showing people "the other side" are likely to be ineffective and may be actively counterproductive. The relational and epistemic context in which information is received matters as much as the content of the information itself.

Third, it raises fundamental questions about what algorithmic interventions to "reduce polarization" by increasing exposure to cross-cutting content would actually accomplish. The Bail et al. finding suggests that such interventions, if they produced any attitude change at all, might increase polarization rather than reduce it.

Tariq, who had been quietly reading the paper on his laptop, looked up. "So the standard liberal solution — just show people more diverse information — might be exactly wrong. Which means the problem is much harder than filter bubble theorists claimed."

"Or it means," Webb replied, "that the solution is social and relational, not informational. Which is a less comfortable conclusion for people who want to solve this with a platform redesign."


Section 8: Primary Source Analysis — Algorithmic Amplification and January 6

The Event and Its Information Environment

The January 6, 2021 assault on the United States Capitol building by supporters of President Donald Trump attempting to disrupt the certification of the 2020 electoral college results was the most significant act of domestic political violence in the United States since the Civil Rights era. Understanding it requires understanding many things, including the legislative and political context, the specific decisions of political leaders, and the role of security failures. But it also requires understanding the information environment in which the events were organized and the beliefs that animated participants — and that information environment was substantially shaped by algorithmic amplification.

The "Stop the Steal" Ecosystem on Facebook

The "Stop the Steal" movement — which promoted the false claim that the 2020 presidential election had been stolen through widespread fraud — originated as a hashtag and informal network but became, through algorithmic amplification, a massive organized presence on Facebook.

Internal Facebook documents disclosed by Haugen and examined by congressional investigators showed that the Stop the Steal Facebook group, created on November 3, 2020, the day of the election, reached 300,000 members within twenty-four hours before being taken down. But the ecosystem was not limited to a single group. Dozens of groups with similar themes proliferated, and Facebook's recommendation algorithm — the "Groups You May Like" and "Pages You May Like" features — systematically recommended them to users who had shown interest in related content.

The mechanism was the same engagement optimization logic documented throughout this chapter. Stop the Steal content was emotionally intense — it invoked stolen elections, betrayed patriots, imminent catastrophe — and it generated enormous engagement. Users who reacted with anger or shared the content were algorithmically recommended more content from the same ecosystem. The recommendation system, optimizing for engagement, effectively functioned as a recruitment and amplification infrastructure for the movement.

Applying the Five-Part Anatomy of Propaganda

Consider a representative viral post from the Stop the Steal ecosystem: a meme widely circulated in late November and December 2020 that showed graphs purportedly showing vote-count anomalies in Michigan at 4 AM on November 4, with text reading "You cannot explain this. This is fraud. Share before they delete it."

Source and Credibility Claim: The post does not identify its source, which is itself a rhetorical move — the claim is presented as self-evident, requiring no authority beyond the visual data it presents. The "share before they delete it" imperative implies suppression and makes the sharer feel like a participant in resisting censorship.

Content: The graphs were based on real vote count data but misrepresented the context — late-night counting spikes in mail-in ballots were predictable and explained by the specific protocols different counties used for counting, which were publicly documented. The "anomaly" was an artifact of how data was reported, not of fraud.

Framing and Emotional Appeal: The framing invokes irreversibility ("stolen"), certainty ("you cannot explain this"), and mobilization ("share"). The emotional register is outrage and urgency — both strong engagement triggers. The us-versus-them structure is embedded in the claim of suppression: the enemy is not just the alleged fraudsters but the platforms that might "delete it."

Distribution and Amplification: The algorithmic dynamics documented above meant that this post, and thousands like it, were systematically recommended to users who had shown engagement with related content. Users who had been engaged with election coverage, with conservative political content, and with content questioning mail-in voting were algorithmically positioned to receive exactly this content.

The January 6 Committee's Assessment: The House Select Committee to Investigate the January 6th Attack on the United States Capitol found, in its final report, that the "Stop the Steal" movement was "a critical component of President Trump's plan to overturn the election results" and that social media platforms "played a key role in spreading and amplifying false claims about the 2020 election." The committee's findings on social media documented that platforms had failed to act on their own internal research identifying the spread of election misinformation, and that algorithmic amplification had substantially accelerated the movement's growth.


Section 9: Debate Framework — Should Algorithm Design Be Regulated?

The question of how democratic societies should respond to the structural harms documented in this chapter is among the most contested in contemporary technology policy. Three principal positions can be articulated and evaluated.

Position A: No Algorithmic Regulation

Core claim: Algorithmic design is a private business decision. When a platform chooses how to rank, recommend, and amplify content, it is making an editorial decision. Government regulation of algorithmic design is therefore government regulation of editorial decisions — a direct intrusion into the freedom of the press and freedom of speech. The appropriate remedy for harms created by platform design choices is not regulation but competition: users who dislike a platform's algorithmic choices can choose a different platform.

Strongest version: This position has genuine constitutional grounding in the United States, where the First Amendment has historically been interpreted to prevent government regulation of editorial decisions. The government does not have the authority to tell a newspaper which stories to run on page one; by analogy, it does not have the authority to tell a social media platform which posts to surface in a user's feed. The "marketplace of ideas" argument holds that diverse platforms with diverse algorithmic approaches will produce a healthier overall information environment than any regulated standard.

Key weaknesses: The analogy to editorial freedom breaks down when applied to the specific case of engagement optimization. A newspaper's editorial choices are made by human editors with professional norms, legal liability for defamation, and accountability to their readership. Algorithmic ranking has none of these properties — it is a mechanical optimization process that selects content based on engagement metrics without any editor's judgment, professional norms, or accountability. The "competition" argument is weakened by network effects that produce near-monopoly conditions in social media markets: for most users, "choosing a different platform" means leaving the communities they have built over years.

Position B: Structural Regulation

Core claim: Governments should not regulate the specific content decisions of algorithms but should require platforms to offer structural alternatives that give users genuine choice. Specifically: platforms should be required to offer non-algorithmic, chronological feeds as a genuine default option (not buried in settings that most users never access). Additionally, platforms should be required to make the logic of their recommendation algorithms auditable — meaning that independent researchers, with appropriate privacy protections, should be able to analyze how recommendation algorithms work and what their effects are.

Strongest version: Structural regulation does not tell platforms what content to surface; it requires transparency and user choice. Auditing requirements would allow academic researchers to conduct the kind of independent analysis that Ribeiro et al. and others have attempted to conduct without platform cooperation — with much more reliable data. Chronological feed requirements would allow users to opt out of engagement optimization without leaving the platform entirely.

Key weaknesses: Auditing requirements, while analytically appealing, are technically complex. Recommendation algorithms are not transparent systems that can be easily read and interpreted by external researchers; they are trained machine learning models that encode billions of parameters. "Auditing" such a system requires not just access to its code but the ability to run controlled experiments — a form of access that would be genuinely difficult to implement. Chronological feed requirements, meanwhile, may not address the core problem: if a user's social graph is itself the product of algorithmic friend and follow recommendations, a chronological feed of that graph is still algorithmically shaped.

Position C: The EU Digital Services Act Approach

Core claim: Rather than prescribing specific algorithmic designs or prohibiting specific content, the EU DSA requires very large platforms to conduct systemic risk assessments — structured evaluations of whether their design choices, including algorithmic recommendation, create systemic risks to democratic discourse, civic engagement, electoral integrity, or public health — and to implement mitigation measures for identified risks. The requirement is not that platforms produce a specific outcome but that they take identified risks seriously, document their assessments, and submit to independent audit.

Strongest version: This approach responds directly to what the Haugen disclosures revealed: Facebook's own research had identified systemic risks, but there was no external requirement to act on that research. The DSA creates exactly such a requirement. It does not require platforms to make any specific design choice; it requires them to justify their design choices in terms of their systemic effects on democratic information quality. It is analogous to requiring pharmaceutical companies to conduct safety research and disclose findings — not prescribing what drugs can be made, but requiring a systematic process for identifying and addressing harm.

Key weaknesses: The DSA's effectiveness depends on the quality and independence of the risk assessment and audit process. If risk assessments are conducted by platforms themselves and reviewed by regulators who lack the technical capacity to challenge them, the process may become a compliance ritual without substantive effect. The DSA applies only within the EU; platforms that are primarily American businesses may face difficult questions about how to implement EU-specific algorithmic requirements without affecting their global systems.


Section 10: Action Checklist — Evaluating Algorithmic Influence

The following questions are designed to help you assess whether your own information environment is shaped by algorithmic curation in ways that may advantage propaganda. The goal is not to escape algorithms entirely — that is neither possible nor necessarily desirable — but to understand the structure of your information environment with sufficient clarity to exercise conscious agency within it.

1. Platform Audit - Which platforms do you use for political and civic information, and which of those platforms use recommendation algorithms to curate content (as opposed to presenting chronological feeds from accounts you have explicitly chosen to follow)? - Have you ever compared what appears in your algorithmic feed with what you would see in a chronological feed of the same accounts? Most platforms offer the option; the comparison is often illuminating.

2. Feedback Loop Recognition - Can you identify moments when a platform's recommendations seemed to track your emotional state or recent engagement? Note specific instances where content that provoked a strong reaction seemed to produce more similar content. - Have you ever gone down a "rabbit hole" of recommended content? What was the starting point, what was the endpoint, and what was the pathway between them?

3. Diversity Audit - How many of the accounts you follow on each platform represent political or cultural perspectives that differ substantially from your own? (Be honest — "following" someone to argue with them counts differently from following them to genuinely understand their perspective.) - When you encounter a political claim that confirms your existing views, what is your process for evaluating its accuracy before sharing or amplifying it?

4. Source Tracing - When you see a piece of political content in your feed, can you trace it to its original source? Is the original source an account you chose to follow, or did the algorithm surface it independently? How did it get from its origin to your feed?

5. Emotional Signal Awareness - The Haugen documents showed that the "angry" reaction generates 5x more distribution than "like." Before reacting to a piece of content that makes you angry, pause and ask: is this content designed to make me angry? Who benefits from me amplifying this content, and to whom will my amplification distribute it?


Section 11: Inoculation Campaign — Algorithmic Amplification Audit

Channel Audit: Week Five (Chapters 13–18)

This component of your Channel Audit project asks you to investigate whether algorithmic amplification is a factor in the distribution of propaganda or disinformation to your target community.

Step 1: Identify the relevant platforms. For the community you identified in Week One of the Channel Audit, identify which social media platforms are most used and most influential. Different communities have different platform distributions: older voters may be primarily on Facebook; younger audiences may be primarily on TikTok or Instagram; political activists may use Twitter/X disproportionately. Your platform selection should reflect where your target community actually is.

Step 2: Map the algorithmic architecture. For each platform you've identified, research (using public sources, platform transparency reports, and academic literature) what is publicly known about its recommendation algorithm. What engagement signals does it prioritize? What content categories has it identified as generating high engagement? Are there known recommendation pathways that have been documented by researchers?

Step 3: Conduct a recommendation audit. Create a new, clean account on the most relevant platform (do not use your existing account — its history will confound the results). Engage with content typical of your target community's information consumption patterns. Document what the algorithm recommends after each engagement. Are you seeing a drift toward more emotionally intense content? Are specific political framing patterns being consistently surfaced?

Step 4: Trace a specific piece of propaganda content. Choose one piece of propaganda content you identified in your earlier channel audit work. Trace its distribution pathway: who originally posted it, which accounts shared it, how widely it spread, and whether algorithmic recommendation appears to have played a role in its distribution. (Evidence for algorithmic distribution: content that spread far beyond the original poster's follower network without explicit sharing by accounts with large followings is likely algorithmically amplified.)

Step 5: Write up and present. Prepare a 500-word writeup and a 5-minute presentation for next week's class that answers the question: Is algorithmic amplification a significant factor in the distribution of propaganda to your target community? If so, which platforms' algorithms are most relevant, and what design features appear most responsible?


Chapter Summary

This chapter has traced the structural logic by which algorithmic recommendation systems interact with the economics of the attention economy to create an information environment that systematically advantages propaganda over accurate information.

The foundational insight — Herbert Simon's 1971 observation that attention, not information, is the scarce resource in an information-abundant world — established the frame within which all subsequent analysis sits. When platforms are in the attention business rather than the information business, they optimize for content characteristics that generate engagement: emotional intensity, outrage, novelty, identity threat, and community formation. Propaganda, by its nature, is engineered to exploit exactly those characteristics.

The filter bubble concept (Pariser) and the echo chamber concept describe two related but distinct dynamics by which algorithmic and social mechanisms fragment shared information environments, creating conditions in which propaganda can circulate within targeted communities without encountering effective counter-argument. The empirical research has established that both phenomena are real, though their magnitude varies across platforms and over time, and that the most powerful driver of political information cocooning is human choice rather than algorithmic curation alone.

The Bail et al. (2018) finding — that exposing people to cross-cutting political content makes them more extreme, not less — complicates both the filter bubble diagnosis and the available remedies. It suggests that the problem is not simply one of information exposure but of the social and relational context in which information is received. This finding has significant implications for platform design interventions aimed at reducing polarization.

The YouTube radicalization pipeline (Ribeiro et al., 2019; Lewis, 2018) and the Facebook internal research (Haugen, 2021) document the practical consequences of engagement optimization at scale: systematic pathways from mainstream to extreme content on YouTube, and documented awareness at Facebook that its design choices generated significant harms to civic information quality and individual well-being — awareness that was not acted upon.

Microtargeted behavioral advertising, exemplified by but not limited to Cambridge Analytica's activities, enables the fragmentation of political messaging across audiences with no visibility to external observers — a structural propaganda infrastructure built into the normal operations of digital political advertising.

The regulatory debate maps three positions — no regulation, structural regulation, and the DSA systemic risk model — onto the genuine tensions between free speech principles, platform accountability, and technical feasibility. The EU DSA represents the most systematic regulatory response yet enacted.

For students of propaganda, the central takeaway is this: the algorithms that shape contemporary information environments are not neutral distribution mechanisms. They are designed systems with specific optimization objectives, and those objectives systematically advantage certain types of content over others. Propaganda, as content specifically engineered to exploit emotional responses and mobilize audiences, is precisely the type of content that algorithmic engagement optimization rewards. Understanding this structural reality is a prerequisite for understanding how propaganda functions in twenty-first-century information environments — and for developing meaningful forms of resistance.


Key Terms

Attention Economy — The economic framework in which human attention is the scarce resource competed for by information producers, and in which advertising-supported media captures and sells attention to advertisers.

Engagement Metrics — Measurable proxies for attention (clicks, shares, reactions, watch time) that determine content distribution and advertising revenue on digital platforms.

Recommendation Algorithm — A system for predicting and surfacing content predicted to be engaging for a specific user, typically combining collaborative filtering, content-based filtering, and engagement optimization.

Filter Bubble — Algorithmic personalization that creates individual information environments shaped by past behavior, potentially reducing exposure to cross-cutting or counter-attitudinal content.

Echo Chamber — A socially maintained information environment in which a community primarily encounters content confirming its existing beliefs, independent of algorithmic curation.

Feedback Loop — The self-reinforcing dynamic in recommendation systems by which engagement with content generates more content of the same type, potentially escalating toward more extreme or emotionally intense content.

Radicalization Pipeline — The documented pathway on YouTube by which recommendation algorithms lead users from mainstream content to progressively more extreme content through a series of incremental recommendations.

Cambridge Analytica — The British political consulting firm that improperly obtained Facebook data on 87 million users and used behavioral targeting for political clients, whose actual capabilities were less than claimed but whose activities highlighted the opacity of digital political advertising.

Frances Haugen — The Facebook whistleblower who disclosed tens of thousands of internal documents in 2021 showing the company's awareness of harms generated by its platform design.

Digital Services Act (DSA) — The EU regulation enacted in 2022 requiring very large online platforms to conduct systemic risk assessments of their design choices and implement mitigation measures for identified risks to democratic discourse and public health.


Chapter 17 of 40 | Part 3: Channels | Progressive Project: Channel Audit — Week Five