35 min read

Pull up any mainstream news website. Wait for it to load fully. Now, if you have the right browser extension installed, look at the tracker count in the upper right corner of your browser: 47 trackers blocked. 63 trackers. 89 trackers.

Chapter 12: Browser Cookies, Tracking Pixels, and the Third-Party Data Ecosystem

Opening: The Hidden Architecture of Every Webpage

Pull up any mainstream news website. Wait for it to load fully. Now, if you have the right browser extension installed, look at the tracker count in the upper right corner of your browser: 47 trackers blocked. 63 trackers. 89 trackers.

Each of those trackers represents a separate organization — an advertising network, an analytics company, a data broker, a social media platform, a video streaming service — that was notified of your presence on that webpage the moment you loaded it. You didn't click anything. You didn't log in. You didn't even finish reading the headline. But forty-seven (or sixty-three, or eighty-nine) companies now know you were there, what time it was, what device you're using, approximately where you are, and — if they've seen you before — how your visit fits into a months-long behavioral record.

The technical infrastructure that makes this possible evolved over three decades from a modest, practical solution to a website memorization problem into the most comprehensive behavioral surveillance system in human history. Understanding that evolution — and the specific technologies that compose the current system — is essential for understanding both the scope of commercial surveillance and the limits of the countermeasures available against it.


The story of the cookie begins not with advertising but with shopping carts.

In 1994, a Netscape Communications engineer named Lou Montulli was working on the company's nascent web browser when he encountered a fundamental problem with the HTTP protocol, the language of the World Wide Web. HTTP is, by design, stateless: each request a browser makes to a server is treated as independent, with no memory of previous requests. For browsing static documents, this is fine. But for commerce — which requires the server to remember what you put in your cart as you navigate from page to page — statelessness is a fatal limitation.

Montulli's solution was elegant: a small text file, stored on the user's computer by the browser, that the server could write to and subsequently read from on future visits. The server sends an instruction — "set this cookie: ShoppingCart=book123" — and the browser stores it. On the next page load, the browser automatically sends the cookie back to the server, which reads it and knows the cart contains a book. The shopping cart problem is solved.

Montulli named his invention a "cookie" after the "magic cookie" concept in Unix computing — a token passed between programs that carries identifying information. He included a feature in the original specification that would eventually become significant: the domain attribute, which specified which servers could read a particular cookie. In the original design, only the server that set a cookie could read it. Montulli intended this as a privacy protection.

What Montulli did not anticipate — what few could have anticipated in 1994 — was that the web would evolve into an ecosystem in which a single webpage might load resources (images, scripts, advertisements) from dozens of different servers. Each of those servers could set and read cookies on the user's browser. The domain attribute protected cookies from other websites' servers. It did not protect users from the commercial ecosystem that would develop around embedding third-party resources in webpages.

💡 Intuition: Think of the cookie as a sticky note on your forehead as you enter a store. When you return, the store can read the note and remember you. Montulli designed the note to be readable only by the store that wrote it. But the internet evolved into a shopping mall, and the mall architecture means that dozens of vendors operating stalls inside every store can each put their own sticky note on your forehead — and read all the others while they're there.

For readers who will encounter cookie-related technical documentation, the basic attributes of cookies are worth understanding:

  • Name/Value: The data stored (e.g., sessionID=abc123)
  • Domain: Which server(s) can read the cookie
  • Path: Which paths on the server can read the cookie
  • Expires/Max-Age: When the cookie expires (session cookies expire when the browser closes; persistent cookies can survive for years)
  • Secure: Flag indicating the cookie should only be sent over HTTPS
  • HttpOnly: Flag preventing JavaScript from reading the cookie (a security measure)
  • SameSite: Attribute controlling cross-site cookie behavior (increasingly important for privacy)

The Secure and HttpOnly attributes are security features added years after the original specification. The SameSite attribute, which restricts cookies from being sent in cross-site requests, was added much later still and is at the center of current efforts to limit third-party tracking.


12.2 First-Party vs. Third-Party Cookies: The Core Distinction

The most important distinction in cookie privacy is between first-party cookies and third-party cookies.

First-party cookies are set and read by the website you are directly visiting. When you log into your bank's website and the bank sets a cookie that maintains your session, that is a first-party cookie. When a news site saves your reading preferences, that is a first-party cookie. First-party cookies are, in general, technically necessary for websites to function as users expect. They implement the features users want: remembered logins, shopping carts, saved preferences.

Third-party cookies are set by servers other than the one running the webpage you are visiting. When you visit a news site that loads an advertisement from an ad network, the ad network's server can set a cookie on your browser — even though you never visited the ad network's website. The next time you visit a different website that loads ads from the same ad network, the network can read the cookie it set earlier and recognize you as the same person who visited the news site.

This recognition capability is the foundation of cross-site behavioral tracking. A user who visits one hundred different websites that all participate in the same ad network is, from the ad network's perspective, a single identifiable individual with a one-hundred-site browsing history. The sites themselves may know nothing about the user — no account, no login — but the ad network has assembled a detailed behavioral profile from the user's browsing across those sites.

The scale of this is larger than most users appreciate. The major advertising networks — Google's DoubleClick (now Google Ad Manager), the Meta Audience Network, and numerous others — are embedded in tens of millions of websites. A user who simply browses the web without any special privacy protection will encounter the same advertising network trackers repeatedly across thousands of sites, allowing those networks to maintain detailed browsing histories that span years of online activity.

📊 Real-World Application: Research by the Web Privacy Measurement project at Princeton found that DoubleClick (Google's advertising network) appeared on over 80% of the top 10,000 websites by traffic. This means that Google's third-party tracking infrastructure has potential visibility into the browsing behavior of virtually any user who visits mainstream websites — regardless of whether that user has a Google account or has ever knowingly interacted with Google's products.


12.3 Tracking Pixels: Surveillance in a Single Transparent Dot

Alongside cookies, the second major technical mechanism of web tracking is the tracking pixel, also known as a web beacon or pixel tag. Where cookies store data on the user's computer, tracking pixels work by triggering a server request that the tracking company logs.

A tracking pixel is an image — often literally one pixel by one pixel, transparent, and invisible to the naked eye — embedded in a webpage or an email. When your browser loads the page, it automatically requests every image on the page, including the pixel. That request travels to the tracking company's server, which logs: your IP address, your user agent (browser type and version, operating system), the time of the request, the URL of the referring page, and — if a cookie is present — a link to your behavioral profile.

The word "request" understates the disclosure: when your browser "requests" the tracking pixel, it is, in effect, announcing your presence, device, location, and identity to the tracking company. No action on your part is required. No click, no login, no form submission. Simply loading the page triggers the disclosure.

Email tracking pixels are a particularly significant application. Marketing emails routinely embed tracking pixels that report when an email has been opened, from what device, at what time, and from what approximate location. The sender learns not just that you opened the email, but when — which correlates with your schedule — and where — which correlates with your location. Open multiple emails from the same sender over months, and the sender accumulates a behavioral profile that reveals your daily routine, your device ecosystem, and your reading patterns. Email clients like Apple Mail have introduced tracking pixel blocking; many others have not.

⚠️ Common Pitfall: Students sometimes assume that tracking pixels are limited to advertising contexts. In fact, tracking pixels are used across many contexts: market research firms embed them in web surveys to track completion rates and demographic correlates; political campaigns embed them in fundraising emails; even some government agencies have used pixel tracking in public communications. The technology is not inherently commercial, and the surveillance it enables is not limited to advertising.

Tracking Pixels in Practice: The Email Client Problem

Consider the following scenario. A company sends a promotional email to its subscriber list. The email contains a tracking pixel hosted by a marketing analytics service called, for example, Mailchimp, HubSpot, or Klaviyo. When a subscriber opens the email:

  1. The subscriber's email client (Gmail, Outlook, etc.) loads the email's images, including the tracking pixel
  2. The pixel request is sent to the analytics service's server
  3. The server logs the subscriber's IP address, device information, timestamp, and email identifier
  4. This data is added to the subscriber's profile in the analytics platform
  5. The company can now see exactly who opened the email, when, and from what device

If the email is opened from a workplace IP address, the company may be able to infer the subscriber's employer. If opened from a home IP address at 11:30 PM, the company learns something about the subscriber's schedule. If opened multiple times, the company learns the subscriber re-read the email — a signal of interest.

None of this requires the subscriber to click any link, respond to the email, or take any action whatsoever. Simply opening the email discloses this information. Most email users have no awareness that this is occurring.


12.4 The Third-Party Data Ecosystem: An Architecture of Players

The tracking pixel and third-party cookie technologies described above do not operate in isolation. They are components of a commercial ecosystem with multiple distinct layers, each playing a specialized role in the collection, processing, and monetization of behavioral data. Understanding this ecosystem requires a brief introduction to its major players.

Demand-Side Platforms (DSPs)

A Demand-Side Platform is a technology system that allows advertisers to purchase digital advertising inventory programmatically — through automated auctions rather than direct negotiation with specific publishers. Advertisers (or their agencies) use DSPs to specify who they want to reach (audience targeting), how much they will pay (bid price), and where they want their ads to appear (inventory preferences). The DSP then participates in real-time auctions on the advertiser's behalf. Major DSPs include Google Display & Video 360, The Trade Desk, Amazon DSP, and Verizon Media (now Yahoo DSP).

Supply-Side Platforms (SSPs)

A Supply-Side Platform is the publisher-facing complement to the DSP. Publishers (websites and apps that want to sell advertising space) use SSPs to make their inventory available for programmatic purchase, set floor prices below which they will not sell, and manage their relationships with multiple ad exchanges simultaneously. Major SSPs include Google Ad Manager (formerly DoubleClick for Publishers), Magnite, and PubMatic.

Data Management Platforms (DMPs)

A Data Management Platform is a data aggregation and segmentation system that collects behavioral data from multiple sources, organizes it into audience segments, and makes it available for targeting. DMPs are where the data pipeline's aggregation stage happens: they ingest cookie data, behavioral logs, CRM data, and third-party purchased data, and output audience segments that DSPs can use to target specific audience types. Major DMPs include Oracle BlueKai, Salesforce DMP (formerly Krux), and Nielsen Marketing Cloud.

Ad Exchanges

An Ad Exchange is the marketplace where advertising inventory is bought and sold through real-time auctions. Publishers make their inventory available through the exchange; advertisers bid on specific impressions through DSPs. The ad exchange is the mechanism of real-time bidding (RTB), which Chapter 14 examines in depth. Major ad exchanges include Google Ad Exchange, OpenX, and Xandr (formerly AppNexus).

Identity Resolution Services

Newer entrants to the ecosystem, identity resolution services (also called "identity graphs") specialize in linking behavioral data across devices, contexts, and time to maintain a persistent profile of individual users. When a third-party cookie expires or is blocked, identity resolution services use probabilistic matching (behavioral similarity) and deterministic matching (shared identifiers like email addresses) to maintain continuity of profile. Companies like LiveRamp, Neustar, and Tapad specialize in this function.

🎓 Advanced: The distinction between deterministic and probabilistic identity matching is significant for both privacy law and surveillance effectiveness. Deterministic matching uses a shared identifier — an email address, phone number, or device ID — that definitively links two records to the same person. Probabilistic matching uses statistical inference — you connected from the same IP at the same time of day on a device with the same browser fingerprint — to estimate that two records belong to the same person. Probabilistic matching carries uncertainty; deterministic matching does not. Privacy regulations typically treat these methods differently: deterministic matching often involves "personal data" under GDPR's definition; probabilistic matching may or may not, depending on the precision and the jurisdiction.

🔗 Connection: The DMP-DSP-SSP-ad exchange ecosystem is the commercial infrastructure within which behavioral targeting (Chapter 14) operates. Understanding the technical architecture of the ecosystem is a prerequisite for understanding how real-time bidding works, what data flows through it, and what privacy implications each transaction carries.


Third-party cookies are, in principle, deletable. Users can clear their cookies, use private browsing mode, or install cookie-blocking extensions. This was a known limitation of cookie-based tracking, and the advertising industry developed a powerful alternative: browser fingerprinting.

Browser fingerprinting works not by storing information on the user's computer but by reading the distinctive combination of technical characteristics that every browser reveals when it makes a request. These characteristics — individually unremarkable, collectively distinctive — include:

  • User agent string: Browser type, version, and operating system
  • Screen resolution and color depth
  • Installed fonts: Which fonts are present on the system
  • Plugin list: Which browser plugins are installed
  • Time zone and language settings
  • Canvas fingerprint: How the browser renders a specific test drawing (varies subtly by graphics hardware, driver, and OS)
  • WebGL fingerprint: How the browser handles 3D graphics (highly distinctive)
  • Audio context fingerprint: How the browser processes a specific audio signal
  • Battery status (in some implementations): Current battery level and charging status
  • CPU and memory specifications
  • Network characteristics: Connection type, DNS response timing

The Panopticlick project (now "Cover Your Tracks") at the Electronic Frontier Foundation demonstrated in 2010 that 84% of browsers had a combination of characteristics unique in the dataset of millions of tested browsers. More recent research has found even higher uniqueness rates, with some studies reporting that 90–95% of browsers are globally unique based on fingerprint alone.

This means that browser fingerprinting can identify and track individual users without storing any data on their computers, without cookies, and — critically — across private browsing sessions, which clear cookies but do not change the browser's hardware and software characteristics.

📊 Real-World Application: The research firm Inria conducted a large-scale study of browser fingerprinting across 98,000 users, finding that 89.4% of users with JavaScript enabled could be uniquely identified from browser fingerprints alone. Critically, they found that fingerprints remained stable enough to re-identify 65.4% of returning visitors even when their browser had been partially updated. The practical implication: a user who believes they have evaded tracking by clearing cookies may still be tracked through fingerprinting, without any stored data and without any cookie consent banner triggering.

Canvas Fingerprinting: Technical Deep Dive

Among fingerprinting techniques, canvas fingerprinting deserves particular attention because of its precision and its difficulty to spoof. Canvas fingerprinting works as follows:

  1. A tracking script (invisible to the user) uses JavaScript to instruct the browser to draw a specific image (typically text with specific fonts, colors, and effects) on an invisible HTML5 canvas element
  2. The browser renders the image using its graphics hardware, drivers, and OS-level text rendering
  3. The script reads the pixel values of the resulting image
  4. These pixel values are hashed (mathematically condensed) into a fingerprint

Because different graphics hardware, driver versions, operating systems, and browser versions render the same canvas instruction slightly differently, the resulting fingerprint is highly distinctive. A user who has cleared cookies, changed their IP address through a VPN, and switched to private browsing mode may still be identifiable through canvas fingerprinting, because none of those actions change their graphics hardware or driver.

A 2014 study by researchers at Princeton found canvas fingerprinting used on 5% of the top 100,000 websites — a number that has grown substantially since. The tracking tool AddThis was among the most prolific deployers in the early study, having embedded canvas fingerprinting in websites of Fortune 500 companies, government websites, and news organizations without their full understanding of the technique.


12.6 Cross-Device Tracking: Following You from Phone to Laptop to TV

A single user in 2026 typically operates across multiple devices: a smartphone, a laptop, possibly a tablet, a smart TV, a work computer. Each device has its own IP address, its own cookies, and potentially its own browser fingerprint. From a raw data perspective, they look like different people.

Cross-device tracking refers to the techniques used to link behavioral profiles across these different devices, establishing that the person who searched for "refinancing a mortgage" on their phone this morning is the same person who later browsed real estate listings on their laptop and watched financial advice videos on their smart TV.

The techniques break into two categories:

Deterministic cross-device tracking uses shared identifiers that definitively link devices to a single person. If a user logs into Facebook on both their phone and their laptop, Facebook links the behavioral data from both sessions through the common account. Email addresses are particularly powerful deterministic identifiers: if the same email address appears in behavioral records from three different devices (registered to an app on the phone, used to log into a service on the laptop, linked to a smart TV account), they can be definitively linked. Major platforms with large logged-in user bases — Google, Facebook, Amazon — have extensive deterministic cross-device graphs based on account logins across devices.

Probabilistic cross-device tracking uses statistical inference to link devices without a shared identifier. If two devices repeatedly appear on the same Wi-Fi network at the same hours, have overlapping behavioral patterns (both show interest in the same product categories), and have similar demographic signals, a probabilistic model estimates that they belong to the same person. The precision varies — estimates of probabilistic cross-device matching accuracy range from 50% to 85% in the literature — but even imprecise matching dramatically improves the completeness of behavioral profiles.

Ultrasonic audio beacons represent a more exotic form of cross-device tracking. Ultrasonic beacons are high-frequency sound signals (above the threshold of human hearing) that can be embedded in television advertisements, in-store audio, or website audio content. A mobile app with permission to access the microphone can detect these beacons and report them to the tracking company, linking the phone's identity to the TV screen's content at a specific moment. The FTC received complaints about this technique from a privacy research group in 2015, and while several companies responded to regulatory pressure by disabling the feature in their apps, the technical capability remains available.


Anyone who has used the internet in Europe since 2018, or increasingly in any jurisdiction with digital privacy regulations, has encountered the cookie consent banner: a dialog that appears when you first visit a website, presenting options about your tracking preferences and asking for consent.

Cookie consent banners exist because the GDPR (discussed below) and similar regulations require "informed consent" for non-essential cookies. In principle, they are a user empowerment mechanism — they give users meaningful choice about whether to be tracked.

In practice, they have become what researchers call dark patterns — interface designs that exploit cognitive biases and choice architecture to nudge users toward choices they might not make if the options were presented neutrally.

Pre-checked boxes: Options to accept tracking cookies are pre-checked, while options to reject are unchecked by default. Users must actively uncheck boxes to opt out, exploiting the status quo bias — the human tendency to go with whatever is already selected.

Asymmetric button design: "Accept all" is displayed as a prominent, brightly colored button; "Manage preferences" or "Reject all" is displayed in smaller, grayed-out text, often requiring navigation to a settings page with dozens of individual toggles.

Option burial: The "reject all" option, if present, requires navigating through multiple layers of menus. The "accept all" option is one click. Research by the Norwegian Consumer Council (Forbrukerrådet) found that on some major platforms, accepting all cookies required one click while rejecting all cookies required thirty-two clicks across multiple nested menus.

Forced interaction: Some banners prevent users from accessing the website until they have engaged with the consent dialog — effectively making the website inaccessible unless the user clicks "accept all," even though the GDPR requires that consent be freely given.

Deceptive framing: Consent dialogs sometimes describe accepting cookies as "helping us improve your experience" or "supporting journalism" while describing rejection as "limiting" or "reducing" the user's experience — framing consent in value-laden terms that make rejection seem negative.

Redundant consent: Some sites require users to set their preferences on every visit, making the process so tedious that users eventually click "accept all" to make it stop.

A 2019 study by academics at MIT, Carnegie Mellon, and the University of Michigan analyzed consent banners on the top 10,000 websites and found that websites using dark patterns achieved 23 percentage points higher consent rates than sites that used neutral banner designs. The design of consent banners is not incidental to their privacy implications; it is central to them.

⚠️ Common Pitfall: A pervasive misconception is that clicking "accept all" on a cookie consent banner is equivalent to accepting tracking only from the website you are visiting. In fact, most consent banners describe consent to multiple layers of tracking: the first-party website, the advertising networks the website participates in, the DMPs that aggregate data, and dozens of other third parties. The "purposes" described in the banner (analytics, advertising personalization, measurement) may correspond to hundreds of distinct companies. Reading the full partner list on a major news site's consent banner — a list that can include 500+ companies — provides a visceral sense of the ecosystem's scale.


The General Data Protection Regulation, which took effect in the European Union in May 2018, represented the most significant legal intervention in the third-party tracking ecosystem to that point. Among its provisions with the most immediate practical consequences were requirements for:

  • Lawful basis for processing: Data processors must identify a lawful basis — consent, legitimate interest, legal obligation, vital interest, public task, or contractual necessity — for each processing activity involving personal data
  • Consent standards: Consent must be freely given, specific, informed, and unambiguous; pre-ticked boxes do not constitute consent; consent must be as easy to withdraw as to give
  • Data subject rights: Rights to access, rectification, erasure, portability, restriction of processing, and objection to processing
  • Data minimization: Only data necessary for the stated purpose should be collected
  • Purpose limitation: Data collected for one purpose cannot be repurposed for another incompatible purpose without fresh consent

In practice, GDPR did not eliminate the third-party tracking ecosystem in Europe. It created a complex, contested, and partially compliant consent management infrastructure — the cookie consent banner system — that the Norwegian Consumer Council and others have argued fails to provide genuine consent in the majority of implementations.

Enforcement has been uneven. Major actions include: - A €150 million fine against Twitter for making it unnecessarily difficult for users to opt out of targeted advertising - A €60 million fine against Google for failing to obtain valid consent in its Google.fr cookies implementation - An €390 million fine against Meta for relying on contractual necessity rather than consent as the basis for behavioral advertising

The cumulative fine total from GDPR enforcement, while significant in absolute terms, represents a small fraction of the revenues the practices generate. Critics have argued that GDPR fines, as currently calibrated, function as a cost of doing business rather than a deterrent.

🌍 Global Perspective: The GDPR's influence has extended well beyond Europe. Brazil's Lei Geral de Proteção de Dados (LGPD), California's CCPA and CPRA, Virginia's CDPA, and proposed federal privacy legislation in the United States have all drawn on GDPR concepts and structures. This "Brussels Effect" — the tendency of the EU's large market to make its standards globally influential — has made GDPR the de facto baseline for privacy policy globally, even in jurisdictions where it does not technically apply. Companies often apply GDPR-like consent mechanisms globally because the implementation cost of different systems for different jurisdictions exceeds the benefit of looser practices in non-EU markets.


12.9 Google's Privacy Sandbox: What It Does and Doesn't Do

In 2020, Google announced that it would deprecate third-party cookies in its Chrome browser — at the time, the dominant global web browser with roughly 65% market share. The announcement was significant because Chrome's support for third-party cookies had been a major reason the tracking ecosystem had continued to operate even as Firefox and Safari had already blocked them by default.

Google's replacement proposal, eventually called the Privacy Sandbox, generated immediate and sustained controversy. The core technology proposal was called Federated Learning of Cohorts (FLoC) — later replaced by the Topics API after FLoC was withdrawn following significant criticism.

How the Topics API Works

Under the Topics API, the browser (Chrome) analyzes the user's browsing history and assigns them to topic categories from a predefined taxonomy (e.g., "Sports/Baseball," "Finance/Investing," "Health/Fitness"). When a website requests a "topic" for targeting purposes, the browser shares one of the user's recent top topics — but the full browsing history never leaves the device.

The claimed privacy benefit is that targeting is based on topics (determined on-device) rather than individual cross-site behavioral profiles (assembled by third parties). The advertiser learns that a user is interested in "Finance/Investing" without learning which specific finance sites they visited, when, or how often.

Criticisms and Limitations

The Privacy Sandbox proposals received significant criticism from multiple directions:

Privacy researchers argued that even topic-level data can enable re-identification when combined with other signals — device fingerprint, IP address, on-site behavior. The privacy improvement over cookies is real but modest.

Advertisers and publishers argued that Topics-based targeting would be substantially less effective than cookie-based behavioral targeting, reducing ad revenue for publishers and ROI for advertisers.

Regulators raised concerns about competitive implications: by eliminating third-party cookies while maintaining first-party data capabilities through its own advertising products (Google Ads, Google Ad Manager), Google could disadvantage competitors who relied on third-party cookies. The UK Competition and Markets Authority (CMA) opened an investigation into the Privacy Sandbox in 2021 and reached a negotiated agreement with Google that included ongoing monitoring and commitments not to treat Google's own advertising operations preferentially.

Civil liberties organizations pointed out that replacing one form of tracking with a browser-level tracking mechanism maintained by the world's largest advertising company is not obviously a privacy improvement — it potentially concentrates more user data in Google rather than distributing it among thousands of third parties.

By 2024, Google had repeatedly postponed the deprecation of third-party cookies in Chrome, citing ecosystem readiness concerns and regulatory uncertainty. The transition is ongoing and unresolved as of this writing — a testament to how deeply embedded the third-party cookie ecosystem is in the commercial web.

📝 Note: The Privacy Sandbox debate illustrates a recurring tension in surveillance studies between competing privacy framings. One framing prioritizes the distribution of surveillance power — many companies tracking users is worse than one company tracking users. Another framing prioritizes the reduction of surveillance power — any reduction in behavioral data collection is an improvement regardless of who controls what remains. These framings can lead to opposite conclusions about the same proposal, and both are arguably valid responses to different aspects of the problem.


12.10 Jordan's Scenario: Seeing the Trackers

Jordan had heard the phrase "you're being tracked" so many times that it had lost most of its meaning. It was one of those things everyone said but that felt too abstract to really grapple with. Until the evening Jordan installed Ghostery.

Ghostery is a browser extension that blocks and identifies tracking scripts. The installation took thirty seconds. The effect was immediate.

The first website Jordan loaded — a local newspaper's homepage — showed a tracker count of 71. Seventy-one separate tracking organizations had been notified of Jordan's presence on a single page. Jordan clicked the extension icon and scrolled through the list: Google, Facebook, Criteo, DoubleClick, Quantcast, comScore, Amazon, LinkedIn, Twitter, TikTok, and fifty more companies Jordan had never heard of. The newspaper itself was one actor among seventy-two.

Jordan navigated to a health information site to look up information about managing anxiety — something Jordan had been dealing with more since the warehouse hours increased. The tracker count: 43. The list included insurance industry trackers Jordan had never heard of.

"Does that mean someone at those insurance companies knows I looked at anxiety information?" Jordan texted Yara.

"Not exactly," Yara replied. "They know something looked at that page from your device at that time. Whether it's linked to you specifically depends on whether they have your identity from somewhere else."

"That's not as reassuring as you probably intended," Jordan replied.

In class the next week, Jordan raised the experience with Dr. Osei. "I knew I was being tracked in some general sense. But seeing the actual count on a page I thought I was just reading — it changed how it felt."

Dr. Osei nodded. "That's the value of making the invisible visible, even briefly. The trackers are always there. The experience of actually seeing them is rare. Most people never have that experience, so most people can't process what it means in any concrete way. The architecture of the system depends on invisibility. Making it visible, even temporarily, is a form of counter-surveillance."

✅ Best Practice: Browser extensions that reveal tracking infrastructure — Ghostery, uBlock Origin, Privacy Badger — are useful not primarily as blocking tools (though they do block trackers) but as educational tools that make invisible surveillance infrastructure visible. Using one for even a brief period can change how you experience the web in a way that no description of trackers can fully replicate. However, be aware that tracker-blockers have their own limitations: they rely on known tracker lists, can be evaded by sophisticated tracking techniques, and may introduce new privacy concerns depending on the extension's own data practices.


12.11 The Limits of Technical Counter-Measures

The tracking ecosystem described in this chapter has prompted a corresponding ecosystem of counter-measures: ad blockers, cookie managers, VPNs, privacy-focused browsers, and tracking protection features in mainstream browsers. Understanding what these tools do — and do not — accomplish is essential for evaluating their role as privacy protections.

Cookie blocking and clearing addresses one layer of tracking — the persistent identifier stored in the browser — but does not address fingerprinting, IP-based tracking, or server-side behavioral logging.

VPNs (Virtual Private Networks) conceal the user's IP address from the websites they visit, but do not block cookies or fingerprinting. They also transfer trust from the website operator to the VPN provider — the VPN sees the same traffic the website would have.

Private browsing mode prevents cookies and browsing history from being stored on the device but does not prevent tracking by the websites visited — which log server-side. It also does not prevent fingerprinting.

Ad blockers like uBlock Origin block the loading of resources from known advertising and tracking domains, preventing those domains from setting cookies and receiving tracking requests. They are among the most effective single tools for reducing third-party tracking. But ad blockers rely on maintained blocklists and can be evaded by first-party data collection (which becomes more important as third-party cookies decline).

Privacy-focused browsers (Firefox with enhanced tracking protection, Brave, or the Tor Browser) offer varying levels of protection. Firefox's Enhanced Tracking Protection blocks third-party cookies from known trackers. Brave goes further, blocking ads and fingerprinting by default. The Tor Browser routes traffic through the Tor network and normalizes fingerprinting parameters to make all users look the same — the gold standard for browser-level anonymity, with significant performance costs.

The broader point is that these technical tools exist within the same structural dynamic as all individual-level responses: they can reduce surveillance exposure at the margins but cannot address the underlying economic incentives and legal frameworks that produce mass surveillance infrastructure. A user who installs every available privacy tool is still operating in a web environment designed for comprehensive behavioral tracking; they are swimming against a very strong current.

🎓 Advanced: The concept of differential privacy offers a technical approach to the privacy problem that differs from blocking. Rather than preventing data collection, differential privacy adds mathematical noise to datasets such that individual records cannot be identified even when the aggregate statistical properties are preserved. This allows companies to learn from behavioral data without learning about specific individuals. Apple has adopted differential privacy for some of its data collection, and it appears in the Privacy Sandbox proposals. Critics note that differential privacy provides formal mathematical privacy guarantees but depends on implementation correctness and parameter choices that are not always publicly audited.


12.12 Historical Continuity: Tracking Before the Web

The web tracking ecosystem described in this chapter is genuinely new in its technical mechanisms and scale. But the impulse it implements — to know who is visiting your commercial space, how they behave, and what they can be sold — is not new at all.

Before cookies, direct mail marketers tracked customer behavior through the catalog orders they received, noting which products were ordered, which were returned, which promotions generated response, and which customers were most valuable. The Retail Analysis department of Sears, Roebuck in 1930 was doing, with paper and pencil, a version of what data management platforms do automatically today.

Before tracking pixels, publishers knew which advertisements generated reader response through coded coupons and order forms — a version of conversion tracking. Before browser fingerprinting, market researchers conducted in-store observation studies to track which products shoppers touched, examined, and ultimately purchased.

The technical innovations of the web tracking ecosystem are real and consequential. They operate at a scale and with a precision that has no historical parallel. But they are not a rupture with prior commercial surveillance practice; they are an acceleration of it. The economic incentives that drove Sears' catalog analytics in 1930 are the same economic incentives that drive Google's advertising infrastructure in 2026. The tools are unrecognizably different; the motivation is the same.

🔗 Connection: This historical continuity connects to one of the textbook's central themes: that surveillance technologies change, often dramatically, while underlying social and economic motivations show remarkable stability. The behavioral tracking ecosystem is the current implementation of a commercial interest in understanding consumers that is as old as commerce itself. Understanding this continuity helps resist the temptation to treat the current situation as uniquely unprecedented — while also recognizing that unprecedented scale produces genuinely new consequences, even from familiar motivations.


Summary: The Collection Layer

The technical infrastructure of web tracking is vast, layered, and largely invisible to the users whose behavior it monitors. Cookies — invented to solve a shopping cart problem — became the foundation of a global behavioral tracking ecosystem. Tracking pixels extended that ecosystem into email and across contexts where cookies cannot operate. Browser fingerprinting provides tracking that persists even when cookies are blocked or cleared. Cross-device tracking knits together behavioral data from phones, laptops, and televisions. The third-party data ecosystem — DSPs, SSPs, DMPs, and ad exchanges — provides the commercial infrastructure within which tracked data is processed and monetized.

Cookie consent banners, the primary legal response to this infrastructure in much of the world, have been captured by dark pattern design and provide, in the majority of implementations, the form rather than the substance of consent. GDPR enforcement has produced real changes — and real compliance theater. The Privacy Sandbox represents the current frontier of contested transition, with the commercial web working to preserve behavioral targeting capability while satisfying at least some regulatory and public relations demands for privacy improvement.

Jordan's experience of installing Ghostery and seeing seventy-one trackers on a local newspaper's homepage — including insurance industry trackers on a health information site — provides a concrete anchor for the otherwise abstract claim that "the web tracks you." It is not abstract. It is seventy-one organizations, notified simultaneously, every time you load a page.

Chapter 14 will examine what the tracking ecosystem produces: the behavioral targeting and real-time bidding systems that are its commercial output. Chapter 32 will return to counter-surveillance tools with a more systematic treatment of their effectiveness and limitations.


Key Terms

Browser fingerprinting — Tracking technique that identifies users based on the distinctive combination of technical characteristics their browser reveals, without storing data on the device.

Canvas fingerprinting — A fingerprinting technique that uses subtle variations in how different systems render a canvas drawing to identify individual browsers.

Cookie — A small text file stored on a user's computer by a website, used to maintain state across multiple requests.

Cross-device tracking — Techniques for linking behavioral data across multiple devices used by a single individual.

Dark patterns — Interface design techniques that exploit cognitive biases to nudge users toward choices they might not make if options were presented neutrally.

Data Management Platform (DMP) — A system that aggregates behavioral data from multiple sources and organizes it into audience segments for targeting.

Demand-Side Platform (DSP) — A technology system allowing advertisers to purchase digital advertising inventory through automated auctions.

First-party cookie — A cookie set and readable only by the website the user is directly visiting.

Privacy Sandbox — Google's set of proposed replacements for third-party cookies in the Chrome browser.

Supply-Side Platform (SSP) — A technology system allowing publishers to make advertising inventory available for programmatic purchase.

Third-party cookie — A cookie set by a server other than the website the user is visiting, enabling cross-site behavioral tracking.

Tracking pixel — A tiny, invisible image embedded in webpages or emails that triggers a server request logging user presence and behavior.

Web beacon — Another term for a tracking pixel.


Discussion Questions

  1. Lou Montulli invented the cookie to solve a shopping cart problem, not to enable commercial surveillance. How should we think about the moral responsibility of technology inventors for the downstream uses of their inventions? Does the tracking ecosystem's origin in a practical, neutral solution complicate the ethical analysis of its current consequences?

  2. The dark patterns literature shows that consent banner design has a large and predictable effect on consent rates — neutral designs produce much lower consent rates than designs that nudge toward acceptance. What does this tell us about the quality of consent obtained through typical cookie banners? Does it matter that users technically had a choice?

  3. Google's Privacy Sandbox would replace third-party cookies — controlled by many third parties — with a browser-level system controlled by Google. Privacy researchers are divided about whether this is a net privacy improvement. What values are in tension in this debate? What would you need to know to adjudicate it?

  4. Browser fingerprinting is technically superior to cookies in several ways: it cannot be deleted, does not require storage on the user's device, and persists across private browsing sessions. But it also does not require the user's browser to make any request to a tracking company — the fingerprint is inferred from normal browser operation. Does this difference in mechanism make fingerprinting more or less ethically problematic than cookie tracking?

  5. The chapter notes that tracking has a long pre-digital history. Does the scale of contemporary web tracking — billions of users, real-time, across all digital contexts — change the ethical analysis, or are scale differences merely differences in degree rather than kind?


Chapter 12 of 40 | Part 3: Commercial Surveillance Backward reference: Chapter 11 (The Data Economy) Forward references: Chapter 14 (Behavioral Targeting), Chapter 32 (Counter-Surveillance Tools)