> "The right to be left alone — the most comprehensive of rights, and the right most valued by a free people."
Learning Objectives
- Explain the HIPAA framework for health data governance, including its key provisions, covered entities, and major limitations
- Analyze the privacy implications of direct-to-consumer genomics and evaluate the adequacy of GINA protections
- Distinguish between facial recognition, fingerprint scanning, gait analysis, and voice recognition as biometric identifiers and assess the unique privacy risks of each
- Evaluate Illinois BIPA as a model biometric privacy law and explain why biometric data requires special legal protection
- Apply care ethics (Chapter 6) to evaluate data practices involving health, genetic, and biometric information
- Articulate why health, genetic, and biometric data are categorically more sensitive than other personal data
In This Chapter
- Chapter Overview
- 12.1 Health Data: The HIPAA Framework
- 12.2 Genetic Data: The New Frontier
- 12.3 Biometric Data: Your Body as Your Password
- 12.4 Illinois BIPA: A Model Biometric Privacy Law
- 12.5 The Special Sensitivity Argument
- 12.6 VitraMed Under HIPAA Scrutiny
- 12.7 Eli's Detroit: Biometric Surveillance and Community Impact
- 12.8 Case Study References
- 12.9 Chapter Summary
- What's Next
- Chapter 12 Exercises -> exercises.md
- Chapter 12 Quiz -> quiz.md
- Case Study: 23andMe and the Golden State Killer -> case-study-01.md
- Case Study: Robert Williams — Wrongful Arrest by Facial Recognition -> case-study-02.md
Chapter 12: Health Data, Genetic Data, and Biometric Privacy
"The right to be left alone — the most comprehensive of rights, and the right most valued by a free people." — Justice Louis Brandeis, Olmstead v. United States (1928), dissenting opinion
Chapter Overview
In Chapters 7 through 11, we examined privacy as a concept, as a right under threat from surveillance, as a design challenge, and as an economic problem. Those chapters treated personal data as a broad category. This chapter narrows the focus to three data types that occupy a special status in privacy law and ethics: health data, genetic data, and biometric data.
Why single these out? Because data about your body is different from data about your purchases, your browsing history, or your social connections. If your shopping preferences are exposed, you might receive targeted advertisements. If your health records are exposed, you might lose your insurance, your job, or your social standing. If your genetic data is exposed, the consequences extend not just to you but to every blood relative you have -- people who never consented to anything. If your biometric data is compromised, you cannot change it the way you change a password. Your fingerprints, your face, and your gait are permanent.
These data types share three characteristics that justify heightened protection: 1. Immutability. You can change your address, your name, even your Social Security number. You cannot change your genome or your fingerprints. 2. Intimacy. Health and genetic data reveal the most private aspects of human existence -- disease, vulnerability, mortality, heredity. 3. Externality. Your genetic data is not only yours. It belongs, in a biological sense, to your parents, your children, and your siblings. Your biometric data, once in a database, can be used to identify you without your knowledge or consent in any context where the biometric is captured.
This chapter examines the legal frameworks that attempt to govern these sensitive data types, the emerging technologies and business models that threaten them, and the ethical principles that should guide our response.
In this chapter, you will learn to: - Evaluate whether existing health data governance (HIPAA) is adequate for the digital age - Analyze the privacy implications of genetic data through both legal and ethical lenses - Assess the risks of biometric surveillance and the case for biometric-specific privacy law - Apply the care ethics framework from Chapter 6 to health and biometric data contexts - Connect VitraMed's health data practices to the HIPAA framework and identify emerging risks
12.1 Health Data: The HIPAA Framework
12.1.1 Origins and Structure
The Health Insurance Portability and Accountability Act (HIPAA), signed into law in 1996, is the primary federal framework governing health data privacy in the United States. Its Privacy Rule (effective 2003) and Security Rule (effective 2005) establish national standards for the protection of individually identifiable health information -- what HIPAA calls protected health information (PHI).
PHI includes any information that: - Relates to the past, present, or future physical or mental health of an individual - Relates to the provision of health care to an individual - Relates to payment for health care - Identifies the individual, or could reasonably be used to identify the individual
This is deliberately broad. It covers not just diagnoses and treatment records but also billing information, appointment schedules, lab results, and even the fact that someone is a patient at a particular facility.
12.1.2 Covered Entities and Business Associates
HIPAA applies to covered entities -- health plans, health care clearinghouses, and health care providers who transmit health information electronically -- and their business associates (organizations that handle PHI on behalf of covered entities, such as billing companies, cloud storage providers, and data analytics firms).
The covered entity/business associate framework creates a chain of responsibility:
| Entity | HIPAA Obligation | Example |
|---|---|---|
| Covered entity | Full compliance with Privacy and Security Rules | Hospital, insurance company, clinic |
| Business associate | Compliance per Business Associate Agreement (BAA) | Cloud storage vendor, billing service, EHR software provider |
| Subcontractor of BA | Compliance per subcontractor BAA | Server hosting company used by the EHR vendor |
VitraMed, as an EHR software provider to small clinics, is a business associate under HIPAA. This means VitraMed must comply with HIPAA's security requirements, sign BAAs with each clinic client, and report breaches to the covered entities it serves. Mira was already familiar with the mechanics of HIPAA compliance from her work at the university's Office of Institutional Research, but VitraMed's obligations went further than what she had encountered in the educational context.
12.1.3 Key HIPAA Provisions
The Minimum Necessary Standard: Covered entities and business associates must limit PHI access, use, and disclosure to the minimum necessary to accomplish the intended purpose. This is HIPAA's version of data minimization (Chapter 10), applied specifically to health data.
Individual Rights: Patients have the right to: - Access their own health records - Request corrections to inaccurate information - Receive an accounting of disclosures (who has received their PHI) - Request restrictions on certain uses and disclosures - File complaints with the Department of Health and Human Services (HHS)
The De-identification Standard: HIPAA provides two methods for de-identifying PHI: 1. Expert determination: A qualified statistical expert certifies that the risk of re-identification is "very small" 2. Safe harbor: Removal of 18 specific identifiers (name, address, dates more specific than year, phone numbers, email addresses, SSN, medical record numbers, etc.)
Once data is de-identified per HIPAA standards, it is no longer considered PHI and can be used or disclosed without restriction.
Connection to Chapter 10: HIPAA's de-identification safe harbor relies on removing direct identifiers, but as Sweeney's research demonstrated, quasi-identifiers can re-identify individuals even after direct identifiers are removed. The HIPAA safe harbor does not require removal of all quasi-identifiers -- only the 18 specified types. This means data that technically qualifies as "de-identified" under HIPAA may still be re-identifiable using the linkage attacks we examined in Chapter 10.
12.1.4 HIPAA's Limitations
HIPAA was enacted in 1996 -- an era when most health records were on paper. Its limitations in the digital age are significant:
Scope gaps. HIPAA applies only to covered entities and their business associates. It does not apply to: - Health and fitness apps (Fitbit, Apple Health, period-tracking apps) unless they share data with a covered entity - Direct-to-consumer genetic testing companies (23andMe, Ancestry) unless they share data with a covered entity - Social media posts about health - Employer wellness programs that collect health data outside the covered entity relationship - Data brokers who purchase and resell health-related data
This means a vast and growing volume of health-related data exists entirely outside HIPAA's protection.
"This is the gap that terrifies me," Mira told Dr. Adeyemi. "VitraMed is regulated under HIPAA because we're a business associate. But the period-tracking app on my phone -- which knows things about my health that my doctor doesn't -- has zero HIPAA obligation. It could sell that data to anyone."
Enforcement weakness. The HHS Office for Civil Rights (OCR), which enforces HIPAA, is chronically underfunded relative to its mandate. Enforcement has been inconsistent, with large breaches sometimes resulting in significant fines and smaller violations going unaddressed.
Consent architecture. HIPAA's notice-of-privacy-practices model suffers from the same problems we documented in Chapter 9. Patients receive a multi-page privacy notice at their first visit, sign it without reading, and never see it again. Meaningful informed consent about health data practices is rare.
Research exception. HIPAA permits use of PHI for research purposes under certain conditions (Institutional Review Board approval or waiver of authorization). While this enables valuable health research, it also creates a pathway for health data to move beyond the clinical context in which it was collected -- a contextual integrity concern (Chapter 7).
12.1.5 The Electronic Health Records Revolution
The Health Information Technology for Economic and Clinical Health (HITECH) Act of 2009 accelerated the adoption of electronic health records (EHRs) through financial incentives and penalties. By 2023, over 96% of hospitals and 78% of physician offices in the United States used certified EHR systems.
The shift from paper to electronic records brought enormous clinical benefits: improved care coordination, reduced medical errors from illegible handwriting, easier access to patient history across providers, and the ability to conduct population health research at scale.
But the digitization of health records also created new privacy risks:
Scale of exposure. A breach of paper records at a single clinic might compromise hundreds of files. A breach of an EHR system can expose millions of records simultaneously. The 2015 Anthem breach exposed 78.8 million records. The 2023 Change Healthcare breach affected approximately 100 million individuals.
Interoperability vs. privacy. The push for EHR interoperability -- making health records accessible across different providers and systems -- creates tension with privacy. More connectivity means better care coordination but also more potential access points for unauthorized viewing.
Secondary use. EHR data has become a valuable resource for pharmaceutical companies, insurance actuaries, and health AI developers. De-identified EHR data is bought and sold in a growing market. While de-identification removes direct identifiers, the richness of clinical data (rare diagnoses, unusual treatment combinations, geographic specificity) makes re-identification a persistent concern, as Chapter 10 documented.
The vendor ecosystem. The EHR market is dominated by a handful of companies -- Epic Systems, Cerner (now Oracle Health), and a few others -- that hold patient data for hundreds of millions of individuals. These vendors operate as HIPAA business associates, but their market dominance and the difficulty of switching EHR systems gives them enormous leverage over the clinics and hospitals they serve.
Mira had seen this dynamic firsthand. "Epic's contract with our university hospital runs to 2032. The hospital can't leave even if they wanted to -- the switching costs are in the hundreds of millions. That gives Epic enormous power over how the data is managed, and the hospital has limited ability to demand changes to data practices."
Connection to Chapter 3: The power dynamics of the EHR vendor market connect directly to Chapter 3's analysis of data ownership. Who "owns" the patient data in an EHR system -- the patient, the provider, or the vendor? Legally, the provider is the custodian and HIPAA's covered entity. Practically, the vendor controls the infrastructure, the access, and the terms of service. The patient, whose body generated the data, has access rights under HIPAA but limited practical control.
12.2 Genetic Data: The New Frontier
12.2.1 The Direct-to-Consumer Revolution
In 2007, 23andMe began selling $999 saliva-based genetic testing kits directly to consumers. By 2024, the price had dropped below $100, and the company had collected genetic data from over 14 million customers. Ancestry.com's DNA service had tested over 22 million. Together with smaller competitors, the direct-to-consumer (DTC) genomics industry holds genetic information on tens of millions of individuals.
These companies offer consumers insights into their ancestry, health predispositions, carrier status for genetic conditions, and connections to genetic relatives. The value proposition is compelling: learn about your heritage, understand your health risks, find long-lost family members.
The privacy implications are profound.
12.2.2 Why Genetic Data Is Uniquely Sensitive
Genetic data differs from other forms of personal data in several critical respects:
Immutability. You can change your password, your address, even your name. You cannot change your DNA. Once your genetic data is compromised, it is compromised permanently.
Familial exposure. Your DNA is approximately 50% identical to each of your parents and each of your children, 25% identical to your grandparents and grandchildren, and 12.5% identical to your first cousins. When you submit your DNA to 23andMe, you are not making a decision about only your own data. You are exposing information about every biological relative you have -- including those who may not have consented, and those who may not yet exist (your future children and their descendants).
Predictive power. Genetic data can reveal predispositions to diseases (Alzheimer's, certain cancers, Huntington's), carrier status for genetic conditions, and other health-relevant information. This predictive power creates risks of discrimination by employers, insurers, and others.
Permanence of databases. Genetic databases are not just snapshots -- they are growing, permanent repositories. As genomic science advances, data that seems innocuous today may reveal new information tomorrow. A genetic sample submitted in 2024 may be re-analyzed in 2040 with techniques that don't yet exist, revealing health risks that couldn't be predicted when the sample was collected.
Reflection: Before submitting a DNA sample, you might ask yourself whether your cousins, your siblings, or your future grandchildren would consent to having their genetic information inferred from yours. The answer is unknowable -- which is precisely the problem. Individual consent fails when the data is inherently collective.
12.2.3 Legal Protections: GINA and Its Gaps
The Genetic Information Nondiscrimination Act (GINA), enacted in 2008, prohibits discrimination based on genetic information in two contexts:
-
Employment (Title II): Employers with 15 or more employees cannot use genetic information in hiring, firing, promotion, or other employment decisions. They also cannot request, require, or purchase genetic information about employees or their family members.
-
Health insurance (Title I): Health insurers cannot use genetic information to make coverage or premium decisions.
GINA was an important step, but its limitations are significant:
| What GINA Covers | What GINA Does NOT Cover |
|---|---|
| Employer discrimination | Life insurance discrimination |
| Health insurance discrimination | Disability insurance discrimination |
| Long-term care insurance discrimination | |
| Education discrimination | |
| Housing discrimination | |
| Law enforcement use | |
| Military use |
The gaps are not hypothetical. In 2018, the New York Times reported cases of individuals who lost long-term care insurance after genetic test results revealed Alzheimer's risk. GINA did not protect them.
12.2.4 The Consent Problem in Genetic Testing
The consent challenges with DTC genomics are distinct from -- and more severe than -- the consent problems documented in Chapter 9 for general data collection. Consider what a 23andMe customer consents to:
At the point of purchase: The customer agrees to terms of service that include provisions about data storage, use for research (opt-in), and sharing with third-party partners. The terms are lengthy and technical, but this is at least a moment of active engagement.
For their relatives: The customer's genetic data reveals information about biological relatives who did not consent to anything. A person who has never heard of 23andMe may have their genetic predispositions inferrable because a second cousin submitted a sample. There is no mechanism for these third parties to be notified, let alone to consent or object.
For their future selves: Genetic data is permanent, and genomic science is advancing rapidly. When a customer submits a sample in 2024, they are consenting based on what the data reveals today. But the data will be reanalyzable with future techniques that may reveal information the customer never anticipated -- disease risks not yet understood, ancestry details not yet discoverable, behavioral propensities not yet correlatable. Consent given today may be inadequate for revelations ten years from now.
For corporate transitions: When a customer submits their sample to 23andMe, they are trusting that specific company with their data. But companies are acquired, restructured, and bankrupted. When 23andMe filed for bankruptcy in late 2024, the genetic data of 14 million customers became a corporate asset that could potentially be sold to the highest bidder. The customer who consented to 23andMe's stewardship did not consent to an unknown future owner's stewardship.
Debate Box: Some genetic privacy advocates argue for a "right to genetic oblivion" -- the right to demand that a DTC genomics company destroy your sample and delete your data permanently. 23andMe and others offer this option. But if your data has already been used in aggregate research, incorporated into ancestry databases, or matched with relatives, can deletion truly undo the exposure? What does "deletion" mean when your genetic information has been inferred by others from their own data? These questions reveal the limits of individual rights frameworks when applied to inherently relational data.
12.2.5 Law Enforcement and Genetic Databases: The Golden State Killer
In April 2018, law enforcement arrested Joseph James DeAngelo -- the suspected Golden State Killer, responsible for at least 13 murders and 50 rapes in California between 1974 and 1986. The arrest was made possible not through traditional police work but through forensic genetic genealogy.
Investigators uploaded DNA from crime scenes to GEDmatch, a public genetic genealogy database where users voluntarily share their data to find relatives. The DNA matched distant relatives of DeAngelo. Through traditional genealogical research, investigators narrowed the list of suspects to DeAngelo, then confirmed his identity with DNA collected from a discarded item.
The technique was celebrated as a breakthrough in cold-case investigation. It also raised profound privacy questions:
The consent problem. DeAngelo never submitted his DNA to GEDmatch. His relatives did -- and through their voluntary participation, they inadvertently provided the information needed to identify him. The people who "consented" to GEDmatch's terms of service did not -- and could not -- consent on behalf of their relative.
The scope creep problem. GEDmatch was created for genealogy enthusiasts. Its terms of service at the time did not specifically address law enforcement use. After the Golden State Killer case, GEDmatch changed its terms to require users to opt in to law enforcement matching. But the precedent was set: a database created for one purpose had been repurposed for another -- a textbook contextual integrity violation (Chapter 7).
The equity problem. Forensic genetic genealogy depends on the representation of specific populations in genetic databases. If certain communities are overrepresented (European-descended Americans are heavily represented in consumer genomics), they are more susceptible to identification. As databases grow and diversify, the technique becomes more powerful -- and more concerning.
Character Moment: Eli leaned back in his chair after reading about the Golden State Killer case. "I'm glad they caught him. He was a serial killer. But the method? Using a genealogy database that people uploaded to for fun? That's like finding a murderer by reading everyone's diary."
"Would you feel differently," Dr. Adeyemi asked, "if the database had been purpose-built for law enforcement, with explicit consent for forensic use?"
"Yes," Eli said immediately. "Context matters. That's the whole point of Chapter 7. People shared their DNA to find cousins, not to become a surveillance tool."
12.2.6 The 23andMe Privacy Crisis
In October 2023, 23andMe disclosed a data breach that exposed the personal information and genetic ancestry data of approximately 6.9 million users. Hackers used credential stuffing -- trying username/password combinations leaked from other sites -- to access individual accounts, then exploited 23andMe's "DNA Relatives" feature to scrape data from the matching relatives of compromised accounts.
The breach disproportionately affected users of Ashkenazi Jewish and Chinese ancestry, raising concerns about targeted ethnic data theft. In the aftermath, 23andMe faced dozens of lawsuits and, in late 2024, filed for bankruptcy protection, raising alarming questions about what would happen to the genetic data of its 14 million customers in the event of an asset sale.
This scenario -- a genomics company going bankrupt, with its genetic database potentially sold as an asset to the highest bidder -- was precisely the kind of risk that privacy advocates had warned about since the DTC genomics industry began.
12.3 Biometric Data: Your Body as Your Password
12.3.1 What Is Biometric Data?
Biometric data is data derived from the measurement of a person's physical or behavioral characteristics, used to identify or authenticate their identity. The most common forms include:
| Biometric Type | How It Works | Common Uses |
|---|---|---|
| Fingerprints | Unique ridge patterns on fingertips are captured optically or capacitively | Phone unlocking, border control, law enforcement |
| Facial geometry | The distances and angles between facial features are measured and encoded as a mathematical template | Surveillance cameras, phone unlocking, social media tagging |
| Iris/retina scans | The unique patterns in the iris or retina are captured by specialized cameras | High-security access control, national ID systems |
| Voice recognition | Vocal characteristics (pitch, cadence, formant frequencies) create a voiceprint | Call center authentication, voice assistants |
| Gait analysis | The unique patterns of how a person walks are captured by cameras or floor sensors | Surveillance, health monitoring |
| Vein patterns | The unique pattern of veins in the hand or finger is captured by infrared imaging | Access control (primarily in Japan) |
12.3.2 Why Biometric Data Is Different
Biometric data shares the immutability characteristic of genetic data but adds a dimension of continuous vulnerability. Consider:
You cannot revoke it. If your password is compromised, you change your password. If your biometric template is compromised, you cannot change your face. A stolen biometric is stolen forever. This makes biometric data breaches categorically more serious than other data breaches.
It can be collected without your knowledge or consent. Your face is visible whenever you are in public. Facial recognition systems can capture and identify you without your awareness, let alone your agreement. This is fundamentally different from other forms of authentication, which require your active participation.
It bridges the digital and physical worlds. A data broker who knows your browsing history can target you with ads online. A facial recognition system that knows your face can track your movements through physical space -- where you go, when, and with whom. Biometric surveillance collapses the distinction between online and offline privacy.
It is probabilistic, not deterministic. Biometric matching systems produce confidence scores, not certainties. A facial recognition system might report a "92% match." When the consequences of a false match include arrest and incarceration, this probabilistic nature creates serious risks -- risks that fall disproportionately on those whose demographic groups are less accurately represented in training data.
12.3.3 Gait Analysis and Emerging Biometrics
Beyond the well-known biometric categories, researchers have developed increasingly exotic methods of identification:
Gait analysis identifies individuals by the distinctive patterns of their walking -- stride length, arm swing, posture, pace, and the subtle asymmetries unique to each person's body mechanics. Unlike facial recognition, gait analysis works at a distance, from behind, and even when the subject is wearing a mask or hat. Research published in IEEE Transactions on Pattern Analysis and Machine Intelligence has demonstrated gait recognition accuracy exceeding 90% in controlled conditions.
Keystroke dynamics identifies individuals by their typing patterns -- how long each key is pressed, the intervals between keystrokes, and characteristic error-correction behaviors. This technique can authenticate users continuously during a computer session without any explicit biometric capture.
Heart rate signature uses radar or other remote sensing to identify individuals by the unique electrical signature of their heartbeat. The Pentagon's "Jetson" laser-based system can reportedly identify individuals by their cardiac signature from over 200 meters away.
These emerging biometrics share a troubling characteristic: they can be captured without the subject's knowledge, cooperation, or proximity. A world in which gait, keystroke, and cardiac biometrics are routinely collected is a world in which anonymity in public spaces becomes technically impossible.
12.3.4 Facial Recognition: Promise and Peril
Facial recognition technology has advanced dramatically in the past decade, driven by deep learning and the availability of massive training datasets (some scraped from social media without consent -- Clearview AI, which we'll examine in Chapter 17, scraped over 30 billion images from the internet).
Accuracy disparities. A landmark 2018 study by Joy Buolamwini and Timnit Gebru -- the "Gender Shades" project -- demonstrated that leading commercial facial recognition systems had dramatically higher error rates for darker-skinned women (up to 34.7% error) compared to lighter-skinned men (0.8% error). These disparities have narrowed since 2018 but persist, with the National Institute of Standards and Technology (NIST) consistently finding demographic differences in accuracy across commercial systems.
Deployment contexts. Facial recognition is being deployed in an expanding range of contexts: - Law enforcement: Real-time identification of suspects from security camera footage - Border control: Automated passport verification and traveler screening - Retail: Customer identification, shoplifting prevention, targeted advertising - Schools: Student attendance, campus security - Workplaces: Employee time-tracking, access control - Public spaces: "Smart city" surveillance in parks, transit systems, and streets
Each deployment context raises distinct ethical questions, but all share the fundamental concern that facial recognition enables persistent identification -- the ability to know who you are, wherever you are, without your knowledge or consent.
12.3.5 The Robert Williams Case
On January 9, 2020, Robert Williams, a Black man from the Detroit suburb of Farmington Hills, was arrested in front of his wife and two young daughters and held in custody for 30 hours. He was accused of stealing watches from a Shinola store in downtown Detroit.
The accusation was based on a facial recognition match. The Detroit Police Department had submitted surveillance footage from the store to a state facial recognition system, which returned a match to Williams's driver's license photo. A detective placed Williams's photo in a "digital lineup" shown to a Shinola security guard, who identified him as the shoplifter.
Williams was not the shoplifter. The facial recognition system had generated a false match -- a match that a human investigator failed to independently verify before pursuing an arrest.
The case became a landmark in the debate over facial recognition in law enforcement for several reasons:
The accuracy problem. Williams's case was not an outlier. NIST studies have consistently shown that facial recognition systems have higher false positive rates for Black individuals, particularly Black men. The systems are trained on datasets that overrepresent lighter-skinned faces, and they perform less reliably on darker skin.
The investigation failure. The detective who pursued the arrest treated the facial recognition output as conclusive evidence rather than as an investigative lead requiring independent corroboration. This is a recurring problem: technology that is designed to assist human decision-making instead replaces it.
The contextual injustice. Williams was arrested in front of his children. He was held in a cell for 30 hours. The charges were eventually dropped, but the experience of wrongful arrest -- the humiliation, the fear, the violation of dignity -- cannot be undone by dropping charges.
The systemic pattern. Williams was the first publicly documented case of wrongful arrest by facial recognition in the United States. He was not the last. In 2020, Michael Oliver and Nijeer Parks were also wrongfully accused based on facial recognition misidentification. All three were Black men.
Character Moment: Eli presented the Robert Williams case to Dr. Adeyemi's class. "Robert Williams is from Farmington Hills," he said. "That's twenty minutes from where I grew up. This isn't an abstract hypothetical. This is my community."
He paused. "The Detroit Police Department says they've since changed their policy -- facial recognition is supposed to be an investigative tool, not the sole basis for arrest. But how do we verify that? How do we know they're following their own policy? And who is liable when they don't?"
"Those are governance questions," Dr. Adeyemi said. "And we'll address them in Chapter 17. But you've identified the core issue: technology that makes mistakes, deployed in a system that amplifies those mistakes, causing harm that falls disproportionately on people who already face systemic injustice."
"It's not a bug in the algorithm," Eli said. "It's the algorithm interacting with a bug in the system."
12.4 Illinois BIPA: A Model Biometric Privacy Law
12.4.1 The Law
The Illinois Biometric Information Privacy Act (BIPA), enacted in 2008, is the most significant biometric privacy law in the United States and has become a model for other jurisdictions. BIPA requires:
- Written informed consent before collecting, capturing, purchasing, or otherwise obtaining a person's biometric identifier or biometric information.
- Written retention and destruction policy that specifies when biometric data will be destroyed (no later than 3 years from the last interaction, or the purpose of collection has been satisfied, whichever comes first).
- No profit from biometric data. Entities in possession of biometric data cannot sell, lease, trade, or otherwise profit from it.
- Reasonable security. Biometric data must be stored using reasonable standards of care, at least as protective as those used for other confidential and sensitive information.
- Private right of action. Individuals whose biometric data has been collected in violation of BIPA can sue for statutory damages of $1,000 per negligent violation and $5,000 per intentional or reckless violation.
12.4.2 Why BIPA Matters
BIPA's most consequential feature is the private right of action -- the ability of individuals to sue. Most privacy statutes rely on enforcement by government agencies (the FTC, state attorneys general, data protection authorities). Government enforcement is inherently limited by resources, political will, and competing priorities.
BIPA's private right of action has generated hundreds of lawsuits, including:
- Rosenbach v. Six Flags (2019): The Illinois Supreme Court ruled that individuals do not need to demonstrate actual harm to sue under BIPA -- the violation of the statute itself is sufficient. This removed a major barrier to privacy litigation (the standing problem).
- Class action against Facebook (2020): Facebook settled for $650 million over its use of facial recognition in the "tag suggestions" feature without obtaining written consent from Illinois users.
- Class action against Clearview AI: Multiple suits alleging that Clearview scraped facial images from the internet and built a facial recognition database without consent, violating BIPA.
- BIPA v. TikTok (2021): TikTok settled for $92 million over allegations of collecting biometric data (facial geometry from uploaded videos) from Illinois users without consent.
12.4.3 Criticism and Replication
BIPA has been criticized by industry groups as overly burdensome, particularly the private right of action, which they argue encourages litigation rather than compliance. In 2023, the Illinois Supreme Court ruled in Cothron v. White Castle that each individual scan of a biometric identifier constitutes a separate violation -- meaning that an employer who scans an employee's fingerprint at every shift for several years could face millions of dollars in liability per employee. This ruling intensified industry opposition.
Despite criticism, BIPA has inspired similar legislation in Texas, Washington, and other states, and its influence is visible in proposals for federal biometric privacy legislation. The European Union's AI Act (Chapter 21) bans certain uses of real-time biometric identification in public spaces, reflecting a similar concern about biometric surveillance.
Debate Box: BIPA's private right of action is its most powerful and most controversial feature. Supporters argue that it provides meaningful deterrence that government enforcement alone cannot achieve. Critics argue that it creates a "litigation lottery" where companies face billions in potential damages for technical violations that may not cause actual harm. Is the private right of action a necessary corrective for the enforcement gap in privacy law, or an unjust burden on businesses? Consider: would BIPA have changed Facebook's behavior if the only enforcement mechanism were FTC complaints?
12.5 The Special Sensitivity Argument
12.5.1 Why These Categories Need Extra Protection
Throughout this chapter, we've documented the unique characteristics of health, genetic, and biometric data. The case for heightened protection rests on several converging arguments:
The irreversibility argument. Most privacy harms can be at least partially mitigated. A compromised credit card number can be replaced. A leaked email address can be abandoned. But compromised genetic data reveals permanent information about your biology, and a stolen biometric template cannot be "re-issued." The consequences of a breach are permanent.
The inference argument. Health, genetic, and biometric data enable inferences that go far beyond the data itself. A genetic test revealing BRCA1 mutation status doesn't just tell you about cancer risk -- it reveals information about your parents, siblings, and children. A facial recognition template doesn't just authenticate your identity -- it enables persistent tracking across every context where cameras exist. The inferential reach of these data types magnifies their sensitivity.
The discrimination argument. Health conditions, genetic predispositions, and physical characteristics have historically been -- and continue to be -- bases for discrimination. People have been denied insurance, employment, and housing based on health status. Genetic information reveals ancestry and ethnic heritage, creating risks of ethnic targeting. Facial recognition systems exhibit racial bias. These data types exist in a context of historical and ongoing discrimination that general data types do not.
The power asymmetry argument (Chapter 5). The organizations that collect and process health, genetic, and biometric data -- hospitals, insurance companies, employers, law enforcement -- have enormous power relative to the individuals whose data they hold. A patient in a hospital, a job applicant undergoing a background check, a pedestrian walking past a surveillance camera -- all are in positions of vulnerability where meaningful consent is difficult or impossible.
12.5.2 Care Ethics and Sensitive Data
The care ethics framework from Chapter 6 is particularly relevant to health, genetic, and biometric data. Care ethics asks us to attend to relationships of vulnerability and to take responsibility for those who depend on us.
Health data and care relationships. When a patient shares health information with a doctor, they do so within a relationship of trust and vulnerability. They are sick, frightened, or seeking help. The doctor holds asymmetric power -- expertise, institutional authority, access to treatment. A care ethics perspective demands that this relationship of trust be honored, that the patient's vulnerability not be exploited, and that the use of their data serve the care relationship that motivated its disclosure.
When VitraMed processes patient data, it inherits this care obligation. The data did not originate in a commercial transaction. It originated in a moment of human vulnerability -- a patient seeking help. Using that data for purposes beyond care (monetization, marketing, undisclosed research partnerships) violates the trust that generated it.
Genetic data and familial relationships. Care ethics emphasizes our responsibilities to particular others. Genetic data implicates relationships of care in an unusually direct way: the decision to share your genetic data affects the people you care about most -- your parents, children, and siblings. A care ethics analysis would require considering the impact on these relationships before disclosing genetic information, and would be skeptical of frameworks that treat genetic data as belonging solely to the individual who submitted a sample.
Biometric data and community relationships. When facial recognition is deployed in a neighborhood, it affects the entire community -- changing the experience of public space from one of relative anonymity to one of persistent identification. A care ethics perspective attends to this communal impact, asking not just "were individual rights violated?" but "what has this done to the community's experience of safety, trust, and freedom?"
Connection to Chapter 6: The five-framework analysis from Chapter 6 produces especially strong convergence on sensitive data types. Utilitarianism: the potential harms of misuse are severe and the affected populations are large. Deontology: using health data as a mere means (for profit, for surveillance) violates Kantian dignity. Virtue ethics: a virtuous practitioner would exercise heightened care with sensitive data. Care ethics: these data types arise from relationships of vulnerability that demand protection. Justice theory: behind the veil of ignorance, you would demand robust protections for data types that could expose you to discrimination. When all five frameworks agree, the ethical ground is strong.
12.5.3 The Regulatory Patchwork
The heightened sensitivity of health, genetic, and biometric data is widely recognized, but the legal frameworks that protect these data types are fragmented and inconsistent:
| Data Type | U.S. Federal Law | U.S. State Law | EU Law |
|---|---|---|---|
| Health data | HIPAA (covered entities only) | State health privacy laws (vary) | GDPR Art. 9 (special category) |
| Genetic data | GINA (employment + health insurance only) | Some states (CA, IL, others) | GDPR Art. 9 + additional member state laws |
| Biometric data | No federal law | IL BIPA, TX CUBI, WA (others pending) | GDPR Art. 9; AI Act (biometric identification) |
The patchwork creates several problems. First, protection depends on geography -- an Illinois resident has strong biometric privacy rights; a resident of a neighboring state may have none. Second, protection depends on the collector rather than the data type -- health data held by a hospital is protected under HIPAA, but the same information entered into a fitness app is not. Third, enforcement is inconsistent -- HIPAA violations can result in millions in fines from HHS, while genetic privacy violations outside GINA's narrow scope may have no remedy at all.
The European approach, by contrast, treats health, genetic, and biometric data as "special categories" under GDPR Article 9, subject to heightened protections regardless of who collects them or where they are collected. Processing of special category data is generally prohibited unless one of several specific conditions is met (explicit consent, vital interests, substantial public interest, etc.). This approach -- category-based rather than entity-based -- provides more consistent protection, though it is not without its own limitations.
Dr. Adeyemi summarized the comparison: "The American system asks, 'Who has the data?' and protects it only if the holder falls within a regulated category. The European system asks, 'What kind of data is it?' and protects it based on its inherent sensitivity. For data types that can flow freely between regulated and unregulated entities -- as health data increasingly does -- the European approach provides more robust protection."
12.6 VitraMed Under HIPAA Scrutiny
12.6.1 The First Incident
Six months after Mira began her privacy remediation plan (Chapter 10), VitraMed experienced its first privacy incident.
A clinic nurse in Tempe, Arizona, reported that she had been able to access patient records from a different clinic -- one in Raleigh, North Carolina -- through VitraMed's analytics dashboard. The access was not intentional; the nurse had been running a routine report and noticed that the results included patient records from a facility she had never heard of.
The root cause was a misconfigured access control in VitraMed's multi-tenant database architecture. Clinic A's analytics queries, under certain conditions, could return records from Clinic B. No patient names were visible in the analytics interface -- the dashboard showed de-identified aggregate data -- but the underlying query was pulling identifiable records from the wrong clinic's partition before aggregating them.
12.6.2 Assessing the Severity
The incident was minor by industry standards: - No patient names, addresses, or Social Security numbers were displayed to unauthorized users - The exposure was limited to aggregated analytics (patient counts by diagnosis category, average treatment duration) - The misconfiguration was present for 11 days before discovery - There was no evidence that any unauthorized user noticed or exploited the issue - The nurse who discovered it reported it immediately
But the implications were serious: - HIPAA violation: Any unauthorized access to PHI, even if the PHI was not viewed in identifiable form, is potentially a HIPAA violation. The underlying query accessed identifiable records, even if the displayed output was aggregated. - Breach notification question: HIPAA requires notification when unsecured PHI has been "accessed, acquired, used, or disclosed" in an unauthorized manner. Did this incident meet that threshold? VitraMed's legal team and privacy team disagreed. - Trust impact: If the incident became public, VitraMed's reputation with its clinic clients -- who had entrusted it with their patients' data -- would be damaged.
12.6.3 Mira's Response
Mira, now informally leading VitraMed's privacy remediation, pushed for full transparency.
"We notify the affected clinics, we report to HHS, and we fix the architecture," she argued. "The HIPAA Breach Notification Rule has a risk assessment provision -- if we can demonstrate that the probability of PHI being compromised is low, we may not need to notify patients. But we absolutely notify the clinics and HHS."
VitraMed's lawyer was more cautious. "Voluntary disclosure exposes us to regulatory scrutiny. If HHS investigates and finds other issues, we've invited them in."
"If they find other issues," Mira said, "those issues exist whether we invite them or not. I'd rather they find us forthcoming than find us hiding."
Vikram sided with Mira. VitraMed notified the affected clinics and filed a breach report with HHS. The investigation concluded with no fine -- the Office for Civil Rights determined that VitraMed had acted in good faith and remediated the issue promptly. But the incident was logged, and VitraMed was now on HHS's radar.
Foreshadowing: This minor incident -- quickly caught, promptly reported, efficiently resolved -- stands in contrast to the larger crisis that will unfold in Chapter 30. The patterns that Mira identified in her remediation plan -- insufficient access controls, multi-tenant architecture vulnerabilities, ambiguity about breach notification thresholds -- are the same patterns that will produce a far more serious incident when VitraMed is larger, more complex, and under greater commercial pressure. Minor incidents are often previews of major ones.
12.7 Eli's Detroit: Biometric Surveillance and Community Impact
12.7.1 Facial Recognition in the Motor City
Detroit's adoption of facial recognition technology has made it one of the most surveilled cities in the United States. Project Green Light, launched in 2016, connects real-time security camera feeds from hundreds of businesses, gas stations, churches, and public housing developments to the Detroit Police Department's Real Time Crime Center, where officers can run facial recognition on live footage.
The program has been controversial since its inception. Civil liberties organizations, including the ACLU of Michigan, have documented its disproportionate impact on Black residents, who constitute approximately 77% of Detroit's population. The Robert Williams wrongful arrest (Section 12.3.4) occurred within the context of Project Green Light.
12.7.2 Community Organizing
Eli's political science training gave him a framework for understanding what was happening in Detroit. "This is a classic case of what Chapter 5 calls the power asymmetry in data collection," he wrote in a research paper. "The city decided to deploy facial recognition. The communities most affected -- disproportionately Black, disproportionately low-income -- were not meaningfully consulted. The technology was chosen by police and city officials; the costs are borne by residents."
Eli documented several community responses:
The Detroit Community Technology Project trained residents to understand surveillance technology, document its deployment, and advocate for oversight.
The Algorithmic Justice League (founded by Joy Buolamwini, whose Gender Shades research documented facial recognition bias) provided technical expertise and national visibility to local organizing efforts.
City Council advocacy resulted in a 2023 ordinance requiring the Detroit Police Department to seek City Council approval before purchasing new surveillance technology. The ordinance did not ban facial recognition -- Eli argued it should have -- but it created a governance mechanism that had not previously existed.
12.7.3 The Broader Pattern
Detroit is not unique. San Francisco banned government use of facial recognition in 2019. Portland, Oregon extended the ban to private businesses. Several other cities have enacted moratoriums or restrictions. The European Union's AI Act restricts real-time biometric identification in public spaces, with exceptions for law enforcement.
But the pattern is fragmented. For every city that restricts facial recognition, dozens deploy it without restriction. The patchwork of local ordinances creates a landscape where your biometric privacy depends on which side of a city line you stand on -- a geographic lottery that bears no relationship to the severity of the risk.
Applied Framework: Evaluate Detroit's Project Green Light using Nissenbaum's contextual integrity framework (Chapter 7). In the context of public safety, certain information flows (reporting crimes, identifying suspects from specific evidence) are normatively appropriate. But Project Green Light changes the norms by enabling persistent, real-time identification of everyone in the vicinity of a participating business. The information flow is no longer "a witness identifies a suspect" but "a camera identifies everyone, and the police database stores the result." Does this change violate the norms of the public safety context? Or does it simply apply the existing norms with new efficiency?
12.8 Case Study References
23andMe and the Golden State Killer
The use of genetic genealogy databases to identify the Golden State Killer in 2018 represents both a triumph of investigative science and a watershed moment for genetic privacy. This case study examines the technique (forensic genetic genealogy), the ethical questions it raises, the subsequent changes to GEDmatch's consent policies, and the broader implications for the millions of people who have voluntarily submitted their DNA to consumer genomics companies.
Key questions for analysis: - Apply the contextual integrity framework to the use of GEDmatch for law enforcement purposes. What were the prevailing norms of the genealogy context, and did the law enforcement use violate them? - Evaluate the consent given by GEDmatch users. Did they consent to having their genetic data used to identify their relatives as criminal suspects? Could they have? - The Golden State Killer was guilty. Does the rightness of the outcome justify the method? Apply at least two ethical frameworks from Chapter 6. - What governance framework would allow forensic genetic genealogy for serious violent crimes while protecting the genetic privacy of the general population?
Full case study analysis: case-study-01.md
Robert Williams: Wrongful Arrest by Facial Recognition
Robert Williams's wrongful arrest in January 2020 was the first publicly documented case of facial recognition misidentification leading to an arrest in the United States. This case study examines the technology failure, the investigative failure, the systemic factors that enabled the wrongful arrest, and the policy responses that followed.
Key questions for analysis: - The facial recognition system produced a false match. The detective failed to independently verify the match before pursuing an arrest. Where does responsibility lie -- with the technology, the officer, the department's policy, or the system that deployed the technology? - Apply the care ethics framework to this case. Who was vulnerable? Who had a duty of care? How was that duty violated? - Research the accuracy disparities in facial recognition across demographic groups. How do these disparities interact with existing patterns of racial bias in policing? - Detroit's subsequent policy requires that facial recognition be used only as an "investigative lead," not as the sole basis for arrest. Is this policy sufficient? What additional safeguards would you recommend?
Full case study analysis: case-study-02.md
12.9 Chapter Summary
Key Concepts
- HIPAA: The primary U.S. health data privacy law, establishing protections for PHI held by covered entities and business associates, but with significant scope gaps that leave health-related data from apps, wearables, and DTC genomics unprotected
- Genetic data: Uniquely sensitive due to immutability, familial exposure, predictive power, and permanence; protected by GINA against employment and health insurance discrimination but not against many other forms of discrimination
- Forensic genetic genealogy: The use of consumer genetic databases for law enforcement identification, raising questions about consent, contextual integrity, and the collective nature of genetic information
- Biometric data: Physical and behavioral identifiers (face, fingerprints, gait, voice) that are irrevocable if compromised and can be collected without knowledge or consent
- BIPA: Illinois's biometric privacy law, notable for its private right of action, which has generated significant litigation and become a model for other jurisdictions
- Robert Williams case: The first documented wrongful arrest by facial recognition in the U.S., illustrating accuracy disparities, investigative failures, and the disproportionate impact of biometric surveillance on Black communities
Key Debates
- Is HIPAA adequate for the digital health data landscape, or does it need fundamental reform?
- Should law enforcement be permitted to use consumer genetic databases for criminal investigation? Under what conditions?
- Should facial recognition technology be banned in public spaces, regulated, or deployed without restriction?
- Is the BIPA model of private right of action the right approach to biometric privacy enforcement, or does it create perverse incentives?
- Does the immutability of biometric and genetic data create an absolute argument against certain uses, or can risks be managed through governance?
Applied Framework
The Sensitive Data Assessment asks: (1) Is the data immutable? (2) Does it reveal information beyond the data subject (familial, communal)? (3) Could it be the basis for discrimination? (4) Can it be collected without the subject's knowledge? (5) In what relationship of vulnerability was the data generated? (6) What would a care ethics analysis require? For each affirmative answer, the case for heightened protection strengthens.
What's Next
Part 2 is now complete. You have examined privacy from multiple angles: its theoretical foundations (Chapter 7), the surveillance apparatus that threatens it (Chapter 8), the consent mechanisms that are supposed to protect it (Chapter 9), the design principles and technologies that can embed it (Chapter 10), the economic forces that shape it (Chapter 11), and the special challenges posed by the most sensitive data about our bodies and identities (this chapter).
In Part 3: Algorithmic Systems and AI Ethics, we turn from privacy to a related but distinct challenge: what happens when data is not just collected and stored but used to make decisions? Chapter 13, "How Algorithms Shape Society," introduces the algorithmic systems that sort, rank, recommend, and decide -- systems that touch nearly every aspect of modern life and raise questions of bias, fairness, transparency, and accountability that will occupy us for seven chapters.
Before moving on, complete the exercises and quiz.