39 min read

In Chapter 16, we examined transparency and explainability — the demand that algorithmic systems be understandable. But understanding a system is not the same as holding anyone responsible for what it does. A perfectly transparent algorithm can...

Learning Objectives

  • Define the accountability gap and explain why algorithmic systems create novel challenges for traditional accountability frameworks
  • Distinguish between internal and external algorithmic audits and evaluate the strengths and limitations of each approach
  • Describe at least three methods of algorithmic auditing and identify appropriate use cases for each
  • Explain the structure and process of an Algorithmic Impact Assessment (AIA) and critically evaluate its limitations
  • Analyze the 'many hands' problem and its implications for liability in complex sociotechnical systems
  • Compare strict liability, negligence, and product liability frameworks as applied to algorithmic harm
  • Evaluate the emerging institutional landscape for algorithmic accountability, including audit firms, regulatory sandboxes, and legislative proposals

Chapter 17: Accountability and Audit

"Accountability is the measure of a leader's height." — Jeffrey Benjamin, governance scholar

Chapter Overview

In Chapter 16, we examined transparency and explainability — the demand that algorithmic systems be understandable. But understanding a system is not the same as holding anyone responsible for what it does. A perfectly transparent algorithm can still cause harm, and if no one is responsible for that harm, transparency is cold comfort.

This chapter confronts what may be the most consequential governance challenge in the algorithmic age: the accountability gap. When an algorithm denies you a loan, rejects your resume, flags you as a fraud risk, or recommends that your medical treatment be denied, who is responsible? The developer who wrote the code? The company that deployed it? The data provider whose biased training set shaped the output? The regulator who approved it — or failed to regulate it at all?

The answer, disturbingly often, is no one. Or rather, everyone points to someone else. The developer says they built a tool, not a policy. The deployer says they trusted the vendor's claims. The data provider says the data reflects reality. The regulator says they lack jurisdiction or expertise. Meanwhile, the person harmed has no clear path to redress.

This chapter maps the accountability gap, examines the emerging tools designed to close it — algorithmic auditing, impact assessments, liability frameworks — and wrestles honestly with their limitations.

In this chapter, you will learn to: - Diagnose why algorithmic systems create accountability gaps that traditional governance cannot easily address - Evaluate different approaches to algorithmic auditing and select appropriate methods for specific contexts - Construct an Algorithmic Impact Assessment and identify where the process is most likely to fail - Navigate the legal landscape of algorithmic liability - Assess the institutional infrastructure emerging to support algorithmic accountability


17.1 The Accountability Gap

17.1.1 What We Mean by Accountability

Accountability, in its simplest form, involves three elements:

  1. Answerability: Someone must be able to explain and justify the decision
  2. Attributability: The decision must be attributable to an identifiable person or entity
  3. Enforceability: There must be consequences for unjustified decisions — sanctions, penalties, remedies

Traditional accountability works because these three elements are usually present. When a bank manager denies your loan application, the manager can explain why (answerability), the bank is the identifiable decision-maker (attributability), and you can appeal to a regulator or court (enforceability). The system is imperfect, but the architecture of accountability exists.

Algorithmic decision-making disrupts all three elements simultaneously.

17.1.2 A Historical Perspective on Accountability

Before examining how algorithms break accountability, it is worth noting that accountability challenges are not entirely new. Bureaucratic systems have always diffused responsibility — a welfare applicant denied benefits by "the system" faces a recognizably similar frustration to someone denied a loan by an algorithm. Max Weber's analysis of bureaucratic rationality (1922) warned that modern administrative systems would create an "iron cage" in which individuals are processed according to impersonal rules, with no single person responsible for the outcome.

What makes algorithmic accountability different from bureaucratic accountability is a matter of degree and kind:

  • Speed and scale. A human bureaucracy processes hundreds or thousands of decisions. An algorithm can process millions per second. The volume of potentially harmful decisions — and the number of people who may need recourse — is orders of magnitude greater.
  • Opacity. Bureaucratic rules, however complex, are written in human language and can in principle be read, understood, and challenged. Machine learning models encode patterns in millions of numerical parameters that do not translate into human-readable rules.
  • Adaptability. Bureaucratic rules change through deliberate amendment. ML models change through retraining — a process that may alter the system's behavior in ways that even its developers do not fully anticipate or understand.
  • Perceived authority. Research in human-computer interaction consistently finds that people attribute greater objectivity and authority to automated decisions than to human ones. An algorithmic decision feels more "scientific," more "neutral," even when it is neither. This perception makes algorithmic decisions harder to challenge — psychologically, if not legally.

Understanding these distinctions matters because governance responses designed for bureaucratic accountability — such as administrative appeals, ombudsman offices, and freedom of information requests — may be necessary but insufficient for algorithmic accountability. New institutional forms are required.

17.1.3 How Algorithms Break Accountability

Answerability collapses when the system's reasoning is opaque. As we explored in Chapter 16, many machine learning models — particularly deep neural networks — cannot provide human-interpretable explanations for their outputs. Even when post-hoc explanations are available (LIME, SHAP), they approximate the model's reasoning rather than revealing it. The system can tell you what it decided but not why in any meaningful sense.

Attributability fragments when the system involves many actors. A healthcare risk model might incorporate data collected by hospitals, cleaned by a data vendor, features engineered by one team of developers, a model architecture selected by another team, deployed by a platform company, and used by a hospital administrator who may not understand how it works. When this system misclassifies a patient and they are denied care, attributing the decision to any single actor is genuinely difficult.

Enforceability weakens when the legal framework hasn't caught up with the technology. Existing anti-discrimination law, for instance, was designed for human decision-makers. Proving that an algorithm discriminated often requires access to the algorithm itself — access that the owner resists providing, citing trade secrets. And even when discrimination is demonstrated, the question of who discriminated (the algorithm? the developer? the deployer?) creates legal uncertainty.

Dr. Adeyemi framed the problem starkly: "We have built systems that make consequential decisions about people's lives — their credit, their health, their freedom — while simultaneously making it harder to identify who is responsible, harder to explain the reasoning, and harder to challenge the outcome. That is not a technical glitch. That is a governance crisis."

17.1.4 The VitraMed Accountability Chain

Consider a scenario that Mira has been grappling with since the previous chapter's discussion of transparency at VitraMed.

VitraMed's patient risk model assigns a score to incoming patients at partner clinics. Patients above a certain threshold are flagged as "high complexity" — which, in practice, means they may be routed to different care pathways, prescribed different interventions, or in some cases, deemed ineligible for certain elective procedures because of predicted complication rates.

Now imagine the model makes a wrong prediction. A patient who would have benefited from a procedure is denied it based on an inaccurate risk score. The patient's health deteriorates. Who is liable?

Actor Role Claim of Non-Liability
VitraMed (developer) Built and trained the model "We provided a decision-support tool, not a clinical decision. The physician makes the final call."
Partner clinic (deployer) Integrated the model into workflow "We relied on VitraMed's validated model. We're not AI experts."
Hospital administrator Set the threshold score for care pathway routing "I followed the vendor's recommended thresholds."
Physician Made the care decision informed by the score "The system flagged the patient as high-risk. I exercised clinical judgment consistent with the information I was given."
Data providers Supplied training data "We provided accurate historical records. How they were used is not our responsibility."

Every actor in the chain has a plausible claim of non-liability. The harm is real, but responsibility has been distributed into invisibility.

"That's the accountability gap in action," Mira said, staring at the chart. "It's not that nobody is responsible. It's that the responsibility is so distributed that it's functionally the same as if nobody is."

Recurring Theme — The Accountability Gap: This is the core manifestation of the accountability gap theme that has been building since Chapter 1. When systems are designed so that no single actor bears clear responsibility for outcomes, the people harmed by those systems have no effective recourse. The gap is not accidental — it is a structural feature of how algorithmic systems are built, sold, and deployed.


17.2 Algorithmic Auditing: Concepts and Methods

17.2.1 What Is an Algorithmic Audit?

An algorithmic audit is a systematic examination of an algorithmic system to evaluate its behavior, impacts, and compliance with relevant standards. The concept borrows from financial auditing — the idea that independent, structured review can surface problems that internal processes miss — but adapts it to the distinctive challenges of algorithmic systems.

Algorithmic audits can examine: - Accuracy: Does the system perform as claimed? - Fairness: Does the system produce disparate outcomes across protected groups? - Transparency: Can the system's reasoning be understood and explained? - Compliance: Does the system conform to legal requirements and organizational policies? - Safety: Does the system operate within acceptable risk parameters? - Impact: What are the system's effects on individuals, communities, and society?

17.2.2 Internal vs. External Auditing

A fundamental distinction in algorithmic auditing is between internal and external audits.

Internal audits are conducted by the organization that developed or deployed the system. They have the advantage of full access to the system — code, training data, model architecture, performance metrics — and deep contextual knowledge of how the system is used.

Their limitation is independence. An internal audit team may face organizational pressure to produce favorable results, may share the same assumptions and blind spots as the development team, and may lack the perspective that comes from standing outside the system.

"I've seen internal audits at NovaCorp that were rigorous and valuable," Ray Zhao told the class. "And I've seen others that were compliance theater — beautifully documented exercises designed to produce a reassuring report without actually challenging anything. The difference isn't the methodology. It's whether the audit team has genuine independence and the authority to report unwelcome findings."

External audits are conducted by independent third parties — consulting firms, academic researchers, civil society organizations, or regulatory bodies. They bring independence and potentially different perspectives, but they face access challenges. Algorithmic systems are often proprietary, and companies resist sharing code, training data, or model details with outsiders.

Dimension Internal Audit External Audit
Access Full access to system internals Often limited or negotiated access
Independence Limited — organizational pressures Greater — but funding sources may create bias
Contextual knowledge Deep understanding of deployment context May lack operational context
Cost Lower marginal cost (existing staff) Higher direct cost
Credibility Lower public credibility Higher public credibility
Frequency Can be continuous or frequent Typically periodic
Scope Risk of narrow scope (testing what's convenient) More likely to test what matters

The most effective accountability regimes combine both. Internal audit for continuous monitoring; external audit for periodic independent verification.

17.2.3 Audit Methods

Audit Studies

Audit studies (sometimes called "correspondence studies" or "field experiments") test algorithmic systems by submitting controlled inputs and observing outputs. Researchers systematically vary characteristics — race, gender, age, location — while holding other factors constant, to detect whether the system produces different outcomes for different groups.

This method has deep roots in civil rights research. In housing discrimination studies dating to the 1970s, researchers sent matched pairs of testers — identical in all relevant respects except race — to inquire about apartments. The same method now applies to algorithmic systems.

Example: In 2019, researchers tested Apple Card's credit limit algorithm by submitting applications from married couples with identical financial profiles. In multiple cases, husbands received credit limits 10-20 times higher than their wives. The audit study design — identical financial inputs, different gendered outputs — provided compelling evidence of disparate treatment, even though neither the researchers nor Apple could initially explain the mechanism.

Strengths: Does not require access to the system's internals. Tests real-world behavior. Results are intuitive and communicable. Strong evidential value in legal and regulatory proceedings.

Limitations: Can only detect disparities in outcomes, not explain their causes. Labor-intensive. May violate terms of service. Cannot test all possible input combinations.

Disparate Impact Testing

Disparate impact testing is a statistical method that examines whether an algorithmic system produces significantly different outcomes for different demographic groups, even if the system does not explicitly use demographic characteristics as inputs.

The legal standard in U.S. employment law (established in Griggs v. Duke Power Co., 1971) holds that a practice with disparate impact on a protected group is unlawful unless the employer can demonstrate that it is job-related and consistent with business necessity. The four-fifths rule — if the selection rate for a protected group is less than four-fifths (80%) of the rate for the group with the highest selection rate, there is evidence of adverse impact — provides a practical threshold.

Applied to algorithmic systems:

Selection rate for Group A: 60%
Selection rate for Group B: 40%
Ratio: 40/60 = 0.67 (below the 0.80 threshold)
→ Evidence of disparate impact

Strengths: Quantitative, replicable, legally grounded. Does not require proof of intent. Can be applied to any system with measurable outcomes and demographic data.

Limitations: Requires access to demographic data (which may not be collected). Focuses on group-level disparities, may miss individual harms. Does not explain why disparities exist. The four-fifths threshold is a guideline, not an absolute standard.

User Experience Audits

User experience (UX) audits evaluate algorithmic systems from the perspective of the people affected by them. Rather than testing the system's statistical properties, UX audits examine:

  • How is the system's output communicated to users?
  • Do users understand that an algorithmic decision has been made?
  • Can users access an explanation of the decision?
  • Is there a meaningful appeal or redress process?
  • How do different user populations experience the system?

This method is particularly valuable for detecting harms that statistical testing misses. A system may be statistically fair in aggregate but practically harmful because of how its outputs are communicated (or not communicated) to affected individuals.

Sofia Reyes, from the DataRights Alliance, emphasized this approach in a guest presentation: "We can run all the statistical tests we want. But if a person is denied housing by an algorithm and they don't even know an algorithm was involved — if they receive a form letter that says 'your application was not successful' with no explanation, no appeal process, and no human contact — then the system is unaccountable regardless of what the statistical audit shows. Accountability isn't just about whether the system is fair in aggregate. It's about whether each individual who is affected has meaningful recourse."

The Power Asymmetry in Practice: Notice how auditing itself reflects the power asymmetry theme. Who has the resources and access to conduct audits? Who decides what gets audited and what standards are applied? Who sees the results? If auditing is controlled by the same institutions that build and deploy algorithmic systems, it risks becoming another mechanism of institutional self-legitimation rather than genuine accountability.


17.3 Algorithmic Impact Assessments (AIAs)

17.3.1 The Concept

An Algorithmic Impact Assessment (AIA) is a structured process for evaluating the potential impacts of an algorithmic system before it is deployed — and for monitoring its actual impacts throughout its lifecycle. The concept draws on environmental impact assessments (EIAs), which have been required for major construction projects since the 1970s, and privacy impact assessments (PIAs), which are now standard practice in data governance.

The core logic is proactive rather than reactive: instead of waiting for harm to occur and then investigating, AIAs require organizations to anticipate, evaluate, and mitigate potential harms as a condition of deployment.

17.3.2 AIA Structure and Process

A comprehensive AIA typically includes the following components:

1. System Description - What does the system do? - What decisions does it make or inform? - What data does it use? - Who built it, and who operates it?

2. Stakeholder Identification - Who is affected by the system's decisions? - Are there populations that are disproportionately affected? - Have affected communities been consulted?

3. Risk Assessment - What harms could the system cause? - What is the likelihood and severity of each harm? - Are there disparate impacts across demographic groups? - What happens when the system makes errors?

4. Mitigation Measures - What steps have been taken to reduce identified risks? - Are there human oversight mechanisms? - Is there a process for appealing algorithmic decisions?

5. Monitoring Plan - How will the system's performance be tracked after deployment? - What metrics will be monitored? - What triggers a reassessment?

6. Public Reporting - What information about the system will be made publicly available? - How will affected individuals be notified that they are subject to algorithmic decision-making?

17.3.3 The Canadian Directive on Automated Decision-Making

Canada provides one of the most developed examples of mandatory AIAs for government systems. The Directive on Automated Decision-Making, issued by the Treasury Board of Canada Secretariat in 2019, requires federal agencies to:

  1. Complete an Algorithmic Impact Assessment before deploying any automated decision system
  2. Assign a risk level (I through IV) based on the system's potential impacts on individuals and communities
  3. Implement governance measures proportional to the risk level, including: - Level I (low impact): Basic documentation and testing - Level II (moderate impact): Peer review and notification to affected individuals - Level III (high impact): External review, explanation on request, human override - Level IV (very high impact): External audit, legal review, public reporting, meaningful human involvement in every decision

This graduated approach recognizes that not all algorithmic systems carry the same risk and that governance requirements should be calibrated to the stakes involved.

17.3.4 Limitations of AIAs

AIAs are a significant governance innovation, but they face real limitations:

Self-assessment bias. When the organization conducting the AIA is also the organization deploying the system, there is a structural incentive to underestimate risks. This is the same problem that undermines internal audits — but in the AIA context, it occurs before the system is even deployed, when the pressure to proceed is greatest.

Participation gaps. AIAs are supposed to include stakeholder consultation, but genuine consultation — particularly with marginalized communities — requires time, resources, and trust that are often absent. In practice, "stakeholder consultation" frequently means soliciting input from easily accessible groups (industry partners, academic advisors) while the people most affected by the system (welfare recipients, job applicants, patients) are not at the table.

Snapshot problem. An AIA evaluates a system at a specific point in time, but algorithmic systems change. Models are retrained, data distributions shift, deployment contexts evolve. An AIA completed at deployment may not reflect the system's behavior six months later.

Enforcement gap. Even the best AIA is only useful if its findings are acted upon. Without regulatory enforcement — and without consequences for deploying systems that the AIA identifies as high-risk — the assessment becomes a compliance exercise rather than a governance mechanism.

Dr. Adeyemi drew a parallel: "Environmental impact assessments have been mandatory for decades, and they've prevented some bad projects. But they've also become a box-checking exercise in many contexts — thick documents that nobody reads, produced by consultants hired by the very companies whose projects are being assessed. AIAs risk the same trajectory unless we design enforcement mechanisms from the start."

The Consent Fiction: AIAs often include a requirement for "public consultation" or "stakeholder engagement." But consider the consent fiction operating here: the communities most affected by algorithmic systems are often the least equipped to participate in technical assessment processes. A meaningful AIA requires not just soliciting input but ensuring that affected communities have the knowledge, resources, and power to shape the assessment — and to reject deployment if the risks are unacceptable. Without that, the consultation is another form of consent theater.


17.4 Liability Frameworks for Algorithmic Harm

When an algorithmic system causes harm, the affected person may seek legal redress. But existing liability frameworks were designed for a world in which decisions were made by identifiable humans or caused by identifiable physical products. Algorithmic systems challenge these frameworks in several ways:

  • Causation is complex. The harm results from an interaction between code, data, deployment context, and human decisions. Establishing a clear causal chain from a specific actor's conduct to the harm is difficult.
  • Intent is absent. The algorithm doesn't "intend" to discriminate. It optimizes for an objective function. If the objective function or training data embed bias, the resulting discrimination may be unintentional — but no less harmful.
  • The product is intangible. Software is not a physical product in the traditional sense. Whether algorithms qualify as "products" under product liability law is contested.

17.4.2 Three Liability Approaches

Negligence

Under a negligence standard, the plaintiff must prove four elements:

  1. The defendant owed a duty of care to the plaintiff
  2. The defendant breached that duty
  3. The breach caused the plaintiff's harm
  4. The plaintiff suffered actual damages

Applied to algorithmic systems, the key question is: what constitutes a "reasonable" standard of care for algorithm development and deployment? If a company deploys a hiring algorithm without testing it for racial bias, and the algorithm disproportionately screens out Black candidates, has the company breached a duty of care?

The emerging standard of care: Professional standards for responsible AI development are beginning to crystallize, which may eventually define what "reasonable care" means in the algorithmic context. The NIST AI Risk Management Framework (2023), the IEEE 7000 series of standards, and industry codes of conduct provide reference points. A company that fails to follow widely accepted practices — such as testing for bias before deployment, documenting model limitations, or providing human oversight for high-stakes decisions — may be found to have breached a duty of care even if no specific law mandated those practices.

Advantages: Familiar, well-established legal framework. Incentivizes care and diligence. Adaptable to evolving best practices.

Disadvantages: Places the burden of proof on the person harmed — who typically has far fewer resources than the algorithm developer. Requires establishing what a "reasonable" algorithm developer would do — a standard that is still evolving and contested. Difficult to prove causation in complex algorithmic systems where multiple factors interact to produce outcomes.

Strict Liability

Under strict liability, the defendant is liable for harm regardless of fault. The plaintiff does not need to prove negligence or intent — only that the product caused the harm.

Strict liability is typically applied to abnormally dangerous activities (e.g., blasting, keeping wild animals) and to defective products. The rationale is that certain activities are so inherently risky that the party who profits from them should bear the cost of harm they cause, even if the party exercised all reasonable care.

The case for strict liability for high-stakes algorithms: If an algorithmic system makes decisions about health care, criminal justice, or financial access — decisions with life-altering consequences — there is an argument that the deployer should bear strict liability for harm, on the theory that:

  1. The deployer profits from the system
  2. The deployer is best positioned to prevent harm (by testing, auditing, and monitoring the system)
  3. The affected individual has no ability to inspect or influence the system
  4. The power asymmetry between deployer and affected individual is extreme

"I find the strict liability argument compelling for high-stakes contexts," Eli said. "If VitraMed's model denies someone health care and they're harmed, the patient shouldn't have to prove that VitraMed was negligent. VitraMed chose to deploy the model, VitraMed profits from it, and the patient had no say in the matter. The burden should be on VitraMed to show the system is safe — not on the patient to prove it's dangerous."

The case against strict liability: Critics argue that strict liability would chill innovation, discourage beneficial uses of AI, and create an unmanageable volume of litigation. They note that algorithms are tools, and tools are used by humans who make the final decisions. If the physician at a VitraMed partner clinic overrides the model's recommendation, is VitraMed still liable?

Product Liability

Product liability law holds manufacturers liable for defective products. The question for algorithmic systems is whether software — and specifically, algorithmic decision outputs — qualifies as a "product."

Three types of product defects are relevant:

  1. Design defect: The product is inherently dangerous due to its design. An algorithm trained on biased data that systematically discriminates could be characterized as having a design defect.
  2. Manufacturing defect: The product departs from its intended design. A bug in the code that causes the algorithm to behave differently than designed might qualify.
  3. Failure to warn: The manufacturer fails to provide adequate warnings about the product's risks. A company that sells an algorithmic system without disclosing its known limitations, error rates, or bias characteristics could face failure-to-warn liability.

The EU's 2024 update to the Product Liability Directive explicitly includes software and AI systems within its scope, making it one of the first jurisdictions to clearly apply product liability to algorithmic systems. Under the revised directive, AI system developers and deployers can be held liable for damage caused by defective AI, with a reversal of the burden of proof in certain circumstances — the defendant must prove the system was not defective, rather than the plaintiff proving it was.

Connection to Chapter 16: The transparency demands we examined in Chapter 16 are directly relevant here. Without meaningful transparency about how an algorithmic system works, affected individuals cannot even identify that an algorithm was involved in the decision that harmed them — let alone prove that the algorithm was negligent, defective, or biased. Transparency is a prerequisite for accountability.


17.5 The "Many Hands" Problem

17.5.1 Distributed Responsibility

The philosopher Dennis Thompson coined the term "the problem of many hands" to describe situations in which a harmful outcome results from the actions of many individuals, none of whom is solely responsible. The concept was originally applied to government — where policy failures often emerge from the accumulated decisions of many officials — but it applies with particular force to algorithmic systems.

Consider the full lifecycle of an algorithmic system:

  1. Researchers develop the model architecture
  2. Data engineers collect, clean, and prepare training data
  3. ML engineers train and tune the model
  4. Product managers define the system's objectives and success metrics
  5. Executives approve deployment
  6. Sales teams market the system to customers
  7. Integration engineers adapt the system for specific deployment contexts
  8. Operators use the system in daily practice
  9. Regulators approve (or fail to regulate) the system

If the system causes harm, each actor can point to others. The researchers say they developed a general-purpose technique. The data engineers say they cleaned the data they were given. The ML engineers say they optimized for the metrics they were told to optimize for. The product managers say they relied on the engineers' technical judgment. The executives say they trusted the product team. The operators say they followed the system's recommendations. And so on.

This is not mere evasion. It reflects a genuine structural feature of complex sociotechnical systems: responsibility is distributed across many actors, and no single actor's contribution is sufficient, by itself, to have caused the harm.

Eli offered a visceral example from Detroit's predictive policing system: "The university researchers published a model for predicting crime hotspots. A tech company licensed the model and built it into a commercial product. The city of Detroit purchased the product. The police department configured the deployment parameters. Individual officers acted on the system's recommendations. When Black residents in my neighborhood are disproportionately stopped and searched because the algorithm concentrated police resources there, who's responsible? The researchers say they developed a general predictive technique. The company says they sold a tool. The city says they trusted the vendor. The department says officers exercise discretion. The officers say they went where they were told to go. My neighbors are the ones getting stopped, and nobody is accountable."

The Detroit example is not hypothetical. It mirrors documented patterns in cities including Chicago (where the Strategic Subject List generated heat scores for individuals), Los Angeles (where PredPol directed patrol resources), and New Orleans (where Palantir's predictive policing system operated secretly for years before public disclosure). In each case, the many-hands structure of the system made accountability elusive.

17.5.2 Responses to the Many Hands Problem

Several approaches have been proposed to address distributed responsibility:

Collective responsibility. Hold the organization — not individual actors — responsible for the system's outcomes. This approach treats the algorithmic system as a product of the organization, and assigns liability to the legal entity that deployed it, regardless of how internal responsibilities are distributed.

Role-based responsibility. Assign specific accountability obligations to specific roles in the development and deployment process. The EU AI Act takes this approach, defining distinct obligations for "providers" (developers), "deployers" (operators), and "importers/distributors" — each with defined responsibilities proportional to their control over the system.

Chain of responsibility. Hold each actor in the supply chain jointly and severally liable — meaning the harmed individual can seek full redress from any actor in the chain, and the actors then allocate responsibility among themselves. This approach is used in environmental law and some product liability contexts.

Designated accountability. Require organizations to designate a specific individual who is personally accountable for the system's outcomes — akin to a "responsible AI officer." This approach makes the many-hands problem manageable by creating a single point of accountability, though critics argue it may simply create a scapegoat.

Ray Zhao offered a corporate perspective: "At NovaCorp, we've adopted a version of designated accountability. For every algorithmic system that touches customer-facing decisions, there's a named individual — we call them the 'model owner' — who is personally accountable for the system's performance, fairness, and compliance. They didn't write the code or collect the data, but they own the outcome. That changes the conversation. When there's a name attached to a system, people pay attention."

17.5.3 VitraMed and the Many Hands

Returning to the VitraMed scenario: under a collective responsibility framework, VitraMed as an organization would bear liability for the risk model's outcomes, regardless of which team contributed to the error. Under role-based responsibility, the EU AI Act would classify VitraMed as a "provider" of a high-risk AI system (health care applications are explicitly listed as high-risk in Annex III of the Act) and require it to meet specific obligations — including risk management, data governance, technical documentation, human oversight, and post-market monitoring.

Mira found the role-based approach clarifying. "It doesn't solve everything," she said. "But at least it gives me a framework. If I can identify VitraMed's role in the chain — provider, deployer, or something else — I can identify our specific obligations. That's better than the current situation, where we could argue forever about who's responsible and meanwhile the patient still has no recourse."

Applied Framework — Accountability Mapping: When analyzing an algorithmic system's accountability structure, map each actor in the system's lifecycle, their specific role, their capacity to prevent harm, their awareness of risk, and their benefit from the system's operation. Actors with greater capacity to prevent harm, greater awareness of risk, and greater benefit should bear proportionally greater accountability. This is the proportional accountability principle.


17.6 Emerging Institutions of Algorithmic Accountability

17.6.1 The Algorithmic Audit Industry

A new industry is emerging to meet the growing demand for algorithmic accountability. Algorithmic audit firms — sometimes called "AI audit firms" or "responsible AI consultancies" — offer services ranging from bias testing and fairness assessments to comprehensive algorithmic impact reviews.

Notable early entrants include:

  • ORCAA (O'Neil Risk Consulting and Algorithmic Auditing), founded by Cathy O'Neil, author of Weapons of Math Destruction. ORCAA conducts independent algorithmic audits for companies and government agencies.
  • Holistic AI, a UK-based firm offering AI governance, risk, and compliance services.
  • Arthur AI, which provides model monitoring and bias detection tools.
  • The AI Forensics team at AlgorithmWatch, a European civil society organization that investigates algorithmic systems through journalistic and research methods.

The emergence of this industry is promising — it creates an institutional infrastructure for accountability that didn't exist a decade ago. But it also raises questions:

  • Who hires the auditor? If the company being audited selects and pays the auditor, the independence of the audit is compromised — the same principal-agent problem that plagued financial auditing before the Sarbanes-Oxley reforms.
  • What standards apply? There is no generally accepted set of "algorithmic auditing standards" analogous to Generally Accepted Auditing Standards (GAAS) in financial auditing. Different firms use different methodologies, making comparisons difficult.
  • Who sees the results? If audit results are confidential — shared only with the client — the public accountability function of auditing is lost. Public reporting of audit results is essential for genuine accountability but creates legal risk for the audited organization.

17.6.2 Standards Development

The absence of generally accepted algorithmic auditing standards is a critical gap. Financial auditing benefits from decades of standards development — GAAS, GAAP, IFRS — that define what auditors must examine, what methods they must use, and what their reports must include. Algorithmic auditing has no equivalent.

Several organizations are working to fill this gap:

  • The National Institute of Standards and Technology (NIST) published the AI Risk Management Framework (AI RMF) in 2023, providing a voluntary framework for managing AI risks that can inform audit practices.
  • The International Organization for Standardization (ISO) has published ISO/IEC 42001 (AI Management Systems) and is developing additional standards for AI bias assessment and risk management.
  • The IEEE has developed the 7000 series of standards addressing ethical considerations in system design, including standards for algorithmic bias considerations and transparency.
  • The European Commission has mandated the development of harmonized standards under the EU AI Act, which will eventually provide legally binding technical specifications for high-risk AI systems.

These standards are in early stages. It will take years before the algorithmic auditing field achieves the level of standardization that financial auditing enjoys. In the meantime, auditors operate with varying methodologies, making it difficult to compare results across audits or to establish consistent accountability expectations.

17.6.3 Regulatory Sandboxes

Regulatory sandboxes are controlled environments in which new technologies can be tested under regulatory supervision, with temporary relaxations of certain rules. The concept originated in financial regulation — the UK Financial Conduct Authority launched the first fintech sandbox in 2015 — and has been extended to AI.

The EU AI Act provides for AI regulatory sandboxes, where developers can test high-risk AI systems under the supervision of national competent authorities. The sandboxes allow regulators to observe how AI systems behave in practice, provide guidance to developers, and develop regulatory expertise — while giving developers a pathway to compliance without the risk of immediate enforcement action.

Spain became the first EU member state to establish an AI regulatory sandbox, in 2022, followed by others.

Advantages: Sandboxes build regulatory capacity. They allow learning-by-doing. They provide a structured pathway for innovation within regulatory bounds.

Limitations: Sandbox conditions may not reflect real-world deployment. Companies may use sandbox participation as a marketing signal ("regulatory-approved AI") without meaningfully addressing risks. And the temporary nature of sandboxes means that supervision ends — potentially before the system's long-term impacts are understood.

17.6.4 Legislative Responses

Legislatures around the world are responding to the accountability gap with new laws and proposals.

The Algorithmic Accountability Act (U.S.): First introduced in the U.S. Congress in 2019 and reintroduced in subsequent sessions, this bill would require large companies (those with annual revenues exceeding $50 million or data on more than one million individuals) to conduct impact assessments of automated decision systems. The assessments would evaluate the system's accuracy, fairness, bias, privacy implications, and security risks. The Federal Trade Commission (FTC) would oversee compliance.

As of the time of writing, the Algorithmic Accountability Act has not been enacted. But it represents an important template for algorithmic governance legislation, and its core ideas — mandatory impact assessments, regulatory oversight, public reporting — have influenced state-level legislation and international developments.

New York City Local Law 144 (2021): This law, which took effect in 2023, requires employers using automated employment decision tools (AEDTs) to conduct annual bias audits and to notify candidates when AEDTs are used. It is one of the first laws in the world to mandate algorithmic audits for a specific application.

The law has been criticized — its definition of AEDTs is narrow, its enforcement mechanisms are limited, and some argue it doesn't go far enough — but it represents a practical step toward algorithmic accountability at the local level.

Sofia Reyes tracked the implementation closely: "Law 144 matters not because it's perfect — it's not — but because it establishes the principle that companies that use algorithms to make consequential decisions about people's lives must subject those algorithms to independent scrutiny. That principle, once established, can be expanded. The first environmental impact assessment law wasn't perfect either. But it created a framework that improved over decades."

The Accountability Gap — A Structural View: The accountability gap is not simply a matter of missing laws or insufficient auditing. It is a structural feature of a system in which the entities that profit from algorithmic decision-making are shielded from its consequences by layers of technical complexity, legal uncertainty, and distributed responsibility. Closing the gap requires action on multiple fronts simultaneously — legal frameworks that assign clear liability, institutional capacity for independent auditing, technical tools for bias detection and monitoring, and civic organizations that advocate for affected communities. No single mechanism is sufficient. The architecture of accountability must be as complex as the architecture of the systems it governs.


17.7 Case Studies

17.7.1 The Algorithmic Accountability Act: Legislative Responses to Algorithmic Harm

The Algorithmic Accountability Act represents one of the most comprehensive legislative attempts to address the accountability gap in the United States. Understanding its history, provisions, and political trajectory illuminates both the promise and the difficulty of legislative responses to algorithmic governance challenges.

Background: By the mid-2010s, a growing body of research had documented algorithmic harms — ProPublica's 2016 investigation of the COMPAS recidivism algorithm, studies of discrimination in online advertising, evidence of bias in facial recognition systems. Civil society organizations, including the ACLU, the Electronic Frontier Foundation, and the AI Now Institute, called for regulatory intervention.

The 2019 Bill: Introduced by Senator Ron Wyden (D-OR) and Representative Yvette Clarke (D-NY), the original Algorithmic Accountability Act would have required the FTC to promulgate regulations requiring "covered entities" to conduct automated decision system impact assessments. These assessments would evaluate:

  • The system's purpose, design, and methodology
  • Data inputs and their sources
  • Potential for inaccurate, unfair, biased, or discriminatory outcomes
  • The extent to which the system affects consumer privacy
  • Security risks

The bill required assessments to be performed by an internal team or external auditor, with results reported to the FTC. Crucially, it also required consultation with relevant external stakeholders, including communities affected by the system.

Subsequent Versions: The bill was reintroduced in 2022 with expanded provisions, including requirements for impact assessments to be conducted before deployment (not just after), enhanced transparency provisions, and stronger FTC enforcement authority. The 2022 version also expanded the definition of "automated decision system" to include systems that inform human decisions, not just those that make decisions autonomously.

Political Dynamics: The bill faced opposition from technology industry groups who argued it would impose excessive compliance costs, chill innovation, and create legal uncertainty. Supporters countered that the costs of not regulating — continued algorithmic discrimination, erosion of trust, harm to vulnerable communities — were far greater.

Lessons: The Algorithmic Accountability Act illustrates several governance dynamics:

  1. The lag between harm and regulation. Years elapsed between documented algorithmic harms and legislative response. During that lag, algorithmic systems continued to expand into more consequential domains.
  2. The expertise gap. Legislators often lacked the technical knowledge to evaluate the bill's provisions. This created vulnerability to industry lobbying that emphasized technical complexity and cost.
  3. The template effect. Even without passage, the bill influenced state-level legislation, corporate self-regulation, and international governance developments. Legislative proposals can shape norms even when they don't become law.

17.7.2 Auditing Airbnb: Racial Discrimination in Platform Marketplaces

In 2016, Benjamin Edelman, Michael Luca, and Dan Svirsky published a landmark audit study: "Racial Discrimination in the Sharing Economy: Evidence from a Field Experiment" (American Economic Journal: Applied Economics). The study demonstrated that Airbnb guests with distinctly African American names were approximately 16% less likely to be accepted by hosts compared to identical guests with distinctly white names.

Methodology: The researchers created 20 fictitious Airbnb guest profiles that were identical in all respects except the name, which signaled race. They sent approximately 6,400 messages to hosts across five cities, requesting to book. The profiles included positive review histories and other markers of trustworthiness.

Findings: - Guests with African American names were accepted approximately 42% of the time, compared to approximately 50% for guests with white names - The discrimination was present among hosts of all races, both genders, and across price points - Hosts with multiple properties (professional hosts) discriminated at similar rates to hosts with single properties - The discrimination occurred despite Airbnb's non-discrimination policy

Airbnb's Response: Airbnb initially resisted the findings but eventually acknowledged the problem. The company commissioned a report from former Attorney General Eric Holder and implemented several changes:

  1. Required all users to accept a "Community Commitment" affirming non-discrimination
  2. Expanded "Instant Book" (which bypasses host approval) to reduce opportunities for discrimination
  3. Reduced the prominence of guest photos during the booking process
  4. Created a dedicated anti-discrimination team
  5. Committed to periodic testing for discrimination

The Limits of Platform Self-Regulation: Despite these measures, subsequent research found persistent discrimination on the platform. A 2020 study found that Airbnb listings in predominantly Black neighborhoods were priced lower and reviewed more negatively than comparable listings in white neighborhoods, suggesting that the platform's algorithmic recommendation and pricing systems may embed structural racial biases that individual-level interventions cannot address.

Accountability Analysis: This case illustrates several themes from the chapter:

  • The audit study method was essential for documenting discrimination that Airbnb's internal processes had not detected (or had not disclosed)
  • The accountability gap operated in multiple dimensions: hosts could discriminate without consequence, the platform's design facilitated discrimination, and guests who were rejected often didn't know the reason
  • The many hands problem was present: the discrimination resulted from host behavior, platform design, algorithmic recommendations, and corporate policy (or lack thereof), making it difficult to assign responsibility to any single actor
  • External pressure — from researchers, media, and advocacy organizations — was necessary to compel action. Internal accountability mechanisms alone were insufficient

Eli connected the case to Detroit. "This is the same pattern. The platform says 'we don't discriminate, our users do.' The tech company selling predictive policing software says 'we don't target Black neighborhoods, the data does.' Everyone's hands are clean because the system does the dirty work. That's what the accountability gap looks like on the ground."


17.8 Chapter Summary

Key Concepts

Concept Definition
Accountability gap The structural condition in which algorithmic systems make consequential decisions while no identifiable actor bears clear responsibility for outcomes
Algorithmic audit A systematic examination of an algorithmic system's behavior, impacts, and compliance with relevant standards
Audit study A field experiment that tests a system by submitting controlled inputs and observing differential outcomes
Disparate impact testing Statistical analysis of whether a system produces significantly different outcomes across demographic groups
User experience audit Evaluation of an algorithmic system from the perspective of affected individuals
Algorithmic Impact Assessment A structured process for evaluating an algorithmic system's potential harms before and during deployment
Many hands problem The diffusion of responsibility across multiple actors in complex systems, making individual accountability difficult
Strict liability Liability regardless of fault — the deployer is responsible for harm even if they exercised reasonable care
Negligence Liability based on failure to meet a reasonable standard of care
Product liability Liability of manufacturers for defective products, potentially applicable to algorithmic systems
Regulatory sandbox A supervised environment for testing innovative technologies under regulatory oversight
Proportional accountability The principle that accountability should be proportional to an actor's capacity to prevent harm, awareness of risk, and benefit from the system

Key Debates

  • Should algorithmic systems be subject to strict liability, negligence standards, or a novel liability framework?
  • Is self-regulation (internal audits, voluntary impact assessments) sufficient, or must algorithmic accountability be mandated by law?
  • How can algorithmic auditing be made independent and credible when companies control access to their systems?
  • Does the many hands problem make individual accountability impossible, or can institutional design create meaningful responsibility?
  • Can Algorithmic Impact Assessments avoid becoming compliance theater — boxes checked without genuine accountability?

Applied Framework

The Accountability Chain Analysis: 1. Map the actors: Identify every entity involved in the system's lifecycle (developer, deployer, data provider, operator, regulator) 2. Map the roles: Define each actor's specific function and contribution 3. Assess capacity: Which actors have the capacity to prevent or mitigate harm? 4. Assess awareness: Which actors knew or should have known about risks? 5. Assess benefit: Which actors profit from the system's operation? 6. Assign proportional accountability: Actors with greater capacity, awareness, and benefit bear greater responsibility 7. Identify gaps: Where does accountability fall through the cracks? Which harms have no accountable actor? These gaps are the priority for governance intervention.


What's Next

This chapter has examined who is responsible when algorithmic systems cause harm. Chapter 18: Generative AI: Ethics of Creation and Deception confronts a new frontier in which AI systems do not merely analyze data or automate decisions — they create. Large language models generate text, diffusion models generate images, and deepfake systems generate video that is increasingly indistinguishable from reality. These generative systems raise a distinctive set of ethical questions: Who owns AI-generated content? What happens when models hallucinate — generating plausible but false information? How do we govern synthetic media in a democracy that depends on shared truth? And what happens to the human workers whose creative labor trained these systems?

The accountability frameworks from this chapter will prove essential, because generative AI raises the many-hands problem in its most acute form: when an AI system generates a deepfake that influences an election, who is responsible?


Chapter 17 Exercises → exercises.md

Chapter 17 Quiz → quiz.md

Case Study: The Algorithmic Accountability Act: Legislative Responses → case-study-01.md

Case Study: Auditing Airbnb: Racial Discrimination in Platform Marketplaces → case-study-02.md