40 min read

> "The real problem is not whether machines think but whether men do."

Learning Objectives

  • Distinguish between levels of autonomy and explain how increasing autonomy shifts ethical and governance responsibilities
  • Critically evaluate the trolley problem as a framework for autonomous vehicle ethics, identifying both its insights and its limitations
  • Define lethal autonomous weapons systems (LAWS) and assess the ethical arguments for and against their development and deployment
  • Analyze the MIT Moral Machine experiment and its implications for cross-cultural moral reasoning about autonomous systems
  • Evaluate philosophical arguments about whether machines can be moral agents
  • Compare human-in-the-loop, human-on-the-loop, and human-out-of-the-loop oversight models and identify appropriate applications for each
  • Explain the EU AI Act's classification of high-risk AI and its human oversight requirements

Chapter 19: Autonomous Systems and Moral Machines

"The real problem is not whether machines think but whether men do." — B.F. Skinner, Contingencies of Reinforcement (1969)

Chapter Overview

In Chapter 17, we examined who is responsible when algorithmic systems cause harm. In Chapter 18, we explored what happens when AI systems create content. This chapter confronts what happens when AI systems act — making decisions and executing actions in the physical world, sometimes with irreversible consequences.

A self-driving car encounters an unavoidable collision and must choose a course of action in milliseconds. A military drone identifies a target and must decide whether to fire without waiting for human confirmation. A diagnostic AI flags a patient as critically ill and recommends an immediate treatment protocol. In each case, the system acts with a degree of independence from human control — and in each case, the ethical, legal, and governance questions are acute.

These are not hypothetical scenarios. Autonomous vehicles are on public roads. Military drones with increasing levels of autonomy are deployed in combat zones. Diagnostic AI systems inform clinical decisions in hospitals. The question is not whether autonomous systems will be part of our lives, but how we will govern them — and who will be held accountable when they cause harm.

This chapter examines the spectrum of autonomy, the philosophical question of machine moral agency, the real-world governance challenge of maintaining meaningful human oversight, and the emerging regulatory frameworks — particularly the EU AI Act — that attempt to balance innovation with protection. It is the final chapter of Part 3, and it brings together the themes of bias (Chapter 14), fairness (Chapter 15), transparency (Chapter 16), accountability (Chapter 17), and generative AI (Chapter 18) into a comprehensive view of what responsible AI governance requires.

In this chapter, you will learn to: - Map autonomous systems along a spectrum of human control and evaluate the governance implications of each level - Move beyond the trolley problem to analyze the real engineering and ethical constraints of autonomous vehicles - Assess the legal, ethical, and strategic arguments surrounding lethal autonomous weapons - Evaluate whether machines can be moral agents and what follows from the answer - Design appropriate human oversight mechanisms for different categories of autonomous systems - Apply the EU AI Act's risk-based framework to autonomous systems governance


19.1 Levels of Autonomy

19.1.1 The Autonomy Spectrum

Autonomy is not binary. Systems exist on a spectrum from fully human-controlled to fully machine-controlled, with multiple intermediate levels where humans and machines share decision-making authority.

The most widely used framework for describing this spectrum comes from SAE International (originally the Society of Automotive Engineers), which defines six levels of driving automation:

Level Name Description Human Role
0 No Automation Human performs all driving tasks Full control
1 Driver Assistance System assists with steering OR acceleration/deceleration Monitors and controls
2 Partial Automation System controls both steering and acceleration/deceleration Must monitor constantly, ready to intervene
3 Conditional Automation System handles all driving tasks in certain conditions Must be ready to take over when requested
4 High Automation System handles all driving tasks in certain conditions, even if human doesn't respond May take over optionally
5 Full Automation System handles all driving tasks in all conditions No driving role

While developed for vehicles, this framework generalizes to other domains. Medical AI can range from Level 0 (human reads all scans) to Level 5 (AI diagnoses and initiates treatment without human involvement). Military systems range from Level 0 (human pulls every trigger) to Level 5 (system identifies, tracks, selects, and engages targets without human input).

19.1.2 The Governance Implications of Each Level

The level of autonomy has direct implications for governance:

At lower levels (0-2), the human is clearly in control. When something goes wrong, traditional accountability frameworks apply — the human decision-maker is responsible. The system is a tool.

At Level 3, a critical transition occurs. The system is in control most of the time, but the human must be ready to intervene. This creates what human factors researchers call the "vigilance problem": humans are poor at monitoring systems that work well most of the time. Attention degrades. When the system finally encounters a situation it cannot handle and requests human intervention, the human may not be prepared. The result is a dangerous gap between formal responsibility (the human is supposed to be monitoring) and practical reality (the human is not meaningfully engaged).

At Levels 4-5, the system operates independently. Traditional accountability frameworks — which assume a human decision-maker — no longer apply cleanly. If the system causes harm, the "responsible human" may not exist in any meaningful sense. The driver of a Level 5 vehicle is a passenger. The physician relying on a Level 5 diagnostic AI may never see the underlying data. The soldier overseeing a Level 5 weapons system may not know a target has been engaged until after the fact.

Dr. Adeyemi framed the governance challenge: "Every increase in autonomy is also an increase in the governance demand. Not because autonomous systems are inherently dangerous — many are safer than human alternatives — but because the governance infrastructure we've built over centuries assumes that consequential decisions are made by humans who can explain their reasoning, who can be held accountable, and who can exercise moral judgment. When we remove the human from the loop, we remove the foundation on which our accountability structures rest. Something must replace it."


19.2 Self-Driving Cars: Beyond the Trolley Problem

19.2.1 The Trolley Problem and Its Limits

No discussion of autonomous vehicles is complete without addressing the trolley problem — the thought experiment in which a runaway trolley is headed toward five people, and you can divert it to a side track where it will kill one person instead. Should you act?

Applied to autonomous vehicles: if a self-driving car faces an unavoidable accident, should it be programmed to minimize total deaths (potentially sacrificing its passenger to save multiple pedestrians)? To prioritize its passenger (the person who bought the car)? To choose randomly?

The trolley problem has dominated public discourse about autonomous vehicle ethics. It is the first thing most people think of when they hear "self-driving car ethics." And while it raises genuine philosophical questions, it is a misleading guide to the actual ethics of autonomous vehicles for several reasons:

1. The scenario is extremely rare. Real autonomous vehicle incidents are overwhelmingly caused by failures of perception (the car doesn't see the pedestrian), prediction (the car misjudges the pedestrian's trajectory), or system failure (a sensor malfunction) — not by forced binary choices between equally visible, equally inevitable outcomes. The trolley problem imagines a scenario of perfect information and no time for anything but a binary choice. Real autonomous driving involves imperfect information, uncertain predictions, and continuous decision-making.

2. The framing obscures the real ethical questions. The trolley problem asks "who should the car kill?" when the more important questions are: - How safe must autonomous vehicles be before they are allowed on public roads? - Who decides the acceptable error rate, and how is that decision made? - How are the risks and benefits of autonomous driving distributed across communities? - Who is liable when an autonomous vehicle kills someone? - How are affected communities consulted in deployment decisions?

3. It individualizes a structural problem. The trolley problem focuses on the moment of crisis — a single car facing a single choice. But the ethical questions about autonomous vehicles are primarily systemic: about infrastructure, regulation, insurance, liability, labor displacement (for professional drivers), and the distribution of risk across different populations. Focusing on trolley-problem scenarios diverts attention from these structural issues.

"I'm tired of the trolley problem," Eli said. "Here's what I want to know: when autonomous vehicles are deployed in my neighborhood, will they be tested as rigorously in Black neighborhoods as in white ones? Will the training data include enough examples of Black pedestrians in different lighting conditions? Who decided that my neighborhood is the testing ground? Those are the real ethical questions."

19.2.2 The Real Engineering Constraints

The actual ethical challenges of autonomous vehicles are engineering and governance challenges, not philosophical thought experiments:

Perception reliability: Autonomous vehicles must perceive their environment accurately in all conditions — rain, fog, darkness, glare, construction zones, unusual road configurations. Studies have documented that some computer vision systems perform less accurately on darker skin tones, raising concerns about whether autonomous vehicles detect all pedestrians equally.

Edge cases: Autonomous vehicles encounter situations that were not represented in their training data — a child running into the street, a wheelchair user on the road, a vehicle traveling the wrong way, a road obstruction that doesn't match any trained category. How the system handles novel situations determines its real-world safety.

Operational design domain (ODD): Every autonomous vehicle system is designed to operate within a specific set of conditions — certain road types, weather conditions, speed ranges, geographic areas. When the vehicle encounters conditions outside its ODD, it must respond appropriately. The transition from autonomous to human control (or to a safe stop) is one of the most challenging engineering problems in the field.

Validation and testing: How do you demonstrate that an autonomous vehicle is "safe enough"? The system's error rate may be lower than a human driver's in aggregate — but the types of errors may differ. Humans rarely drive onto a sidewalk because they mistook a pedestrian for a shadow. Autonomous vehicles can make precisely these kinds of perceptual errors. Whether a lower overall error rate with qualitatively different error types constitutes "safer" is not a purely technical question.

19.2.3 The Uber Self-Driving Car Fatality

On March 18, 2018, an Uber self-driving test vehicle struck and killed Elaine Herzberg, a 49-year-old woman who was walking her bicycle across a road in Tempe, Arizona. It was the first known fatality involving a fully autonomous vehicle striking a pedestrian.

The National Transportation Safety Board (NTSB) investigation revealed a cascade of failures:

  • The vehicle's perception system detected Herzberg approximately 5.6 seconds before impact but classified her inconsistently — first as a vehicle, then as "other," then as a bicycle — and repeatedly failed to predict her path
  • Uber had disabled the vehicle's automatic emergency braking system, relying instead on the human safety driver to intervene
  • The safety driver, who was supposed to monitor the road, was looking at her phone at the time of the collision
  • Uber's safety culture prioritized ride experience (minimizing hard braking) over pedestrian safety

The case illustrates virtually every theme in this chapter:

The accountability gap: Uber blamed the safety driver. The safety driver's attorneys argued that Uber's system design — particularly the decision to disable automatic braking — made the accident inevitable. The State of Arizona initially declined to bring criminal charges against Uber, though the safety driver was eventually charged with negligent homicide.

The many hands problem: The failure involved Uber's engineering decisions, the safety driver's inattention, Arizona's permissive regulatory framework, and NHTSA's lack of specific safety standards for autonomous vehicles. No single actor's failure was sufficient to cause the fatality; each contributed.

The consent fiction: Elaine Herzberg did not consent to being a test subject in Uber's autonomous vehicle experiment. She was simply crossing a road. The public roads on which autonomous vehicles are tested become de facto testing grounds for everyone who uses them — with no mechanism for informed consent.

The Power Asymmetry: Uber, a company with billions of dollars in resources, tested its experimental technology on public roads. Elaine Herzberg, a pedestrian with no knowledge of or consent to the experiment, paid the price. The asymmetry is structural: the entities that deploy autonomous systems have the resources, expertise, and political influence to shape the regulatory environment. The people whose safety is at stake — pedestrians, cyclists, other drivers — have no comparable voice.


19.3 Autonomous Weapons Systems (LAWS)

19.3.1 Definitions and Current State

Lethal autonomous weapons systems (LAWS) are weapon systems that can select and engage targets without direct human intervention. The term encompasses a range of capabilities, from systems that autonomously track and target incoming missiles (which have existed for decades) to hypothetical systems that independently identify, track, and kill human beings.

The distinction that matters most for governance is between:

  • Autonomous targeting functions (the system selects which targets to engage) — this is the ethically and legally contested capability
  • Autonomous navigation/positioning (the system controls its own movement) — this is less contested
  • Autonomous execution (the system fires or detonates without a human command) — this is the most contested capability

As of the mid-2020s, no country has publicly acknowledged deploying fully autonomous weapons that independently select and engage human targets without human authorization. However, several systems approach this threshold:

  • Israel's Harpy and Harop loitering munitions can autonomously detect and destroy radar emitters
  • Turkey's Kargu-2 drone is alleged to have autonomously engaged targets in Libya in 2020, though the details are disputed
  • South Korea's SGR-A1 sentry gun, deployed along the DMZ, can detect and track intruders and (in theory) fire autonomously, though it is currently operated in a human-supervised mode
  • Various nations are developing autonomous drone swarms that can operate collectively without individual human control of each unit

19.3.2 The Case Against LAWS

The Campaign to Stop Killer Robots, a coalition of NGOs including Human Rights Watch, the International Committee for Robot Control, and numerous national organizations, has advocated for a preemptive ban on fully autonomous weapons since 2013. Their arguments include:

The accountability gap. If an autonomous weapon kills a civilian, who is responsible? The commander who authorized its deployment? The programmer who designed its targeting algorithm? The manufacturer? As Chapter 17 established, the many-hands problem makes attributing responsibility for algorithmic decisions difficult. In the context of weapons that kill, this accountability gap is morally unacceptable. The laws of armed conflict require that someone be held accountable for every use of lethal force.

The inability to exercise judgment. International humanitarian law requires that combatants exercise distinction (distinguishing between combatants and civilians), proportionality (ensuring that civilian harm is not excessive relative to military advantage), and precaution (taking feasible measures to minimize civilian harm). These principles require the kind of contextual moral judgment that current AI systems cannot perform. An autonomous weapon cannot assess whether a person carrying a farming tool is a combatant or a civilian. It cannot weigh the military value of a target against the likely civilian casualties. It cannot exercise mercy.

The lowered threshold for conflict. If war can be fought by machines — reducing the risk to one's own soldiers — the political cost of going to war decreases. This may make armed conflict more likely, particularly against adversaries who lack equivalent autonomous capabilities. The asymmetry is troubling: wealthy nations deploy robot soldiers while poorer nations bear the human cost.

Human dignity. There is a fundamental argument, independent of consequences, that the decision to take a human life should be made by a human being who can comprehend the moral weight of that decision. Delegating the kill decision to a machine — even a highly capable one — is a violation of the dignity of the person killed, who is denied the minimal recognition of being judged by a moral agent.

19.3.3 The Case For (or Against Banning) LAWS

Proponents of autonomous weapons — or at least opponents of a preemptive ban — advance several arguments:

Precision and reduced civilian casualties. Autonomous systems may be more precise than human soldiers, who are subject to fear, fatigue, anger, and revenge. If autonomous weapons can distinguish combatants from civilians more accurately than humans, a ban on autonomous weapons could increase civilian casualties.

Force protection. Autonomous weapons reduce the risk to one's own soldiers. A nation has a legitimate interest in protecting its military personnel, and autonomous systems serve that interest.

Deterrence. Autonomous weapons may deter aggression by demonstrating overwhelming military capability without requiring large deployments of human soldiers.

Feasibility of a ban. Critics argue that a preemptive ban is unenforceable and that nations will develop autonomous weapons regardless. Better, they argue, to develop norms and regulations that ensure responsible use than to attempt a prohibition that will be circumvented.

Eli was unconvinced by these arguments: "They said the same thing about landmines and cluster munitions — that bans wouldn't work, that the weapons were militarily necessary. The Mine Ban Treaty isn't perfect, but it's saved lives. And the argument that autonomous weapons might be more precise than humans is speculative. We don't have evidence for that claim. What we do have evidence for is that AI systems are biased, that they fail in unpredictable ways, and that the people who deploy them are rarely held accountable when things go wrong. Now we're talking about giving those same systems the power to kill."

19.3.4 The Governance Landscape

Discussions on LAWS have been ongoing at the UN Convention on Certain Conventional Weapons (CCW) since 2014. A Group of Governmental Experts (GGE) has met regularly to discuss potential regulations. However, progress has been slow, with major military powers — including the United States, Russia, China, and Israel — resisting binding restrictions.

Key principles that have emerged from these discussions include:

  • Meaningful human control: The principle that humans must retain meaningful control over the use of lethal force. What constitutes "meaningful" control is contested.
  • Compliance with IHL: Any autonomous weapon must comply with the principles of distinction, proportionality, and precaution under international humanitarian law.
  • Accountability: There must be a clear chain of accountability for every use of lethal force by an autonomous system.

The International Committee of the Red Cross (ICRC) has recommended that states adopt new rules on autonomous weapons, including prohibiting unpredictable autonomous weapons and those designed to use force against people, and requiring sufficient human control over all others.

Recurring Theme — Power Asymmetry: The autonomous weapons debate starkly illustrates the power asymmetry theme. The nations developing LAWS are the wealthiest and most technologically advanced. The communities most likely to be targeted by LAWS are in the Global South. The governance discussions occur at the UN, where the voices of affected communities are mediated through state representatives who may not share their interests. The people whose lives are at stake have the least influence over the decisions that govern the technology.


19.4 The Moral Machine and Cross-Cultural Ethics

19.4.1 The MIT Moral Machine Experiment

In 2018, researchers at MIT published the results of the Moral Machine experiment — an online platform that presented users with variations of the trolley problem applied to autonomous vehicles and asked them to choose which outcome they preferred. The experiment collected over 40 million decisions from users in 233 countries and territories, making it one of the largest studies of moral preferences ever conducted.

Participants were presented with scenarios in which a self-driving car's brakes have failed and it must choose between two harmful outcomes — for example, continuing straight and hitting three elderly pedestrians, or swerving and hitting two young children. The scenarios varied across multiple dimensions: the number of people affected, their ages, their genders, their social status (professional vs. homeless), whether they were pedestrians or passengers, and whether they were obeying traffic laws.

19.4.2 Key Findings

Three universal preferences emerged across all cultures: 1. Sparing human lives over animal lives 2. Sparing more lives over fewer lives 3. Sparing younger lives over older lives

But significant cultural variation appeared in other dimensions:

The researchers identified three "moral clusters" of countries:

  • Western cluster (North America, Europe, Australasia): Stronger preference for sparing the young, weaker preference for sparing higher-status individuals
  • Eastern cluster (East Asia, particularly Japan and China): Weaker preference for sparing the young relative to the old, reflecting cultural respect for elders
  • Southern cluster (Latin America, parts of Africa and the Middle East): Stronger preferences for sparing women and for sparing higher-status individuals

19.4.3 What the Experiment Tells Us (and What It Doesn't)

What it tells us: Moral preferences about life-and-death trade-offs vary significantly across cultures. Any attempt to program universal ethical preferences into autonomous systems must contend with the fact that there is no universal agreement on these preferences.

What it doesn't tell us:

1. Survey preferences ≠ real-world decisions. What people say they prefer in an online survey may differ from what they would actually choose in a real emergency, or what they would accept as the decision their autonomous car makes.

2. The framing constrains the answer. The Moral Machine presents forced binary choices. Real autonomous driving involves continuous decision-making under uncertainty, where the options are not neatly packaged as "swerve left or right."

3. Preferences are not principles. The finding that people in some cultures prefer to spare the young over the old does not mean that autonomous vehicles should be programmed with this preference. An aggregate preference may conflict with principles of equal human worth and non-discrimination.

4. The trolley framing distracts from systemic issues. As discussed in Section 19.2, the most important ethical questions about autonomous vehicles are not about impossible binary choices but about testing rigor, equitable deployment, accountability structures, and community consent.

Dr. Adeyemi used the Moral Machine as a teaching tool with a twist: "I don't show you this study to help you program a car. I show you this study to illustrate the impossibility of the exercise. There is no culturally neutral set of moral preferences to encode. The question is not 'what should the car choose?' The question is 'who has the authority to make that choice, and through what process?' That is a governance question, not a programming question."


19.5 Can Machines Be Moral Agents?

19.5.1 The Philosophical Question

As autonomous systems take on more consequential decisions, a fundamental philosophical question emerges: Can a machine be a moral agent?

A moral agent is an entity that can: 1. Recognize moral dimensions of a situation 2. Deliberate about right and wrong 3. Choose an action based on moral reasoning 4. Be held morally responsible for that choice

A moral patient is an entity that can be wronged — that has interests, experiences, or a welfare that can be harmed. Humans are clearly both moral agents and moral patients. Animals are generally considered moral patients (they can be wronged) but not full moral agents (they cannot be held morally responsible).

Where do autonomous AI systems fall?

19.5.2 Three Positions

Position 1: Machines are not and cannot be moral agents.

This position, held by most contemporary philosophers and ethicists, argues that moral agency requires properties that current AI systems lack:

  • Consciousness and subjective experience: Moral agency requires the ability to understand the moral significance of one's actions, not merely to process rules about morality. An autonomous vehicle that swerves to avoid a pedestrian does not value human life — it follows a programmed objective function.
  • Free will and autonomy: Moral responsibility implies that the agent could have chosen otherwise. AI systems do not choose — they compute. Their outputs are determined by their inputs, training data, and architecture. The appearance of choice is an anthropomorphic projection.
  • Understanding, not just processing: There is a fundamental difference between a system that can process moral rules (if the trolley is heading toward five people, divert it toward one) and a system that understands why human life has value. Current AI systems process. They do not understand.

Position 2: Machines can be functional moral agents.

Some researchers, including Luciano Floridi and J.W. Sanders, argue for a concept of functional morality — the idea that a system can be a moral agent in a functional sense even if it lacks consciousness or subjective experience. On this view, if a system can:

  • Perceive morally relevant features of a situation
  • Process those features through a decision framework
  • Select an action that is morally appropriate
  • Adjust its behavior based on outcomes

...then it functions as a moral agent, regardless of whether it has subjective experience. A medical AI that detects a life-threatening condition and alerts a physician has functioned as a morally responsible agent, even if it doesn't "care" about the patient in any experiential sense.

Position 3: The question is premature but important.

This pragmatic position argues that current AI systems are clearly not moral agents in any robust sense, but that the trajectory of AI development may eventually produce systems whose moral status is genuinely ambiguous. We should develop the philosophical frameworks now — before the ambiguity arrives — rather than scrambling to address it after the fact.

19.5.3 The Responsibility Gap

Regardless of where one stands on machine moral agency, a practical problem emerges: the responsibility gap.

If a machine is not a moral agent, it cannot be held morally responsible for its actions. But if the machine acts autonomously — without a human in the loop — there may be no identifiable human who can be held responsible either. The developer designed the system but did not make the specific decision. The user activated the system but did not control the specific action. The owner deployed the system but did not foresee the specific circumstance.

This creates a gap in the moral and legal framework: a harmful action has occurred, but no moral agent — human or machine — is clearly responsible.

Andreas Matthias, who coined the term "responsibility gap," argues that this gap is a new problem created by autonomous systems. It is not merely the many-hands problem from Chapter 17 (where responsibility is distributed among many humans), but a deeper challenge: responsibility may not be assignable to any human, because the machine's behavior emerged from a learning process that no human fully controlled or predicted.

"This is what keeps me up at night," Mira admitted. "If VitraMed deploys an autonomous diagnostic assistant — one that learns and adapts over time — and it eventually makes a recommendation that no one on our team programmed or predicted, are we still responsible? We designed the system, yes. But if it evolved beyond our design through its own learning process, in what sense did we make the decision?"


19.6 Human Oversight: Models and Limits

19.6.1 Three Models of Human Oversight

The consensus response to the responsibility gap is that autonomous systems must maintain meaningful human control. But what does this mean in practice? Three models are commonly distinguished:

Human-in-the-loop (HITL): A human must approve every significant action before the system executes it. The system recommends; the human decides.

  • Example: A diagnostic AI identifies a potential cancer on a medical scan and flags it for a radiologist, who reviews the scan and makes the diagnosis.
  • Strengths: The human retains full decision authority. Accountability is clear.
  • Limitations: Slow. Doesn't scale. May defeat the purpose of automation. The human may suffer from automation bias — the tendency to defer to the machine's recommendation, effectively rubber-stamping rather than genuinely evaluating.

Human-on-the-loop (HOTL): The system can act autonomously, but a human monitors its actions and can intervene. The system acts; the human oversees.

  • Example: An autonomous vehicle drives itself, but a remote operator monitors its camera feeds and can take control if something goes wrong.
  • Strengths: Faster than HITL. Allows autonomous operation in routine cases while preserving human override for exceptional cases.
  • Limitations: Vulnerable to the vigilance problem — humans are poor at monitoring systems that work well most of the time. When intervention is needed, the human may not be paying attention, may not have sufficient context, and may not be able to react quickly enough.

Human-out-of-the-loop (HOOTL): The system acts fully autonomously. No human monitors or intervenes in individual decisions.

  • Example: An automated missile defense system detects an incoming threat and launches countermeasures without waiting for human authorization.
  • Strengths: Maximum speed. No human reaction time delay. Appropriate when the decision window is too short for human involvement.
  • Limitations: No human accountability for individual decisions. No ability to exercise contextual moral judgment. The highest governance risk.

19.6.2 Automation Bias

Even when humans are nominally in the loop, a well-documented psychological phenomenon undermines the effectiveness of human oversight: automation bias.

Automation bias is the tendency to trust and defer to automated recommendations, even when they conflict with the human's own judgment or other available evidence. Studies across domains — aviation, medicine, criminal justice, financial analysis — consistently find that humans presented with automated recommendations are significantly less likely to override them than they would be to make the same judgment independently.

In a healthcare context: if a diagnostic AI recommends that a patient's condition is benign, a physician reviewing the case may be less likely to investigate further than they would be without the AI recommendation — even if clinical signs suggest the AI may be wrong. The physician is "in the loop" in a formal sense, but the AI's recommendation has effectively shaped the decision.

"This is the dirty secret of human-in-the-loop systems," Dr. Adeyemi said. "We say the human is in charge. But the human is influenced — sometimes decisively influenced — by the machine's recommendation. If you always agree with the machine, you're not in the loop. You're a signature stamp."

19.6.3 VitraMed and the Human Override Question

VitraMed's development of an autonomous diagnostic assistance tool forced Mira to confront the human oversight question concretely.

"The system is designed to analyze patient data and flag potential conditions that the physician might miss," Mira explained. "In the current design, the physician makes all final decisions — classical human-in-the-loop. But my father's team is pushing toward a version where the system can autonomously order certain follow-up tests — blood panels, imaging — without waiting for physician approval. Their argument is speed: earlier detection saves lives."

"And where is the human override?" Dr. Adeyemi asked.

"That's the question. In the current proposal, the physician can review and cancel any autonomously ordered test within a 30-minute window. But in practice, most physicians are too busy to review routine test orders within 30 minutes. So the system would effectively be operating autonomously."

"Then you don't have human-in-the-loop. You have human-on-the-loop at best, and human-out-of-the-loop in practice," Dr. Adeyemi observed. "The label says one thing. The reality is another. And when something goes wrong — when the system orders an invasive test that wasn't warranted, or misses a condition that an attentive physician would have caught — the question of who is responsible becomes very complicated."

The Consent Fiction in Human Override: Notice the consent fiction operating here. The design says the physician can override within 30 minutes. But if the workflow makes override impractical, the "option" to override is formal rather than real — like the "option" to read a 47-page terms of service agreement. Meaningful human oversight requires not just the formal authority to intervene but the practical conditions that make intervention possible: sufficient time, sufficient information, sufficient expertise, and organizational support for overriding the machine.


19.7 The EU AI Act and High-Risk AI

19.7.1 The Risk-Based Framework

The EU AI Act (Regulation (EU) 2024/1689), which entered into force in August 2024, is the world's first comprehensive legislation regulating AI systems. Its approach is risk-based: different levels of regulatory obligation correspond to different levels of risk.

Risk Category Examples Regulatory Approach
Unacceptable risk Social scoring by governments, real-time biometric identification in public spaces (with exceptions), manipulation targeting vulnerable groups Prohibited
High risk AI in critical infrastructure, education, employment, essential services, law enforcement, migration, justice Extensive obligations
Limited risk Chatbots, emotion recognition, deepfake generation Transparency obligations
Minimal risk Spam filters, AI in video games No specific obligations

19.7.2 High-Risk AI Requirements

For high-risk AI systems — the category most relevant to autonomous systems — the EU AI Act imposes detailed requirements:

Risk management system: Providers must establish a continuous risk management process that identifies, evaluates, and mitigates risks throughout the AI system's lifecycle.

Data governance: Training, validation, and testing data must be relevant, representative, free of errors (to the extent possible), and complete. Data governance practices must address potential biases.

Technical documentation: Providers must maintain detailed documentation of the system's design, development, capabilities, limitations, and risk mitigation measures.

Record-keeping: High-risk AI systems must be designed to automatically log events relevant to their operation, enabling post-hoc analysis of decisions and incidents.

Transparency: Users must be provided with clear information about the system's capabilities, limitations, and appropriate conditions of use.

Human oversight: High-risk AI systems must be designed to allow effective human oversight, including: - The ability for the human to understand the system's capabilities and limitations - The ability to monitor the system's operation - The ability to override or interrupt the system - The ability to decide not to use the system or to disregard its output

Accuracy, robustness, and cybersecurity: The system must achieve appropriate levels of accuracy, be resilient to errors and attempts to manipulate it, and include cybersecurity safeguards.

19.7.3 Implications for Autonomous Systems

The EU AI Act's requirements have significant implications for autonomous systems:

Autonomous vehicles that operate in the EU must comply with high-risk AI requirements, including effective human oversight. The Act does not prohibit Level 4 or Level 5 autonomy, but it requires that the system's design enable meaningful human control — even if that control is exercised remotely or through system-level governance rather than individual-decision oversight.

Medical AI used for diagnostic or treatment purposes is explicitly classified as high-risk. VitraMed's autonomous diagnostic assistant, if deployed in the EU, would need to meet all high-risk requirements — including human oversight that is not merely formal but effective.

Autonomous weapons are not directly addressed by the EU AI Act, which exempts military applications. However, the Act's framework has influenced discussions at the UN level about governance principles for LAWS.

Eli's concern about autonomous policing technology fits directly within the Act's framework. AI systems used by law enforcement — including predictive policing, facial recognition, and risk assessment — are classified as high-risk and must meet the Act's full requirements. This includes requirements for data quality (addressing the historical biases that Chapter 14 documented), human oversight (ensuring that police officers exercise genuine judgment rather than blindly following algorithmic recommendations), and transparency (explaining the system's operation to affected communities).

"The EU AI Act isn't perfect," Sofia Reyes said. "Its exemptions are too broad — military and national security are carved out entirely. Its enforcement mechanisms are still being developed. And there's a real risk that 'high-risk AI requirements' become another compliance checklist that companies satisfy on paper without changing their practices. But it's the most ambitious attempt any jurisdiction has made to regulate AI systematically. It sets a floor. Other jurisdictions can build on it."

19.7.4 Eli and Autonomous Policing

Eli had tracked the deployment of autonomous policing technology in Detroit with growing concern:

"They've introduced predictive deployment — an algorithm that tells police where to concentrate patrols based on predicted crime hotspots. The data that feeds the algorithm comes from historical arrest records. And since Black neighborhoods have been over-policed for decades, the historical data says those neighborhoods are where crime happens. So the algorithm sends more officers there. More officers means more arrests. More arrests feed back into the data. The feedback loop reinforces itself."

"That's the bias problem from Chapter 14," Mira noted.

"Right. But now add autonomous drone surveillance. The department is piloting camera drones that autonomously patrol designated areas — the same areas the predictive algorithm identifies as hotspots. So now you have algorithmic targeting combined with autonomous surveillance, concentrated in Black neighborhoods. Nobody made a conscious decision to surveil Black communities more than white ones. The system did it. And when I go to city council and ask who's responsible, they say 'the algorithm.' The algorithm doesn't show up to council meetings."

The convergence of predictive policing and autonomous surveillance technology illustrates how the themes of Part 3 compound. Bias in training data (Chapter 14) produces unfair predictions (Chapter 15) that are opaque to affected communities (Chapter 16), deployed without meaningful accountability (Chapter 17), and reinforced by autonomous systems that operate with minimal human oversight (this chapter). Each chapter addresses one dimension of the problem. Together, they reveal a system of algorithmic governance that is greater — and more concerning — than the sum of its parts.

The Accountability Gap in Autonomous Policing: When a predictive algorithm directs autonomous drones to surveil a community, and residents are stopped, searched, and arrested at higher rates as a result, who is accountable? The software vendor says the algorithm reflects historical patterns. The police department says it follows the algorithm's recommendations. The city council says it trusted the department's expertise. The residents — disproportionately Black, disproportionately poor — have no effective mechanism for challenge or redress. The accountability gap and the power asymmetry converge.


19.8 Case Studies

19.8.1 The Uber Self-Driving Car Fatality: Accountability in Autonomous Systems

The death of Elaine Herzberg in Tempe, Arizona, on March 18, 2018, is the first documented fatality involving a fully autonomous vehicle striking a pedestrian. The case provides a comprehensive study of the accountability challenges that autonomous systems create.

The Incident: Herzberg was walking her bicycle across a four-lane road at approximately 10:00 PM. An Uber autonomous test vehicle, traveling at 39 mph, struck and killed her. The vehicle's LIDAR, radar, and camera systems detected Herzberg approximately 5.6 seconds before impact. The system classified her variously as a vehicle, as "other," and as a bicycle, never settling on a consistent classification. It did not initiate emergency braking. The vehicle's standard automatic emergency braking system had been disabled by Uber to prevent "erratic" braking behavior that degraded ride quality.

The Safety Driver: Rafaela Vasquez, the human safety driver, was supposed to monitor the road and intervene if the autonomous system failed. Dashboard camera footage showed that Vasquez was looking at her phone — streaming a television show — in the moments before the collision. She looked up approximately 0.5 seconds before impact, too late to intervene.

The Investigation: The NTSB determined that the probable cause was the failure of the safety driver to monitor the driving environment, combined with Uber's inadequate safety culture. Contributing factors included:

  • Uber's decision to disable automatic emergency braking
  • The vehicle's inability to classify and predict the behavior of a pedestrian outside a crosswalk
  • Uber's failure to implement adequate safety oversight procedures for its testing program
  • Arizona's permissive regulatory framework, which allowed autonomous vehicle testing without specific safety requirements

Accountability Outcomes:

Actor Outcome
Rafaela Vasquez (safety driver) Charged with negligent homicide; pleaded guilty; sentenced to 3 years supervised probation
Uber (developer/deployer) Not criminally charged; paid undisclosed settlement to Herzberg's family; suspended autonomous testing for 9 months
Arizona regulators No consequences; subsequently adopted modest additional testing requirements
NHTSA (federal regulator) No enforcement action; had not established specific safety standards for autonomous vehicle testing

Analysis:

The distribution of consequences is revealing. The lowest-ranking actor in the system — the safety driver — bore the most severe legal consequences. Uber, whose engineering decisions (disabling emergency braking, inadequate perception systems) and safety culture (inadequate driver training and monitoring) were the primary causes, faced no criminal liability. The regulators whose permissive framework enabled the testing faced no consequences at all.

This pattern — accountability flowing downward, away from institutional power — is a manifestation of the power asymmetry. The safety driver was a gig-economy contractor with limited resources. Uber was a multinational corporation with teams of lawyers. The regulatory framework had been shaped, in part, by lobbying from the autonomous vehicle industry.

Lessons for governance: 1. Human safety drivers are an inadequate substitute for robust autonomous systems. The vigilance problem ensures that human oversight in long-duration monitoring tasks degrades. 2. Accountability structures must address institutional decisions (system design, safety culture, testing protocols) — not just individual operator failures. 3. Regulatory frameworks for autonomous vehicles must specify safety standards before testing, not develop them in response to fatalities. 4. The consent of the public — who shares roads with autonomous test vehicles — must be obtained through democratic processes, not assumed.

19.8.2 Autonomous Weapons: The Campaign to Stop Killer Robots

The Campaign to Stop Killer Robots is a global coalition of organizations, founded in 2013, that advocates for a preemptive ban on fully autonomous weapons. The campaign represents one of the most significant civil society efforts to govern an emerging technology before it is widely deployed.

Origins: The campaign was launched by Human Rights Watch, the International Committee for Robot Control, and a network of NGOs concerned about the development of weapons that could select and engage targets without meaningful human control. The campaign's founding principle: "Machines should not be allowed to make life-and-death decisions on the battlefield."

Strategy and Activities:

The campaign has pursued a multi-track strategy:

  1. Diplomacy: Lobbying at the UN Convention on Certain Conventional Weapons (CCW), where discussions on LAWS have been ongoing since 2014. The campaign has pushed for a legally binding instrument — a new international treaty — prohibiting fully autonomous weapons.

  2. Public mobilization: Building public awareness and support for a ban. Surveys in multiple countries show that a majority of people oppose autonomous weapons, and the campaign has leveraged this public sentiment.

  3. Technology sector engagement: Engaging with AI researchers and technology companies. In 2018, over 4,500 AI researchers signed a pledge not to participate in the development of lethal autonomous weapons. Google employees protested the company's involvement in Project Maven (a Pentagon AI program), leading Google to withdraw and adopt AI principles that include a prohibition on weapons applications.

  4. Expert advocacy: Producing research reports, policy analyses, and expert testimony to inform diplomatic discussions.

Achievements and Challenges:

The campaign has achieved significant normative impact: the concept of "meaningful human control" over weapons systems is now widely accepted as a governance principle, even among states that oppose a binding ban. The UN Secretary-General has called for restrictions on autonomous weapons. Over 70 states have called for new international rules.

However, a binding treaty remains elusive. Major military powers — the United States, Russia, China, the United Kingdom, and Israel — have resisted binding restrictions, arguing that existing international humanitarian law is sufficient and that a ban would be premature given the uncertain state of the technology.

Analysis:

The campaign illustrates several governance dynamics:

  • Preemptive governance is harder than reactive governance. It is easier to regulate a technology after it has caused documented harm than before. The Campaign to Stop Killer Robots is asking states to ban a capability that has not yet been fully deployed — a harder political ask than responding to an existing crisis.
  • Civil society can shape norms even without legal outcomes. The campaign has not achieved a binding treaty, but it has shifted the terms of debate, established "meaningful human control" as a governance principle, and created reputational costs for companies and countries that develop autonomous weapons.
  • Technical governance and political governance are intertwined. The debate about LAWS is simultaneously a technical question (can autonomous weapons comply with IHL?) and a political question (should the power to kill be delegated to machines?). Technical arguments are necessary but not sufficient — the ultimate decision is a moral and political one.

19.9 Chapter Summary

Key Concepts

Concept Definition
Levels of autonomy A spectrum from full human control (Level 0) to full machine control (Level 5), with intermediate levels of shared human-machine decision-making
Lethal autonomous weapons (LAWS) Weapon systems that can select and engage targets without direct human intervention
Moral agency The capacity to recognize moral dimensions of a situation, deliberate about right and wrong, choose an action, and be held responsible
Moral patient An entity that can be wronged — that has interests or welfare that can be harmed
Responsibility gap The condition in which harmful actions by autonomous systems cannot be attributed to any moral agent — human or machine
Human-in-the-loop A human must approve every significant action before the system executes it
Human-on-the-loop The system acts autonomously while a human monitors and can intervene
Human-out-of-the-loop The system acts fully autonomously with no human monitoring of individual decisions
Automation bias The tendency to trust and defer to automated recommendations, even when they conflict with one's own judgment
Meaningful human control The principle that humans must retain genuine — not merely formal — authority over consequential decisions by autonomous systems
EU AI Act risk categories Unacceptable risk (prohibited), high risk (extensive obligations), limited risk (transparency), minimal risk (no obligations)
Trolley problem A thought experiment about forced binary moral choices, useful for philosophical illustration but misleading as a guide to autonomous vehicle governance

Key Debates

  • Can machines be moral agents, and does the answer matter for governance? Or is the practical question of accountability more important than the philosophical question of moral agency?
  • Should lethal autonomous weapons be banned preemptively, or should governance focus on ensuring meaningful human control?
  • Is the trolley problem a useful framework for autonomous vehicle ethics, or does it distract from the more important systemic questions of safety, equity, and accountability?
  • Can human-in-the-loop oversight be maintained in practice, or does automation bias inevitably degrade human judgment?
  • Does the EU AI Act's risk-based approach represent an adequate governance framework for autonomous systems, or are its categories and requirements insufficient?

Applied Framework

The Autonomous Systems Governance Assessment: 1. Classify the level of autonomy. Where does the system fall on the autonomy spectrum (Levels 0-5)? 2. Identify the decision stakes. What are the consequences of the system's decisions? Are they reversible or irreversible? High-stakes or low-stakes? 3. Map the human oversight model. Is there a human in the loop, on the loop, or out of the loop? Is the human oversight effective in practice, or merely formal? 4. Assess automation bias risk. Are the humans who oversee the system likely to defer to its recommendations? What safeguards exist against automation bias? 5. Evaluate accountability. Who is responsible when the system causes harm? Is there a clear accountability chain, or a responsibility gap? 6. Apply the proportionality principle. Are governance requirements proportional to the system's risk level? (Higher autonomy + higher stakes = stronger governance requirements) 7. Test for consent and community voice. Have the people affected by the system been consulted? Do they have meaningful input into deployment decisions?


What's Next

Chapter 19 closes Part 3: Algorithmic Systems and AI Ethics. Over seven chapters, we have examined how algorithms shape society (Chapter 13), how data and algorithmic bias operate (Chapter 14), what fairness means and why it's contested (Chapter 15), why transparency matters and why it's hard (Chapter 16), who is responsible when things go wrong (Chapter 17), what happens when AI creates (Chapter 18), and what happens when AI acts autonomously (this chapter).

The threads converge: bias compounds through opaque systems deployed without meaningful accountability, generating harms that fall disproportionately on communities with the least power to resist. The accountability gap, the consent fiction, and the power asymmetry operate in every domain we've examined.

But awareness of these problems is not the same as solving them. Part 4: Governance and Regulation turns to the institutional and legal responses that societies are building — beginning with Chapter 20: The Regulatory Landscape: A Global Survey, which maps the diverse approaches that different nations and regions are taking to govern data and AI. We move from diagnosing the problem to evaluating the solutions.


Chapter 19 Exercises → exercises.md

Chapter 19 Quiz → quiz.md

Case Study: The Uber Self-Driving Car Fatality: Accountability in Autonomous Systems → case-study-01.md

Case Study: Autonomous Weapons: The Campaign to Stop Killer Robots → case-study-02.md