49 min read

In November 2020, Timnit Gebru received an email from a manager at Google telling her that she had been given the option to either retract a research paper she had co-authored or resign. She had not been consulted about the deadline. She had not...

Chapter 22: Whistleblowing and Ethical Dissent in AI Organizations

When the Ethics Researchers Leave: Dissent and Its Costs

In November 2020, Timnit Gebru received an email from a manager at Google telling her that she had been given the option to either retract a research paper she had co-authored or resign. She had not been consulted about the deadline. She had not been informed of the specific objections to the paper beyond a brief note suggesting it did not meet Google's publication standards. She responded, not with retraction, but with a message to colleagues expressing her frustration at the process. That message was treated by Google management as her resignation. Gebru says she was fired. Google says she resigned. The distinction matters legally; the facts are not in material dispute.

The paper in question — "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?" — examined the risks of large language models: their enormous computational and environmental costs, their tendency to reproduce and amplify social biases present in training data, and the concentration of the capacity to build and deploy them in the hands of a small number of powerful organizations. The paper did not claim that language models should not be built. It argued that the risks deserved serious attention and that the field's orientation toward building ever-larger models without adequate attention to these risks was a governance failure.

The paper's publication was ultimately delayed and then published after Gebru's departure, with her name removed from the author list — a decision that generated significant controversy in the research community. Margaret Mitchell, the other co-lead of Google's Ethical AI team, was fired two months later, reportedly while investigating the circumstances of Gebru's dismissal.

Two senior Black women — the most prominent AI ethics researchers at the world's largest AI company — had been pushed out in quick succession. The organization's response to internal dissent had told the AI ethics field something important: that the relationship between large AI companies and their internal ethics functions was more precarious than the organizational charts suggested, and that the protection available to employees who raised serious concerns about their employer's most consequential products was, in many cases, inadequate.

This chapter examines the role of ethical dissent in AI organizations. Why is dissent a critical governance mechanism? Why is it so difficult? What legal protections exist — and where do those protections fall short? What organizational conditions make it possible for employees to raise ethics concerns without facing career destruction? And what should individuals do when they face situations where they cannot reconcile their professional responsibilities with their ethical convictions?


Learning Objectives

By the end of this chapter, students will be able to:

  1. Explain why internal ethical dissent is a critical AI governance mechanism and articulate the organizational value of effective dissent channels.
  2. Identify the psychological, social, and organizational barriers that cause employees to remain silent about AI ethics concerns, including Bandura's moral disengagement mechanisms.
  3. Assess the legal protections available to AI ethics whistleblowers in the United States, identify the significant gaps in those protections, and compare them to international frameworks.
  4. Analyze high-profile AI whistleblowing and dissent cases — including Timnit Gebru, Margaret Mitchell, Sophie Zhang, and Frances Haugen — identifying patterns in how organizations respond to internal ethical dissent.
  5. Evaluate the organizational conditions that support or suppress ethical dissent, including the role of psychological safety, structural mechanisms, and leadership behavior.
  6. Apply the principles of responsible disclosure to evaluate when and how AI ethics concerns should be escalated externally, including the ethics of going public with organizational misconduct.
  7. Articulate the organizational business case for welcoming internal dissent and describe how organizations benefit from cultures that make ethical concern-raising safe and normal.
  8. Describe the emerging legal and institutional landscape for AI ethics whistleblowing and identify the mechanisms through which it is likely to evolve.

Section 22.1: The Role of Internal Dissent in AI Ethics

Dissent as a Governance Mechanism

Chapter 21 examined the formal structures through which organizations govern their AI development: ethics boards, responsible AI teams, principles documents, and review processes. This chapter examines something different but equally important: the informal, human-scale mechanisms through which governance actually functions or fails to function — specifically, the ability of individual employees to raise ethical concerns, have those concerns heard, and see them translated into action.

Internal dissent is not merely a safety valve for individual employees whose conscience is troubled. It is, properly understood, one of the most important governance mechanisms available to organizations. Employees who work directly with AI systems — engineers who train models, data scientists who curate datasets, product managers who define use cases, user researchers who observe real-world impacts — know things about those systems that no governance body, ethics board, or executive team can independently know. They see the edge cases, the anomalous outputs, the demographic performance gaps that don't appear in summary reports, and the gap between how systems are described in management presentations and how they actually behave.

Organizations that create conditions under which these employees can raise their concerns have an enormous governance advantage: they get early warning about problems while they can still be addressed, before deployment locks them in, before harms reach scale, and before they become public. Organizations that suppress this dissent — through retaliation, structural barriers, or cultural norms that discourage concern-raising — are blind. They learn about problems when they become crises, not when they could be addressed. The governance cost of silencing internal voices is paid in discrimination lawsuits, regulatory enforcement actions, public boycotts, and reputational damage.

The Dissent Spectrum

Ethical dissent in AI organizations exists on a spectrum, from the most informal to the most consequential.

At the informal end, dissent looks like a comment raised in a team meeting: "I'm worried this evaluation metric doesn't capture fairness across demographic groups." It looks like a conversation with a manager: "I've been thinking about this model's behavior in edge cases, and I want to flag something before we proceed." These informal moments are where most ethical concerns begin, and they are where most ethical governance happens or fails to happen. If the culture of the team makes these informal raises normal and welcomed, many concerns will be addressed before they require escalation. If the culture makes them risky — if the engineer who raises them is labeled a troublemaker or the concern is dismissed without genuine consideration — the problem is driven underground.

Formal internal channels — ethics hotlines, ombudspersons, ethics committees, formal complaint processes — represent a more structured level of dissent. These mechanisms allow employees to raise concerns more formally, typically with some degree of anonymity protection, and to have those concerns reviewed by parties outside the immediate management chain. The effectiveness of these channels depends entirely on whether they have genuine independence (not just nominal independence) and whether employees believe, based on organizational track record, that using them will result in genuine review rather than retaliation.

Internal escalation — taking a concern directly to senior leadership, to the board, to the ethics committee — represents a higher-stakes level of dissent. It typically occurs when informal and formal channels have failed to produce adequate response. The willingness of employees to escalate, and the willingness of senior leadership to hear and act on escalated concerns rather than punishing the escalator, is a critical indicator of organizational ethics culture.

External disclosure — going to the media, to regulators, to the public — represents the highest-stakes level of dissent, and the one with the most severe consequences for the disclosing individual. External disclosure typically occurs when internal channels have been exhausted without adequate response to a concern serious enough that the employee believes the public has a right to know.

Finally, resignation — leaving the organization when internal concerns cannot be addressed and the employee cannot continue to participate in what they believe is harmful — is a form of dissent that removes the individual from the problem without necessarily solving it. Strategic resignation, accompanied by public explanation (as in the Axon ethics board case examined in Chapter 21), can be more consequential.

The Organizational Value of Dissent

The governance value of dissent is empirically supported. Organizations with strong internal voice mechanisms — cultures where employees feel safe raising concerns, structural channels for formal concern-raising, and demonstrated track records of acting on legitimate concerns — catch more problems earlier. They also attract and retain employees who care about ethical practice: the engineers, data scientists, and product managers who will self-select for organizations where their ethical commitments will be honored.

The analogy to financial audit is instructive. Internal financial controls are not a concession to employee distrust; they are a governance requirement that protects the organization by catching errors and misconduct before they become disasters. The auditor who flags a problem is providing a service to the organization, not making trouble. AI ethics internal review mechanisms serve exactly the same function. The employee who identifies a bias problem in a model three weeks before deployment has given the organization an enormous gift — the opportunity to fix the problem before it causes harm at scale.


Vocabulary Builder

  • Whistleblowing: The disclosure by an employee or former employee of information about organizational misconduct that the individual believes is illegal, unethical, or otherwise contrary to the public interest.
  • Ethical dissent: The act of expressing disagreement with organizational decisions or practices on ethical grounds, through any mechanism on the dissent spectrum.
  • Retaliation: Adverse employment action taken against an employee because of protected activity, such as reporting legal violations or raising ethical concerns.
  • Organizational silence: The collective phenomenon in which employees do not speak up about problems they observe, typically because they believe raising concerns would be futile or risky.
  • Moral disengagement: Psychological mechanisms that allow individuals to maintain a positive self-image while participating in practices that cause harm, through cognitive reframing of those practices.
  • Psychological safety: The belief that one can speak up, raise concerns, take risks, or make mistakes without facing punishment or humiliation.

Section 22.2: Why Employees Stay Silent — Organizational Barriers

The Documented Pattern of Retaliation

The most direct reason employees stay silent about AI ethics concerns is the most obvious one: they have seen what happens to people who speak up. The firing of Timnit Gebru and Margaret Mitchell at Google, the termination of Sophie Zhang at Facebook after she raised concerns about inauthentic political manipulation on the platform, the reported difficulties faced by employees who objected to Amazon's sale of Rekognition to ICE — these high-profile cases are visible precisely because the individuals involved were prominent enough that their treatment attracted public attention.

For every highly publicized case, there are many more cases that never become public: the engineer whose objection to a discriminatory system was overridden and who was quietly passed over for promotion; the data scientist who pushed back on a problematic training dataset and was moved to a less desirable team; the responsible AI reviewer whose critical assessment was accepted on paper and ignored in practice, who resigned rather than fight the battle again.

Organizations where retaliation against ethics concerns has occurred — whether or not it is labeled as such — send a clear message to every other employee. That message does not require explicit communication. It is legible in what happened to the person who spoke up: their treatment teaches everyone who observes it that raising concerns is risky and that the organizational narrative will not acknowledge the connection between the concern and the consequence.

Moral Disengagement Mechanisms

Albert Bandura's theory of moral disengagement describes the psychological mechanisms through which individuals who hold genuine ethical values participate in organizations that cause harm without experiencing the cognitive dissonance that would otherwise motivate them to object or exit. These mechanisms are not unique to AI development — they are general features of organizational psychology — but they have specific manifestations in the AI context.

Moral justification reframes harmful activity as serving a larger good: "This AI system may have some inaccuracies, but overall it will help police departments allocate resources more effectively, saving lives." The harms to specific individuals who are incorrectly flagged are made cognitively tolerable by reference to the aggregate benefit. In AI contexts, consequentialist framing of this kind is particularly common: the claim that aggregate statistical benefit justifies distributional harm to specific individuals is embedded in how many AI systems are designed and how their performance is evaluated.

Euphemistic labeling substitutes technical or neutral language for moral language that would make the harm visible: AI systems don't "discriminate," they "optimize"; they don't "surveil," they "monitor"; they don't "predict criminality," they "assess risk." The language of data science is remarkably effective at making ethical questions invisible. A system described as "a machine learning classifier trained on historical case outcomes" sounds neutral and technical; described as "a system that predicts who will commit crimes based on data reflecting decades of racially discriminatory policing" it sounds different — because it is different, and the technical framing obscures rather than describes what is happening.

Displacement of responsibility allows individuals to locate accountability elsewhere in the organizational chain: "I just write the code; management decides what to do with it." This mechanism is particularly powerful in large organizations with complex hierarchies, where any individual's contribution to an AI system is genuinely small. The engineer who trained the model can point to the product manager who defined the use case; the product manager can point to the executive who approved deployment; the executive can point to the legal team that cleared it; and the legal team can point to the regulators who have not prohibited it. At each step, responsibility is displaced to somewhere else in the chain, and at no step does anyone experience themselves as the person who decided to cause harm.

Diffusion of responsibility operates at the group rather than individual level: when many people are involved in a decision, each person's sense of individual responsibility is diluted. Teams that deploy AI systems collectively may experience this diffusion: "Someone must be checking for this." "The responsible AI team reviewed it." "Legal cleared it." These representations — which may even be accurate as descriptions of what formally occurred — reduce each individual's sense of moral responsibility for the outcome.

Group Norms and Confirmation Bias

Teams develop norms around what concerns are appropriate to raise and how. Teams that have never raised ethical concerns, or where ethical concerns have been met with dismissal, develop implicit norms that ethics is outside the scope of professional discussion. An individual who raises an ethics concern in such a team is violating a social norm, not just a professional expectation — they are marking themselves as different, as a potential source of delay and friction, as someone who is not a full team player.

Confirmation bias compounds this problem. Teams that are proud of their work — that believe, as most technologists do, that they are building something beneficial — are less likely to notice or credit evidence that their systems cause harm. The disposition that produces excellent engineering (confidence in your design, commitment to your approach, willingness to push through obstacles) also produces resistance to evidence that the approach is causing problems you didn't intend.

The "brilliant jerk" culture that has been documented in parts of the technology industry is particularly hostile to ethical dissent. When technical excellence is treated as a trump card — when the engineer who ships the fastest and builds the most impressive systems is insulated from accountability for the ethics of what they ship — ethical concerns become professionally precarious. The engineer who raises a bias concern about a technically impressive system is challenging the primacy of technical achievement as the dominant professional value. This is a significant social risk.


The Federal Landscape: Existing Protections

The United States has a complex patchwork of federal whistleblower protections, developed over decades in response to specific categories of corporate misconduct. Understanding these protections — and their significant limits in the AI ethics context — is essential for anyone advising employees who are considering raising concerns about AI ethics violations, and for organizations designing governance programs that include employee reporting.

The Sarbanes-Oxley Act (SOX), passed in 2002 following the Enron and WorldCom accounting scandals, provides whistleblower protection to employees of publicly traded companies who report violations of securities laws, SEC rules, or any federal law relating to fraud on shareholders. SOX protection covers reports to supervisors, law enforcement, or Congress; it prohibits retaliation and provides for reinstatement, back pay, and attorney's fees. SOX is relevant to AI ethics primarily when AI misconduct involves financial fraud or securities violations — when, for example, a company misrepresents the accuracy or fairness of its AI systems to investors, or when AI systems are used to commit financial fraud.

The Dodd-Frank Wall Street Reform and Consumer Protection Act (2010) created the SEC's whistleblower program, which allows individuals to report securities law violations to the SEC and provides financial awards (10–30% of sanctions over $1 million) to those whose information leads to successful enforcement. Dodd-Frank also includes strong anti-retaliation provisions and, crucially, limits organizations' ability to use non-disclosure agreements to prevent employees from communicating with the SEC. If an AI company misrepresents its AI capabilities, safety, or ethics record to investors in a way that violates securities laws, Dodd-Frank may protect and incentivize the employee who reports it.

The False Claims Act provides protection to employees who report fraud against the federal government, including through qui tam actions in which the whistleblower can sue on behalf of the government and share in recovered damages. This protection is most relevant when AI misconduct involves government contracts — when, for example, an AI vendor sells systems to federal agencies while knowing the systems do not meet the performance or safety specifications represented in the contract.

OSHA whistleblower protection programs cover retaliation under approximately 25 federal statutes addressing workplace safety, environmental law, and sector-specific regulations. Some of these protections apply in contexts relevant to AI deployment — notably in transportation (autonomous vehicle safety), nuclear energy, and financial services.

The Gap: No Federal AI Ethics Whistleblower Protection

Despite this patchwork of protections, there is no federal law that specifically protects employees who report AI ethics concerns that do not rise to the level of fraud, securities violation, or specific regulatory violation. The engineer who raises concerns about a discriminatory AI hiring tool that is not (yet) illegal under federal law is not protected by any of the statutes described above. The data scientist who reports that her employer has deployed an AI system with known accuracy problems that harm specific demographic groups — but that does not clearly violate any existing federal law — has no federal whistleblower protection if she faces retaliation for doing so.

This is a significant and increasingly recognized gap. AI systems cause harms that existing regulatory frameworks have not yet fully addressed, and the employees best positioned to identify and report those harms have no specific legal protection for doing so. Congress has considered but not enacted legislation that would extend whistleblower protections to AI ethics concerns specifically.

State Protections

State employment law provides some protection that federal law does not. California Labor Code Section 1102.5, one of the broadest state whistleblower protection statutes, prohibits retaliation against employees who disclose information they reasonably believe violates a law, regulation, or standard — including to internal supervisors, not just to regulators. New York Labor Law Section 740 provides similar protections. Given that many major AI companies are headquartered in California, Section 1102.5 is practically significant for a large portion of the AI industry workforce.

State anti-retaliation protections vary significantly, however. An employee working in a state with weaker whistleblower protection statutes may have fewer options than a California-based colleague. The patchwork nature of state protections means that available legal remedies are highly jurisdiction-dependent.

International Frameworks

The European Union's Whistleblower Protection Directive (2019), which member states were required to implement by December 2021, provides significantly more comprehensive protection than US federal law. The directive requires member states to establish protection for whistleblowers who report violations of EU law across a wide range of domains, including product safety, data protection (which directly encompasses AI), financial services, and consumer protection. Importantly, the directive requires organizations above a certain size to establish internal reporting channels and prohibits a broad range of retaliation.

For AI ethics whistleblowing specifically, the GDPR's data protection framework — which the EU Whistleblower Directive covers — means that EU employees who report AI systems' violation of GDPR requirements have protection that US employees lacking parallel federal protection do not. As the EU AI Act's provisions take full effect, employees reporting violations of high-risk AI requirements will have additional protected ground.

The UK's Public Interest Disclosure Act (PIDA) predates the EU directive and provides protection for disclosures made in the public interest, including to regulators and (in some circumstances) to the media. PIDA's public interest standard means that AI ethics concerns — if serious enough — may qualify for protection.

Timnit Gebru's situation illustrates the legal gaps in AI ethics whistleblower protection with painful clarity. Gebru was a Google employee in California — which provides among the broadest state whistleblower protections in the country. California Labor Code 1102.5 protects employees from retaliation for disclosing information they reasonably believe violates a law. But the "Stochastic Parrots" paper was an academic research paper, not a disclosure to a regulator or law enforcement authority. The concerns it raised, while serious, were not framed as violations of specific laws. The circumstances of Gebru's departure — whether she resigned or was fired — were themselves disputed.

The point is not that Gebru had no legal recourse; the legal analysis of her specific situation is complex and depends on facts not publicly available. The point is that an employee at one of the world's most sophisticated technology companies, raising concerns about significant AI risks through academic publication — one of the most established and legitimate mechanisms for scientific concern-raising — found that the legal protection available to her was uncertain. For employees at organizations with fewer resources, without access to sophisticated legal counsel, and working in states with weaker whistleblower protections, the picture is considerably worse.


Section 22.4: What Happens When Employees Speak Out — The High-Profile Cases

Timnit Gebru — Google, 2020

Timnit Gebru joined Google in 2018 as a research scientist and quickly rose to become co-lead of Google's Ethical AI team alongside Margaret Mitchell. Her research focused on bias in AI systems and the intersection of AI with social equity — work that had earned her a prominent reputation in the AI ethics field before she joined Google.

The paper that precipitated her departure — "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?" — was co-authored with Emily M. Bender, Angelina McMillan-Major, and Mitchell. Its argument was not that language models should not be built; it was that the field's orientation toward scale — the assumption that bigger models were always better — came with costs that were inadequately examined. Environmental costs: training a large language model generates carbon emissions comparable to five cars over their lifetime. Data curation costs: massive web-scraped datasets encode social biases at scale, and the larger the model, the larger the scale of the encoding. Concentration costs: the resource requirements for large language models concentrate AI capability in the hands of a small number of powerful organizations.

Google management objected to the paper before its publication — the specific objections have been described variously as concerning the paper's tone, its omission of positive research, and the speed of the review process. Gebru's refusal to retract triggered the chain of events that ended with her departure. The paper was subsequently published in the proceedings of the FAccT conference in 2021; Gebru's name was eventually restored.

The aftermath was significant. Hundreds of Google AI researchers signed an open letter expressing solidarity with Gebru. External AI researchers and ethicists widely condemned Google's handling of the situation. The episode prompted serious discussion in the AI ethics field about the structural position of ethics researchers within large AI companies — whether internal AI ethics teams could operate with genuine independence when the research they produced might challenge the organization's core commercial interests.

Margaret Mitchell — Google, 2021

Margaret Mitchell, co-lead of Google's Ethical AI team alongside Gebru, was fired in February 2021. According to public reports, she was fired after using an automated script to search her own work emails for documentation that might support Gebru's case against Google — documentation she intended to provide to Gebru's lawyers. Google's stated reason was that Mitchell had violated company policy by moving files outside the organization.

The pattern was not lost on observers: two senior Black women, the most prominent AI ethics researchers at the world's largest AI company, had been terminated within two months of each other. The first was terminated in circumstances relating to a paper questioning the ethics of large language model development. The second was terminated while investigating the circumstances of the first termination.

Mitchell, like Gebru, had been a prominent voice advocating for diversity in AI research and for attention to the harms that AI systems could cause. Her firing completed a picture that many in the AI ethics field found deeply troubling: that Google's commitment to internal AI ethics research had its limits, and those limits were reached when the research challenged fundamental assumptions about the company's core business.

Sophie Zhang — Facebook, 2020

Sophie Zhang was a data scientist at Facebook responsible for identifying fake accounts and inauthentic political manipulation on the platform. In 2020, she was fired — a termination she and Facebook characterize differently. Before and after her departure, she documented what she had found: extensive evidence of inauthentic political manipulation — fake accounts, coordinated behavior, government-linked influence operations — in countries including Azerbaijan, Honduras, Ukraine, India, and Spain.

Zhang's internal memo, which she circulated among colleagues before her departure, ran to approximately 6,500 words and documented specific cases of political manipulation that she believed Facebook had failed to adequately address. After her departure, she shared the memo with reporters, and it was published in BuzzFeed News. The memo presented a picture of Facebook's political integrity team as severely under-resourced relative to the scale of the problem, and of senior management as making decisions about which governments' manipulation to address based on calculations of political risk to Facebook rather than scale of democratic harm.

Zhang's case illustrates several aspects of AI ethics whistleblowing that are less visible in the Gebru case. First, the harms she documented were not about AI bias in the narrow technical sense; they were about the use of AI platform tools to manipulate democratic processes. Second, her disclosure strategy — sharing an internal memo with journalists after her departure — is a common and legally risky form of external disclosure. Third, she described the internal experience of trying to address these problems within Facebook before going public: raising concerns through official channels, feeling those concerns were not adequately prioritized, and ultimately concluding that external disclosure was necessary.

Frances Haugen — Facebook, 2021

Frances Haugen's disclosure is the most consequential whistleblowing case in the technology sector to date, both in its scale and in the sophistication of the disclosure strategy she employed. Her case is examined in depth in Case Study 22.2; here, the key narrative elements are introduced in the context of the broader pattern.

Haugen, a former product manager in Facebook's civic integrity team, spent months before her departure copying internal research documents — studies Facebook had conducted on the harms of its platform, including research on Instagram's effects on teenage girls' mental health and research on the amplification of political misinformation. She disclosed the documents to the SEC before going to journalists, establishing a legal record and regulatory engagement before media disclosure. She worked with a team of journalists at the Wall Street Journal, who published a series of investigative stories based on the documents. She testified before Congress. She made disclosures to European regulators. Her disclosures were carefully planned, legally structured, and strategically sequenced.

Haugen's case demonstrates that highly effective whistleblowing — whistleblowing that actually changes organizational behavior and regulatory response — requires preparation, legal advice, strategic judgment, and media sophistication. It is not simply speaking truth to power; it is a complex strategic undertaking that must be executed carefully to achieve its intended effect.

The Google Walkout Organizers — 2018

In November 2018, approximately 20,000 Google employees walked out of their offices in a coordinated global protest over the company's handling of sexual harassment allegations against senior executives, including payments to departing executives accused of misconduct. The walkout was also a statement about Google's Project Maven — the Defense Department contract — and the culture of impunity the organizers believed had developed within the company.

The walkout was striking in its scale and coordination. It demonstrated that organized collective action by technology employees could attract public attention and create organizational pressure. It also demonstrated the limits of collective action: some of the principal organizers subsequently reported experiencing retaliation, including demotions, changes in role responsibilities, and hostile treatment from management. Their experiences were documented in a subsequent report to the National Labor Relations Board.

Pattern Analysis

These cases reveal several common features worth examining as a pattern.

The individuals who spoke out were prominent and credible. Gebru, Mitchell, Zhang, and Haugen were not marginal employees with grievances; they were experienced professionals with strong track records. This prominence did not protect them — it may have increased their threat to management. But it did give their subsequent disclosures credibility that less prominent employees might not have achieved.

The organizations' responses were formal but contested. Google and Facebook both articulated formal policy-based reasons for the departures of Gebru, Mitchell, Zhang, and Haugen. The factual accuracy of these formal justifications was disputed in each case. This pattern — the pretextual termination, in which a policy violation is cited as the formal reason for an exit that critics attribute to retaliation — is common in whistleblower cases and is one reason that legal whistleblower protections require employees to demonstrate that the protected activity was a contributing factor in the adverse action, even if it was not the only stated reason.

Internal channels were tried and found inadequate. In each case, the individual who eventually disclosed externally or publicly had attempted to raise concerns through internal channels before doing so. This pattern challenges the claim, sometimes made by organizations defending their handling of whistleblower situations, that external disclosure was premature or that internal processes were not given adequate opportunity. In each case examined here, internal channels were engaged and found inadequate — the concerns were not acted on, were dismissed, or were met with actions that the individual experienced as hostile.

The costs to the individuals were significant and lasting. Reputational damage, legal costs, career disruption, and personal stress were common consequences for those who spoke out. The people described in this section were unusually prominent and unusually resourced; most employees who raise significant ethics concerns do not have their professional networks, their ability to attract public support, or their access to legal counsel. The costs they bore, substantial as they were, are likely lower than what a less prominent employee would experience in similar circumstances.


Section 22.5: Building Organizational Cultures That Welcome Dissent

Psychological Safety as Governance Infrastructure

Amy Edmondson's foundational research on psychological safety demonstrated that teams whose members feel safe speaking up — raising concerns, admitting mistakes, challenging assumptions — consistently outperform teams where members feel unsafe doing so. The mechanism is not surprising: teams with psychological safety access more information, catch more problems, and learn faster from errors. Teams without it are blind to the information their members hold but will not share.

Psychological safety in the AI ethics context means specifically that employees can raise ethics concerns without fear that doing so will damage their careers, their relationships with their manager, or their position on the team. This is a higher bar than psychological safety in general, because ethics concerns can be uniquely threatening to organizations: they may challenge high-priority projects, create legal exposure by documenting known problems, or publicly signal disagreement with leadership decisions.

Building AI-specific psychological safety requires deliberate organizational work. It is not sufficient to have a general culture of openness if that culture does not explicitly extend to ethics concerns. Organizations should be explicit: raising ethics concerns is a valued professional contribution, and employees who do so will be protected from retaliation and recognized for their contribution to the organization's ethics practice.

Structural Mechanisms

Structural mechanisms for concern-raising can make dissent safer by providing anonymity, independence, and procedural protection.

Anonymous ethics hotlines and reporting systems allow employees to raise concerns without identifying themselves, reducing the career risk associated with concern-raising. The value of anonymity depends on the credibility of the anonymous channel: if employees believe that reported concerns are investigated seriously and that reporters' identities cannot be deduced from the specific concerns they raise, anonymous channels can elicit concerns that employees would not raise under their own names.

Independent ombudspersons provide a confidential, neutral resource for employees who want to discuss ethics concerns without committing to formal reporting. The ombudsperson does not adjudicate concerns or impose solutions; they provide a safe space for employees to think through their concerns, understand their options, and decide how to proceed. The ombudsperson model has been widely used in higher education and is increasingly adopted in large organizations with significant ethics risk exposure.

Explicit non-retaliation policies — written policies that clearly prohibit adverse action against employees who raise ethics concerns in good faith — provide a formal commitment against which the organization can be held accountable. These policies are only as effective as the organizational culture that surrounds them; a non-retaliation policy in an organization with a pattern of retaliating against dissenters is not a meaningful protection. But their absence is itself a governance signal.

Pre-mortems — structured exercises in which teams imagine that a project has failed and work backward to identify what went wrong — are a practical mechanism for surfacing ethics concerns in a context that makes raising them psychologically safer. In a pre-mortem, articulating concerns is the task, not a deviation from the task. Teams that regularly conduct pre-mortems develop the habit of examining failure modes, including ethical failure modes, as a normal part of project planning.

Leadership Signals

The most powerful signal about psychological safety in an organization is what happens to the first person who raises an ethics concern. If the concern is heard, taken seriously, thanked, and acted on, the signal to everyone who observes the interaction is clear: raising concerns is safe and valuable. If the concern is dismissed, and the person who raised it is subsequently disadvantaged in any way, the signal is equally clear: this organization does not welcome dissent.

Leaders' responses to ethics concerns are observed carefully and interpreted accurately by the people who report to them. Leaders who believe they are creating an open culture while privately discouraging or dismissing ethics concerns are not creating an open culture. The culture is what happens, not what is claimed to happen.

The Role of HR

Human resources functions are structurally positioned in a role that creates a tension for AI ethics dissent. HR's primary client, in most organizations, is the organization's leadership, not individual employees. HR's function is to manage employment relationships in a way that minimizes the organization's legal exposure and supports organizational goals. This positioning can make HR an unreliable ally for employees raising ethics concerns, particularly when those concerns implicate senior management.

Organizations that are serious about creating genuine dissent protection need to address this tension explicitly. One approach is to give ethics reporting channels institutional independence from HR — routing concerns through a compliance or legal function, through an independent ombudsperson, or directly to the board or board committee. Another approach is to explicitly define HR's role in ethics concern situations as protective of the reporting employee, with clear protocols for independent review of retaliation allegations.


Section 22.6: Responsible Disclosure — The Ethics of Going Public

The Whistleblower's Ethical Calculus

The decision to disclose organizational misconduct externally — to journalists, regulators, or the public — involves a genuine ethical calculus, not merely a legal risk assessment. The employee who has exhausted internal channels and is considering external disclosure faces real competing obligations: the obligation to address serious harm, the obligation not to cause disproportionate damage to innocent colleagues or the organization as a whole, the obligation of accuracy in the claims they make, and the obligation to choose disclosure targets and disclosure scope proportionate to the harm being addressed.

These obligations are not simply in tension with career self-interest — though they are that. They are genuine ethical obligations that deserve careful thought. External disclosure is a powerful action with consequences that extend beyond the disclosing individual. Done thoughtlessly, it can cause unnecessary harm. Done carefully, it can change organizational behavior, regulatory attention, and public understanding in ways that prevent significant harm.

The first ethical question is whether internal channels have been genuinely exhausted. Going public before making a serious effort through internal channels is generally not justified unless internal channels are structurally captured by the interests being disclosed. If a concern can be addressed internally — if the organization, made aware of the problem through appropriate channels, would take meaningful action — then internal resolution is both faster and less costly. But "genuine exhaustion" means giving internal channels a fair opportunity to work, not simply filing a complaint and immediately going to the press when it is not instantly resolved.

The second question is the severity and urgency of the harm. Not every organizational ethics failure justifies external disclosure. The employee considering external disclosure should ask: how serious is the harm, to how many people, with what certainty? A concern about a theoretical future harm that is speculative and marginal is different from evidence of ongoing harm to identifiable populations at known scale. The threshold for external disclosure should be calibrated to the severity and certainty of the harm.

Responsible Disclosure Standards

Security research has developed a practice of "responsible disclosure" — the practice of informing a software vendor of a discovered vulnerability before publishing it publicly, giving the vendor an opportunity to patch the vulnerability before it can be exploited. The responsible disclosure norm has developed because the alternative — immediate public disclosure — creates a window during which bad actors can exploit the vulnerability before it is patched.

A parallel responsible disclosure norm is developing in AI ethics, though it is less formalized than in security research. The principle is similar: where possible, inform the organization of the problem before going public, give a reasonable opportunity for internal correction, and limit public disclosure to what is necessary to motivate that correction or, when internal correction does not occur, to protect the public from ongoing harm.

The Journalist Relationship

Reporters covering AI ethics and technology misconduct are valuable allies for employees whose internal disclosures have been ineffective. But the relationship requires care. Journalists are not lawyers; they cannot provide legal advice about employment risk. Their interests — in a compelling story — are not identical to the disclosing employee's interests. Documents shared with journalists may be shared with others, may be used in ways the source did not anticipate, and cannot be recalled once provided.

Employees considering going to journalists should consult legal counsel first, understand what disclosures are protected (so that the disclosure itself does not create legal liability), understand what NDAs they have signed and what those NDAs cover (they cannot generally prohibit communication with the SEC or with law enforcement, but they may cover other disclosures), and understand the difference between "on background" and "on the record."

Non-Disclosure Agreements and Their Limits

Many technology employees have signed NDAs as conditions of employment or upon separation. These agreements typically prohibit disclosure of confidential business information. Their scope and enforceability vary, and several federal laws limit what NDAs can cover.

SEC rules prohibit NDAs from preventing employees from communicating with the SEC about potential securities violations. Dodd-Frank specifically protects employees from retaliation for going to the SEC even if an NDA purports to prohibit it. OSHA's whistleblower protection programs similarly protect disclosures to appropriate regulatory authorities from NDA enforcement. State laws also vary in their treatment of NDAs that purport to cover disclosures of illegal activity.

Employees who have signed NDAs and are considering external disclosure should consult an employment attorney who specializes in whistleblower law before making any disclosures, to understand the specific protections and risks that apply to their situation.


Section 22.7: The Organizational Case for Welcoming Dissent

Dissent as Risk Management

The argument for suppressing internal ethics dissent — while rarely made explicitly — rests on a short-term calculation: that allowing employees to raise and escalate ethics concerns will slow development, create legal exposure (by documenting known problems), and complicate management decision-making. These concerns are not imaginary. Documented concerns about known problems can become evidence in litigation. Ethics debates can slow decision-making. Formal escalation processes add friction to development.

But these short-term costs need to be weighed against the long-term costs of suppressing dissent. Organizations that silence internal voices learn about problems when they become external crises, at which point the options for response are narrower and the costs are higher. A discriminatory AI system that reaches deployment over employees' suppressed objections may face regulatory enforcement, class action litigation, and public boycott — each of which is vastly more costly than the development delay that addressing the problem before deployment would have required.

The risk management case for welcoming dissent is not complicated: organizations that hear about problems early can fix them cheaply; organizations that suppress early warning pay for problems at scale.

The PR Cost of Suppression

When internal dissenters go public — as Gebru, Mitchell, Zhang, and Haugen did — the reputational damage is substantially greater than if the problem had been addressed internally. The story is not only "company deployed a harmful AI system." The story is "company deployed a harmful AI system and fired the employee who tried to stop it." The second story is significantly more damaging because it adds a dimension of organizational malfeasance to the underlying product problem.

Organizations that suppress dissent are not avoiding public scrutiny; they are converting internal problems into more damaging external ones. The Frances Haugen case is a clear demonstration: Facebook's internal research on its platform's harms was the problem; the existence and suppression of that research was what made the story consequential to public policy and regulatory response.

Talent

The AI ethics field's most thoughtful practitioners — the researchers and engineers who understand the social implications of AI most deeply — are increasingly choosing their employers based on organizational ethics culture. The engineer who cares about the social impact of her work will choose, when she has options, an organization where her concerns will be welcomed rather than one where they will be suppressed. Organizations that suppress dissent lose access to exactly the employees most likely to catch and flag ethical problems. This is a talent management consequence with direct governance implications.


Section 22.8: The Future of AI Whistleblowing

Legislative and Regulatory Developments

The legal landscape for AI ethics whistleblowing is evolving, driven by the growing recognition that existing federal protections leave significant gaps. Several legislative proposals have been introduced in Congress that would extend whistleblower protections specifically to AI ethics concerns, creating an analogue to the Dodd-Frank financial sector protections in the AI domain. None has yet been enacted, but the legislative attention signals growing recognition of the gap.

The SEC's expanded whistleblower program has seen increasing engagement from individuals raising concerns about technology companies' representations to investors about their AI capabilities and ethics practices. As AI claims become more central to technology company valuations, the potential for securities fraud relating to AI representations grows — and with it, the scope for Dodd-Frank whistleblower protections.

The EU Whistleblower Protection Directive, in combination with the GDPR and the AI Act, creates a more comprehensive protection framework for AI ethics whistleblowers in EU member states than anything currently available at the US federal level. As the EU AI Act's requirements for high-risk AI systems take full effect, the range of EU-law violations that can be the subject of protected disclosure expands accordingly.

Professional Association Mechanisms

Professional associations for technology and AI professionals — ACM, IEEE, and others — have codes of ethics that include obligations to report practices that endanger the public interest. These codes are largely aspirational; professional associations lack the enforcement mechanisms of licensed professions like law and medicine. But they provide a normative framework that supports the legitimacy of ethical dissent and, in some circumstances, may provide community support and credentialing to individuals who speak out.

The development of specific AI ethics professional standards — and potentially, the development of AI ethics as a genuinely licensed professional function — would create stronger mechanisms for collective professional dissent. A licensed AI ethics professional who is obligated by professional rules to report certain categories of harm would have both stronger normative backing and potentially stronger legal protection for their disclosures.

Collective Action and Organized Labor

The Google Walkout of 2018 demonstrated that collective action by technology workers is possible and can attract significant public attention. The walkout was not about AI ethics specifically, but it established that technology workers are willing to organize and act collectively when organizational behavior conflicts with their values.

The formation of the Alphabet Workers Union — a minority union at Google's parent company — and the growing interest in organized labor across the technology sector suggest that collective mechanisms for ethical dissent may become more significant. A union that includes AI ethics concerns in its advocacy agenda — that can raise concerns about harmful AI deployment as a collective labor matter rather than as individual employee dissent — would have significantly more structural protection than any individual whistleblower.

The Long-Term Trajectory

As AI systems become more consequential — as they make more high-stakes decisions about more people, in more domains, with larger effects — the importance of internal ethical dissent as a governance mechanism will grow. The employees of AI organizations are, in many cases, the only people with sufficient access to understand what those systems are doing and with sufficient technical expertise to evaluate whether it is ethically acceptable.

The question is not whether AI ethics dissent will occur — it will, because people of conscience will work in AI organizations and will encounter ethical problems. The question is whether the legal, organizational, and cultural conditions will exist to make that dissent effective rather than merely heroic. Effective dissent — dissent that changes organizational behavior rather than simply ending careers — requires legal protection, organizational structures that make internal escalation safe and effective, and cultural norms that treat ethics concern-raising as a valued professional contribution rather than a threat.

Building those conditions is not just an ethics project. It is a governance project, an organizational effectiveness project, and, ultimately, a project in the practical management of extraordinarily consequential technology.


Section 22.9: The Intersection of Dissent, Power, and Identity

Why the Pattern Is Not Random

Looking across the high-profile AI ethics dissent cases of the past decade, a pattern emerges that is too consistent to be incidental: the individuals who bear the highest personal costs for raising AI ethics concerns are disproportionately women, people of color, and members of other marginalized groups. Timnit Gebru and Margaret Mitchell. Sophie Zhang. The Google Walkout organizers who reported retaliation. This pattern demands examination, because it reveals something about the organizational dynamics of AI ethics dissent that goes beyond the general problem of career risk.

The explanation operates at several levels. First, the people who have the sharpest awareness of AI systems' discriminatory potential are often the people who have personal experience with discrimination. Researchers and practitioners from underrepresented groups — whose communities have historically been the subjects of discriminatory algorithms, invasive surveillance, and biased institutional decision-making — often bring to their work a personal knowledge of the stakes that their more privileged colleagues lack. This knowledge makes them more likely to see problems and more motivated to push for them to be addressed.

Second, within organizations, people from underrepresented groups often have less institutional protection than their more privileged colleagues. They may have fewer mentors in senior positions, less social capital with decision-making leaders, and fewer alternative opportunities that would reduce the career risk of speaking up. The intersection of ethical commitment and institutional vulnerability makes the cost of dissent higher for people who are already marginalized in their organizations.

Third, the specific framing of AI ethics concerns — as matters of fairness, bias, and discrimination — is territory that is politically charged within organizations. Concerns about AI discrimination are easily reframed, by those who wish to dismiss them, as political rather than technical — as activist concerns rather than engineering problems. This reframing is more easily applied to concerns raised by people whose identity positions them as advocates for their own communities, creating a specific additional barrier to having concerns taken seriously.

The governance implication is direct: AI ethics governance structures that do not address the intersection of dissent, power, and identity will systematically underserve the employees who are most likely to identify discrimination problems, and most likely to pay the highest personal cost for raising them. Organizations committed to genuine AI ethics governance must be explicit about this dynamic and must create structural protections that are calibrated to the specific vulnerabilities of the people most likely to raise the most important concerns.

Power, Access, and the Information Asymmetry

A related structural dynamic shapes the AI ethics dissent landscape in ways that are not always explicitly acknowledged: the individuals with the deepest knowledge of AI systems' ethical problems are often not the individuals with the highest organizational power to address them.

The data scientist who identifies a demographic performance gap in a model has the technical knowledge. The product manager who documents the gap between how a system is described in management presentations and how it actually behaves has the organizational knowledge. The junior researcher who notices that an analysis has been structured to find a particular result has the methodological knowledge. But these individuals typically do not have the authority to change the system, delay deployment, or insist on remediation. That authority sits with executives, product leaders, and governance bodies — people who are further from the technical detail and dependent on the people with knowledge to bring concerns to them.

This information asymmetry — where knowledge and power are separated by organizational hierarchy — is the structural reason why psychological safety and formal dissent channels matter so much. The people with power to address AI ethics problems depend on the people with knowledge of those problems communicating them accurately and completely. When the people with knowledge do not feel safe communicating, or when their communications are dismissed without genuine engagement, the organization's decision-makers are operating blind, making consequential decisions without the information they need to make them well.

The organizations that close this gap — that create genuine information flows from technical staff to decision-makers, that treat technical ethics concerns as governance-relevant inputs rather than organizational friction — have a significant governance advantage. They know what their AI systems are doing. Organizations that suppress these information flows may maintain the appearance of smooth operations right up to the moment their AI system's behavior becomes public in the worst possible way.


Section 22.10: Designing for Genuine Voice — A Practical Framework

The Four Conditions for Effective Internal Dissent

Drawing on the analysis throughout this chapter, four conditions are necessary for internal ethical dissent to function as an effective AI governance mechanism. Organizations that meet all four are equipped for genuine governance; those that fall short on any one of them face governance gaps that are difficult to compensate for with other mechanisms.

Condition 1: Safety. Employees must genuinely believe that raising ethics concerns will not harm their careers. This is not a matter of policy alone; it is a matter of organizational track record. Employees observe what happens to the people who raise concerns, and they make rational inferences about their own risk from those observations. Safety requires demonstrated protection — not just stated protection — for employees who raise concerns in good faith.

Condition 2: Access. Employees must have channels through which to raise concerns that reach appropriate decision-makers. Access means more than the existence of an ethics hotline; it means that reported concerns are investigated by parties with genuine independence and appropriate expertise, that investigation findings reach the people with authority to act on them, and that reporting employees receive meaningful feedback about what was found and what was done. Ethics hotlines that generate no visible response are not access; they are the appearance of access.

Condition 3: Competence. The people who receive and review ethics concerns must be competent to evaluate them. An HR generalist receiving a technical report about a demographic performance gap in a machine learning model is not equipped to assess it without technical support. An ethics committee composed of business executives reviewing a complaint about training data practices is not equipped to evaluate it without technical and domain expertise. Competence in the ethics review function is a governance prerequisite.

Condition 4: Authority. The governance bodies and individuals that receive ethics concerns must have genuine authority to act on them: to require remediation, to delay deployment, to escalate to the board, and to protect the employee who raised the concern from retaliation. Authority without competence is decision-making in the dark; competence without authority is analysis without consequence. Both are necessary.

The Role of External Accountability in Supporting Internal Voice

Internal dissent mechanisms are more effective when they exist in an ecosystem of external accountability. When employees know that credible external alternatives exist — regulatory complaint mechanisms, protected media disclosure, independent auditing — the organization's incentive to genuinely address internal concerns increases. If internal channels produce no response but external disclosure is easy, protected, and consequential, organizations have a strong reason to make internal channels work.

This is one reason why the development of robust external AI ethics accountability mechanisms — expanded regulatory whistleblower programs, mandatory independent AI auditing, stronger legal protection for AI ethics disclosures — supports rather than undermines internal ethics governance. Organizations are more likely to take internal concerns seriously when they understand that suppressing those concerns will not make them go away; it will redirect them to external channels that are harder to manage and more publicly damaging.

The governance ideal is an internal ethics culture that is so effective — where concerns are raised early, heard seriously, investigated competently, and addressed genuinely — that external disclosure is rarely necessary. That ideal is far from universally achieved. Closing the gap between the current state and the ideal requires simultaneously investing in internal mechanisms and strengthening external accountability, so that the internal investment is reinforced by appropriate external incentives.


Discussion Questions

  1. The chapter describes a "dissent spectrum" from informal concern-raising to external public disclosure. At what point on this spectrum do you believe employees acquire an ethical obligation to act — not merely an option to do so? What factors should determine where that threshold is set?

  2. Albert Bandura's moral disengagement mechanisms — moral justification, euphemistic labeling, displacement of responsibility, diffusion of responsibility — are presented as explanations for why engineers and other AI practitioners participate in harmful AI development without experiencing themselves as doing something wrong. Do you find this analysis convincing? What experiences from your own professional or personal life might illustrate these mechanisms?

  3. The legal protections available to AI ethics whistleblowers in the United States are described in this chapter as having significant gaps. Should Congress enact specific federal protection for AI ethics whistleblowers? What would such a law need to cover, and what legitimate arguments exist against it?

  4. Frances Haugen's approach to whistleblowing — carefully copying documents, going to the SEC before the press, working with media in a structured way — is described as highly sophisticated and effective. What does it reveal about the current state of whistleblower protection that effective whistleblowing requires this level of preparation and strategy? What would a whistleblowing environment look like in which such elaborate preparation was unnecessary?

  5. The chapter argues that organizations that welcome internal dissent are better positioned to avoid AI failures than organizations that suppress it. If this is true — and the business case is straightforward — why do so many organizations suppress dissent? What organizational dynamics override the risk management logic?

  6. As an executive or board member, how would you design an organization's ethics reporting infrastructure to make it genuinely useful — not just formally compliant? What specific elements would distinguish your design from the typical anonymous ethics hotline that most employees distrust?

  7. Timnit Gebru and Margaret Mitchell were both terminated in circumstances that many observers characterized as retaliation for their research and advocacy on AI ethics. Google disputed that characterization. How should external observers — regulators, journalists, prospective employees, prospective partners — evaluate these competing accounts? What evidence is relevant, and who bears the burden of establishing the true account?


Chapter 22 examines the human dimension of AI ethics governance — the employees who see problems that governance structures should catch, and the conditions under which they can or cannot act on what they see. Chapter 21 examined the structures; Chapter 22 examines the people who must make those structures work.