55 min read

> "Governance isn't bureaucracy. Governance is the immune system of an AI organization. Without it, the first infection becomes fatal."

Chapter 27: AI Governance Frameworks

"Governance isn't bureaucracy. Governance is the immune system of an AI organization. Without it, the first infection becomes fatal."

— Professor Diane Okonkwo


The Board Presentation

The Athena Retail Group boardroom on the thirty-second floor has floor-to-ceiling windows facing Lake Michigan. On most days, the view is the most commanding thing in the room. Today, it is not.

Ravi Mehta stands at the head of the table, one slide illuminated behind him. The slide contains a two-by-two matrix — risk on the vertical axis, oversight on the horizontal. The upper-right quadrant, labeled "High Risk / Low Oversight," is shaded red. Inside it, in plain text, is a single entry: HR Screening Model.

The board members — eleven of them, including three who joined specifically for this session — are silent. They have spent the last forty minutes hearing Ravi describe Athena's new AI Governance Framework. The framework is comprehensive: an AI Ethics Board, a four-tier risk classification system, a model registry, impact assessment templates, and an incident response protocol. It is the product of three months of intensive design work, informed by international standards, peer company benchmarks, and, most uncomfortably, the lessons of Athena's own HR screening crisis.

A board member — Patricia Huang, chair of the Risk Committee, former General Counsel at a Fortune 200 insurer — breaks the silence with the question Ravi has been preparing for.

"Is this going to slow down our AI projects?"

Ravi does not hesitate. "Some of them, yes. The ones that should be slowed down."

He advances to the next slide. It shows Athena's current AI project pipeline: forty-three models in various stages of development and deployment. Each is color-coded by the risk tier Ravi's framework would assign. Eight are green — low risk, internal analytics, documentation requirements only. Fourteen are yellow — medium risk, operational decisions, requiring peer review. Twelve are orange — high risk, customer-facing, requiring ethics board review. Nine are red — critical risk, involving hiring, pricing, or credit decisions, requiring external audit.

"If this framework had been in place eighteen months ago," Ravi says, pointing to the red quadrant, "the HR screening model would have been flagged as critical risk. It would have required an ethics board review before deployment, an impact assessment documenting potential fairness concerns, and an independent validation of its performance across demographic groups. We would have caught the age discrimination pattern before it affected a single candidate."

He pauses. "Governance doesn't prevent innovation. It prevents the kind of innovation that ends up in a lawsuit."

Patricia Huang looks at Grace Chen, Athena's CEO, seated at the far end of the table. Grace nods.

"I want to say something to the board," Grace says. "Publicly and on the record. We move at the speed of trust, not the speed of code. This governance framework is not a constraint on our AI ambitions. It is the foundation of them. I've asked Ravi to have it fully operational within ninety days, and I've authorized the budget for an external ethics advisor and the tooling to support the model registry. This has my full backing."

In the back of the room, watching via video conference, NK Adeyemi types a single note in her document: This is what leadership looks like when the stakes are real.

Tom Kowalski, who has been quietly reviewing the framework document on his laptop, writes in his paper notebook: Governance as competitive advantage? Need to think about this more.


What Is AI Governance?

Let us begin with a definition that we will build on throughout this chapter.

Definition: AI governance is the set of policies, processes, organizational structures, and accountability mechanisms that ensure an organization's use of artificial intelligence is safe, ethical, effective, and aligned with its values, legal obligations, and strategic objectives.

Notice what this definition includes and what it excludes. It includes policies (the rules), processes (how the rules are applied), organizational structures (who has authority), and accountability mechanisms (what happens when things go wrong). It does not specify a particular technology, a particular industry, or a particular moral framework. AI governance is the infrastructure — the plumbing and wiring — that makes responsible AI possible at scale.

This is an important distinction. Ethics tells you what is right. Governance tells you how to ensure that the right thing happens consistently, reliably, and across an organization of hundreds or thousands of people making AI-related decisions every day.

Why Organizations Need Formal AI Governance

Consider what happens in its absence. Without governance, AI development decisions are made locally — by individual data scientists, product managers, or business unit leaders — based on their own judgment, priorities, and risk tolerance. Some of these decisions will be excellent. Some will be catastrophic. And the organization will have no systematic way to distinguish between the two until after the consequences arrive.

The case for governance rests on five pillars:

1. Risk management. AI systems can cause harm — discriminatory decisions, privacy violations, safety failures, financial losses. Governance provides the structures to identify, assess, and mitigate these risks before they materialize. As we saw in Chapter 25, Athena's HR screening model discriminated against older applicants and candidates from non-traditional educational backgrounds. No one intended this outcome. It emerged from historical patterns in the training data that no one examined because no one was required to examine them.

2. Legal and regulatory compliance. The regulatory landscape for AI is evolving rapidly (we will map it in detail in Chapter 28). The EU AI Act, sector-specific regulations in financial services and healthcare, and emerging state and national laws around the world create compliance obligations that require organizational infrastructure — not just legal awareness. Governance ensures that compliance is systematic rather than ad hoc.

3. Stakeholder trust. Customers, employees, investors, regulators, and the public are increasingly demanding transparency about how organizations use AI. A 2024 Edelman Trust Barometer special report found that 63 percent of consumers said they would stop using a company's products if they learned the company was using AI irresponsibly. Governance provides the structures — model cards, impact assessments, audit reports — that demonstrate trustworthy AI practices.

4. Organizational alignment. In any large organization, dozens or hundreds of people may be building, deploying, or purchasing AI systems simultaneously. Without governance, these efforts proceed according to different standards, different risk tolerances, and different ethical frameworks. The result is inconsistency, duplication, and — as we explored in Chapter 22's discussion of shadow AI — invisible risk accumulation.

5. Sustainable innovation. This is the argument that surprises people most, and the one Ravi made to Athena's board. Governance does not impede innovation — it enables sustainable innovation. Organizations that move fast without governance eventually hit a wall: a bias scandal, a regulatory penalty, a data breach, a public backlash. The recovery costs — financial, reputational, and organizational — far exceed the overhead of governance. Companies that invest in governance can move fast and maintain the trust that allows them to continue moving fast.

Business Insight: A 2024 study by the AI governance platform Holistic AI found that organizations with formal AI governance frameworks deployed AI models to production 23 percent faster than those without. The reason: governance reduced rework. Models that went through risk assessment and review before deployment were less likely to fail in production, require emergency rollbacks, or trigger legal reviews. Governance is not the brake — ungoverned failure is the brake.

The Governance Gap

The need is clear. The reality is concerning.

According to Stanford HAI's 2025 AI Index Report, 72 percent of organizations surveyed had deployed AI in at least one business function, but only 35 percent had established any formal AI governance framework. Even among organizations that reported having governance, the depth and maturity varied dramatically — from a one-page acceptable use policy to a comprehensive governance operating model with dedicated staff.

This governance gap — the distance between AI deployment and AI oversight — represents one of the most significant organizational risks in the current business landscape. It is the equivalent of building a fleet of commercial aircraft and skipping the safety inspection regime. The planes might fly perfectly well for a while. But when one doesn't, the absence of governance transforms an incident into a catastrophe.

Caution

The governance gap is not evenly distributed. Regulated industries — financial services, healthcare, insurance — tend to have more mature AI governance, partly because existing regulatory frameworks (OCC SR 11-7, FDA guidance, Solvency II) provide templates. Unregulated industries — retail, media, technology — often have the least governance despite deploying AI most aggressively. If your organization is in an unregulated sector, the absence of external requirements is not evidence that governance is unnecessary. It is evidence that you need to build it yourself.

Several factors contribute to the gap:

  • Speed-to-market pressure. Teams building AI systems face intense pressure to deliver. Governance feels like friction. It requires documentation, review, approval — activities that do not directly advance the product timeline.
  • Diffuse responsibility. No one "owns" AI governance in most organizations. Is it the CTO's responsibility? The Chief Data Officer's? Legal? Compliance? Risk? When everyone is responsible, no one is responsible.
  • Expertise scarcity. Effective AI governance requires a rare combination of technical understanding, legal knowledge, ethical reasoning, and organizational design skill. Few individuals possess all four. Few teams are assembled with all four represented.
  • Absence of standards. Until recently, there were no widely accepted standards for AI governance. Organizations that wanted to build governance frameworks had to design them from scratch. This has changed — and the rest of this chapter will cover the standards and frameworks now available.

The NIST AI Risk Management Framework

The most influential AI governance framework for US-based organizations is the NIST AI Risk Management Framework (AI RMF), published by the National Institute of Standards and Technology in January 2023. The AI RMF is voluntary — it carries no regulatory force — but it has rapidly become the de facto reference standard for organizations building AI governance programs.

Definition: The NIST AI Risk Management Framework (AI RMF 1.0) is a voluntary framework designed to help organizations manage risks associated with AI systems throughout their lifecycle. It is structured around four core functions: Govern, Map, Measure, and Manage.

The Four Functions

The AI RMF is organized around four functions, each containing categories and subcategories that describe specific governance activities. Think of these as concentric rings of a governance system.

Govern is the outer ring — the organizational foundation. It addresses the policies, processes, accountability structures, and culture that enable AI risk management. Govern is not a one-time activity; it is the ongoing organizational commitment that makes the other three functions possible.

Key Govern activities include:

  • Establishing AI risk management policies and procedures
  • Defining roles, responsibilities, and accountability for AI governance
  • Allocating resources (budget, staff, tools) for AI risk management
  • Creating mechanisms for stakeholder input and feedback
  • Building a culture of responsible AI development
  • Ensuring diversity of perspectives in AI governance processes

Map is the process of understanding the context in which an AI system operates. Before you can manage risk, you need to understand what the system does, who it affects, what data it uses, and what could go wrong.

Key Map activities include:

  • Identifying and documenting the purpose and intended use of each AI system
  • Understanding the population affected by the AI system and how they are affected
  • Mapping the data used by the system, including its provenance, quality, and potential biases
  • Identifying the benefits, risks, and potential harms of the AI system
  • Understanding the broader social, legal, and regulatory context

Measure is the process of assessing and quantifying AI risks. This function connects technical metrics (model performance, fairness measures, robustness tests) to organizational risk categories.

Key Measure activities include:

  • Selecting and applying appropriate metrics for evaluating AI system performance
  • Assessing the AI system for bias, fairness, and equity across relevant demographic groups
  • Evaluating the robustness, reliability, and security of the AI system
  • Testing the AI system's behavior under adversarial conditions and edge cases
  • Documenting the results of assessments and tracking metrics over time

Manage is the process of responding to identified risks — allocating resources, implementing controls, and monitoring outcomes. This is where assessment translates into action.

Key Manage activities include:

  • Prioritizing AI risks based on severity, likelihood, and organizational impact
  • Implementing risk mitigation strategies and controls
  • Monitoring AI systems in production for performance degradation, bias drift, and emerging risks
  • Establishing incident response procedures for AI-related failures
  • Continuously improving AI risk management practices based on experience

Business Insight: The four functions of the AI RMF are not sequential steps. They are concurrent, iterative activities. You do not Govern, then Map, then Measure, then Manage in a linear waterfall. You Govern while you Map, Measure, and Manage. And insights from Mapping, Measuring, and Managing feed back into Governance — refining policies, reallocating resources, and updating accountability structures. This iterative nature is one of the framework's greatest strengths: it acknowledges that AI risk management is a continuous process, not a one-time compliance exercise.

Implementing the NIST AI RMF in Practice

The framework is deliberately flexible — it can be adapted to organizations of any size, in any industry, at any stage of AI maturity. But flexibility can be paralyzing. Here is a practical implementation approach:

Phase 1: Foundation (Months 1-3). Establish the Govern function. Appoint an AI governance lead or committee. Develop an initial AI risk management policy. Create a model inventory — a catalog of all AI systems currently in development or production. You cannot govern what you cannot see.

Phase 2: Assessment (Months 3-6). Apply the Map function to existing AI systems. Document the purpose, data sources, affected populations, and potential risks of each system. Prioritize systems by risk level. This is where many organizations discover AI systems they didn't know existed — the shadow AI problem from Chapter 22.

Phase 3: Metrics and Testing (Months 6-9). Implement the Measure function. Select fairness metrics, establish performance baselines, and conduct bias assessments for high-risk systems. Build or purchase the tooling needed for ongoing measurement.

Phase 4: Ongoing Management (Month 9+). Operationalize the Manage function. Implement monitoring for production systems. Establish incident response procedures. Begin the continuous cycle of assessment, mitigation, and improvement.

Athena Update: Ravi's governance framework at Athena is built on the NIST AI RMF skeleton, but customized for Athena's context. His Govern function includes the AI Ethics Board, risk tier definitions, and model registry. His Map function requires impact assessments for all models above the "low risk" tier. His Measure function mandates fairness testing across age, gender, and racial groups for all HR-related and customer-facing models. His Manage function includes quarterly model reviews, an incident response playbook, and a "governance hotline" for employees to report AI concerns anonymously.


ISO/IEC 42001: The AI Management System Standard

While the NIST AI RMF provides a risk-focused framework, ISO/IEC 42001 provides something different: a certifiable management system standard for AI. Published in December 2023, it is the first international standard specifically designed for AI management systems.

Definition: ISO/IEC 42001 is an international standard that specifies requirements for establishing, implementing, maintaining, and continually improving an Artificial Intelligence Management System (AIMS) within an organization. It is a certifiable standard — meaning organizations can be audited and certified as compliant by accredited certification bodies.

What Makes ISO/IEC 42001 Different

If you are familiar with other ISO management system standards — ISO 9001 (quality), ISO 27001 (information security), ISO 14001 (environmental management) — the structure of ISO/IEC 42001 will feel familiar. It follows the Harmonized Structure (formerly Annex SL) that all modern ISO management system standards share, which means it can be integrated with existing management systems rather than requiring a separate parallel structure.

The standard addresses several key areas:

Context of the organization. Understanding internal and external factors that affect AI management, identifying interested parties (stakeholders), and determining the scope of the AIMS.

Leadership and commitment. Top management must demonstrate commitment to the AIMS, establish an AI policy, and ensure roles and responsibilities are clearly defined. This is not a delegation-friendly requirement — the standard explicitly requires senior leadership involvement.

Planning. Identifying risks and opportunities associated with AI, setting objectives for AI management, and planning how to achieve those objectives.

Support. Ensuring adequate resources, competence, awareness, communication, and documented information for AI management.

Operation. Planning and controlling AI development, deployment, and use — including AI impact assessments, data management, and third-party considerations.

Performance evaluation. Monitoring, measuring, analyzing, and evaluating the AIMS — including internal audits and management reviews.

Improvement. Addressing nonconformities, implementing corrective actions, and continually improving the AIMS.

The Certification Process

Unlike the NIST AI RMF, which is voluntary and self-assessed, ISO/IEC 42001 can be independently certified. The certification process typically involves:

  1. Gap analysis. An assessment of the organization's current AI management practices against the standard's requirements.
  2. Implementation. Building or enhancing the AIMS to meet the standard's requirements. This typically takes 6-18 months, depending on organizational size and existing maturity.
  3. Internal audit. A self-assessment to verify readiness for external certification.
  4. Stage 1 audit. An external auditor reviews the AIMS documentation and assesses readiness for a full audit.
  5. Stage 2 audit. The certification audit — an external auditor evaluates the AIMS in practice, interviewing staff, reviewing records, and assessing the effectiveness of the system.
  6. Certification. If the audit is successful, the organization receives ISO/IEC 42001 certification, valid for three years with annual surveillance audits.

When ISO/IEC 42001 Makes Sense

Certification is not free — it requires significant investment in documentation, processes, training, and audit fees. It makes the most strategic sense for organizations that:

  • Operate in regulated industries where demonstrating AI governance maturity is a competitive or compliance advantage
  • Serve enterprise customers who require evidence of AI governance as a procurement condition
  • Want to differentiate their AI practices in markets where trust is a competitive factor
  • Operate internationally and need a governance framework recognized across jurisdictions
  • Already have ISO management systems (ISO 27001, ISO 9001) and can integrate AI management into existing structures

Business Insight: As of early 2026, ISO/IEC 42001 certification is still relatively rare — fewer than 500 organizations worldwide had achieved it. Early adopters are primarily AI vendors, cloud service providers, and companies in financial services and healthcare. But the trajectory mirrors the adoption of ISO 27001 (information security) in the 2000s and 2010s: initially a differentiator, eventually an expectation. Organizations that certify early gain both the internal benefits of a structured AIMS and the external benefits of market differentiation.


OECD AI Principles

Before diving into organizational implementation, it is worth understanding the policy foundation that many national AI governance frameworks build upon. The OECD Principles on Artificial Intelligence, adopted in May 2019, were the first intergovernmental standard on AI. They have since been endorsed by over 46 countries and influenced the development of the EU AI Act, the NIST AI RMF, and numerous national AI strategies.

The five OECD AI Principles are:

1. Inclusive Growth, Sustainable Development, and Well-Being

AI should benefit people and the planet. AI systems should be designed to augment human capabilities, reduce inequalities, and support sustainable development — not merely to maximize efficiency or profit.

In organizational practice: This principle challenges organizations to look beyond immediate business value and consider broader impact. Who benefits from this AI system? Who might be harmed? Are the benefits distributed equitably? Athena's governance framework operationalizes this principle through its impact assessment template, which requires project teams to identify both intended beneficiaries and potentially affected populations.

2. Human-Centered Values and Fairness

AI systems should respect human rights, democratic values, and diversity. They should include safeguards — such as human oversight — to ensure fairness and to prevent discrimination.

In organizational practice: This principle connects directly to the fairness and explainability work from Chapter 26. It requires organizations to define what fairness means in their context, implement metrics to measure it, and provide mechanisms for individuals to challenge AI-driven decisions that affect them. It also reinforces the Human-in-the-Loop theme: AI systems that make consequential decisions about people should include meaningful human oversight.

3. Transparency and Explainability

Organizations should be transparent about when and how AI is used. AI systems should be understandable — people affected by AI-driven decisions should be able to understand the basis for those decisions and to challenge them.

In organizational practice: This principle has operational implications at multiple levels. At the system level, it requires explainability mechanisms (SHAP, LIME, model cards — as covered in Chapter 26). At the organizational level, it requires disclosure policies — when must customers be told they are interacting with AI? At the governance level, it requires documentation — can the organization explain, retrospectively, why a particular AI-driven decision was made?

4. Robustness, Security, and Safety

AI systems should be robust, secure, and safe throughout their lifecycle. Organizations should be able to trace and reproduce AI outputs, and should have mechanisms to address risks as they emerge.

In organizational practice: This principle connects to model risk management (covered later in this chapter), cybersecurity practices (covered in Chapter 29), and operational monitoring. It requires ongoing vigilance — not just validating a model before deployment, but monitoring it continuously for performance degradation, data drift, and adversarial attacks.

5. Accountability

Organizations and individuals who develop, deploy, or operate AI systems should be accountable for their proper functioning. This includes being answerable for the outcomes of AI systems, maintaining documentation, and supporting redress mechanisms for those affected by AI-driven decisions.

In organizational practice: Accountability is perhaps the most challenging principle to operationalize because it requires clear lines of responsibility. Who is accountable when an AI system produces a discriminatory outcome — the data scientist who built it, the product manager who specified the requirements, the executive who approved the deployment, or the governance committee that reviewed it? Effective governance frameworks define these accountability lines explicitly, often through RACI matrices (covered later in this chapter).

Research Note: The OECD AI Policy Observatory (oecd.ai) maintains a comprehensive database of AI policies and initiatives from over 70 countries and territories. It is the best single resource for tracking how these principles are being translated into national policy — and an essential reference for organizations that operate internationally.


Ethics Committees and Review Boards

We now move from frameworks and principles to organizational structures. Perhaps no governance mechanism is more visible — or more frequently misunderstood — than the AI ethics committee.

Why Ethics Committees Exist

An ethics committee provides institutional oversight for AI-related decisions that involve significant ethical considerations. Its purpose is not to replace individual judgment, but to ensure that consequential decisions receive diverse perspectives, rigorous analysis, and organizational accountability.

The case for a dedicated committee (rather than relying on existing governance structures) rests on three observations:

  1. AI ethics decisions are cross-functional. They involve technical considerations (model behavior, data quality), legal considerations (regulatory compliance, liability), business considerations (market impact, customer trust), and ethical considerations (fairness, autonomy, harm). No single department has all four perspectives.

  2. AI ethics decisions have unique characteristics. They often involve probabilistic outcomes (the model will discriminate against some percentage of people, but we cannot predict which ones), emergent behaviors (the model may behave in ways not anticipated by its designers), and long time horizons (the effects of an AI-driven decision may not manifest for months or years).

  3. AI ethics decisions create organizational liability. An AI system that discriminates, violates privacy, or causes harm creates legal, financial, and reputational risk for the entire organization. A committee provides a formal mechanism for organizational due diligence — evidence that the organization took reasonable steps to identify and mitigate ethical risks.

Committee Composition

The composition of an AI ethics committee is critical. It must include sufficient diversity of perspective — technical, legal, ethical, business, and affected-community — to challenge assumptions and identify blind spots.

A well-composed AI ethics committee typically includes:

Role Perspective Why Essential
Chief Data/AI Officer (or equivalent) Technical leadership Can assess technical feasibility and limitations
General Counsel (or designee) Legal and regulatory Evaluates compliance risk and liability
Chief Human Resources Officer Employee impact Particularly important for workforce AI
Chief Risk Officer Enterprise risk Connects AI risk to broader risk framework
Business unit leader (rotating) Operational reality Ensures governance reflects business needs
External ethics advisor Independent perspective Provides objectivity and academic rigor
Employee representative Frontline perspective Represents the people most directly affected
Customer/community advocate External perspective Represents affected populations

Athena Update: Ravi designed Athena's AI Ethics Board with six permanent members: himself as chair (VP Data & AI), the CHRO, the General Counsel, the CTO, an external ethics advisor (Dr. Amina Fayad, a professor of technology ethics at Northwestern), and an employee representative selected through an internal nomination process. The employee representative position rotates annually. Business unit leaders attend when their projects are under review. This composition was deliberate — Ravi wanted enough diversity to challenge assumptions without making the committee so large that it became unwieldy.

Charter and Authority

An ethics committee without clear authority is a discussion group. The committee's charter should define:

Scope: Which AI systems fall under the committee's purview? All AI systems? Only high-risk systems? Only customer-facing systems? Defining scope prevents both mission creep (reviewing trivial systems) and gaps (missing critical ones).

Authority level: Can the committee approve, delay, or block AI deployments? Or is it advisory only? Most mature organizations give their ethics committees binding authority over high-risk and critical-risk systems, with advisory authority over medium-risk systems.

Trigger criteria: What triggers a committee review? Risk tier classification? A flagged concern from a team member? A new data source? A change in the regulatory environment? Clear triggers prevent both over-review (everything goes to committee) and under-review (nothing goes to committee).

Decision-making process: How does the committee reach decisions? Consensus? Majority vote? Chair's authority with committee input? What happens when the committee is split?

Escalation path: What happens when a project team disagrees with the committee's decision? There should be a clear escalation mechanism — typically to the CEO or a board-level committee — that provides an appeal path without undermining the committee's authority.

Documentation requirements: What records does the committee maintain? Meeting minutes, decision rationales, dissenting opinions, follow-up actions? These records serve both as institutional memory and as evidence of due diligence.

Meeting Cadence and Decision-Making

The operational rhythm of an ethics committee matters more than most organizations realize.

Regular review meetings (monthly or bi-monthly) handle the ongoing pipeline of AI projects — new deployments, significant model updates, changes to data sources, and periodic reviews of deployed systems.

Ad hoc reviews handle urgent situations — a discovered bias, a data incident, a regulatory change, or an employee concern. The committee should have a mechanism for convening quickly (within 48-72 hours) when urgency requires it.

Annual governance reviews assess the committee's own effectiveness. Is the composition still appropriate? Are the risk tier definitions working? Are project teams engaging with the process constructively? Is the committee catching the right things? What has it missed?

Caution

The most common failure mode for AI ethics committees is not opposition — it is irrelevance. If the committee meets too infrequently, reviews too slowly, or lacks binding authority, project teams will learn to work around it. The committee becomes a rubber stamp or, worse, a box that gets checked after deployment is already complete. To remain effective, the committee must be responsive (fast enough to keep pace with development timelines), authoritative (empowered to delay or block deployments), and embedded (integrated into the development workflow rather than bolted on at the end).


AI Impact Assessments

If the ethics committee provides the who of governance oversight, the AI impact assessment provides the how. An AI impact assessment is a structured evaluation of an AI system's potential effects — intended and unintended, positive and negative — on individuals, communities, and the organization.

When to Conduct an Impact Assessment

Not every AI system requires a full impact assessment. An internal dashboard that visualizes sales trends poses different risks than a model that determines which customers receive credit offers. The governance framework should define clear criteria for when an impact assessment is required.

A practical approach ties impact assessment requirements to risk tiers:

Risk Tier Examples Assessment Requirement
Low Internal analytics, reporting dashboards, data visualization Documentation only — a brief description of the system's purpose, data sources, and intended users
Medium Operational optimization, demand forecasting, inventory management Abbreviated assessment — a structured analysis of the system's purpose, data, potential biases, and mitigation measures
High Customer-facing recommendations, marketing personalization, pricing algorithms Full impact assessment — comprehensive analysis including stakeholder consultation, fairness testing, and ongoing monitoring plan
Critical Hiring, credit decisions, insurance underwriting, healthcare recommendations Full impact assessment plus independent review — all elements of the full assessment, plus validation by an independent party (internal audit, external auditor, or ethics committee)

The Impact Assessment Framework

A well-structured AI impact assessment addresses seven domains:

1. System description. What does the AI system do? What decisions does it inform or automate? What is its intended purpose, and what are its known limitations?

2. Data assessment. What data does the system use? Where does the data come from? How representative is it? What biases might exist in the data? Are there any gaps in coverage that could lead to disparate outcomes?

3. Stakeholder analysis. Who is affected by this AI system? This includes direct users, the people whose data is used, the people whose decisions are informed or replaced, and broader communities that may be affected. How have these stakeholders been consulted?

4. Fairness and equity. How does the system perform across different demographic groups? What fairness metrics have been applied? Are there disparate impact concerns? What mitigation measures have been implemented?

5. Transparency and explainability. Can the system's decisions be explained to affected individuals? What explainability mechanisms are in place? Is the use of AI disclosed to affected parties?

6. Risk analysis. What could go wrong? What is the severity of potential harms? What is the likelihood? What controls are in place to prevent or detect failures? This section should include both technical risks (model failure, data drift) and organizational risks (misuse, over-reliance, automation bias).

7. Monitoring and review. How will the system be monitored after deployment? What metrics will be tracked? How frequently will the system be reviewed? What are the criteria for taking the system offline?

Try It: Select an AI system your organization uses (or a hypothetical one relevant to your industry). Complete a brief impact assessment covering each of the seven domains above. You will likely find that domains 2 (Data Assessment) and 4 (Fairness and Equity) are the most difficult — which is precisely why they are the most important.

Risk Categorization

The impact assessment should produce a clear risk categorization that determines the level of oversight and approval required. A four-level categorization is common:

Low Risk. The AI system has minimal potential for individual or societal harm. Failure modes are well-understood and consequences are limited. Examples: internal reporting tools, data visualization aids, document summarization for internal use.

Medium Risk. The AI system affects operational decisions but does not directly determine outcomes for individuals. Failure could result in operational inefficiency or financial loss but not direct harm to individuals. Examples: demand forecasting, inventory optimization, anomaly detection in internal systems.

High Risk. The AI system directly affects individuals — customers, employees, or other stakeholders. Failure could result in discrimination, privacy violations, financial harm, or other consequential impacts on individuals. Examples: customer-facing recommendation engines, dynamic pricing, fraud detection systems.

Unacceptable Risk. The AI system poses risks that no mitigation can adequately address given the current state of technology, data, or organizational maturity. This is the category that results in a decision not to deploy. Examples might include a facial recognition system with known accuracy disparities across racial groups deployed in a high-stakes context, or a fully automated credit denial system with no human review mechanism.

Business Insight: The "unacceptable risk" category is the hardest to use because it requires saying no to something that may have significant business value. But it is also the most important. The willingness to classify a project as unacceptable risk — and to decline the revenue or efficiency it would generate — is the ultimate test of an organization's governance commitment. If the category exists but nothing ever lands in it, the governance framework is performative.


Model Risk Management

For organizations in financial services, model risk management is not new. The Office of the Comptroller of the Currency (OCC) published Supervisory Letter 11-7 (SR 11-7) in 2011, establishing comprehensive requirements for the management of model risk at banks. As AI models have become more prevalent and more consequential, the principles of SR 11-7 have become relevant far beyond banking.

Definition: Model risk is the potential for adverse consequences from decisions based on incorrect or misused model outputs. Model risk arises from two sources: the model itself (errors in design, implementation, or data) and the use of the model (misunderstanding its limitations, applying it outside its intended context, or over-relying on its outputs).

The Three Lines of Defense

SR 11-7 and its successors establish a "three lines of defense" approach to model risk management:

First line: Model development and use. The teams that build and use models are responsible for ensuring model quality — appropriate design, clean data, thorough testing, clear documentation, and appropriate use. This is not a governance function; it is a quality function embedded in the development process.

Second line: Model validation and oversight. An independent function — separate from the development team — validates models before deployment and periodically thereafter. Validation includes assessing model design, evaluating data quality, testing model performance, and reviewing documentation. Independence is critical: the validators must not report to the same management chain as the developers.

Third line: Internal audit. An independent audit function assesses the overall effectiveness of the model risk management framework — not individual models, but the system of governance itself. Are policies being followed? Are validations thorough? Are issues being tracked and resolved?

Applying Model Risk Management to AI

Traditional model risk management was designed for statistical and financial models — regression models, credit scoring models, pricing models. AI models introduce additional challenges:

Complexity. Many AI models — particularly deep learning models — are significantly more complex than traditional statistical models. This makes them harder to validate, harder to explain, and harder to monitor.

Data dependency. AI models are more sensitive to data quality and data drift than traditional models. A change in the distribution of input data can degrade model performance in ways that are difficult to detect without continuous monitoring.

Emergent behavior. AI models, particularly large language models and complex deep learning systems, can exhibit behaviors that were not anticipated or intended by their developers. This makes pre-deployment validation necessary but not sufficient — ongoing monitoring is essential.

Rapid iteration. AI development cycles are often faster than traditional model development cycles, creating tension between the speed of development and the thoroughness of validation.

Despite these challenges, the core principles of model risk management apply:

  • All production models should be inventoried in a model registry with metadata including purpose, owner, data sources, performance metrics, known limitations, and review dates.
  • High-risk models should be independently validated before deployment and periodically thereafter.
  • Model documentation should be thorough enough that a qualified person who was not involved in development could understand the model's design, data, performance, limitations, and intended use.
  • Ongoing monitoring should track model performance in production, detect data drift and performance degradation, and trigger re-validation when necessary.
  • Incident response should include procedures for responding to model failures, including the authority to take models offline when they pose unacceptable risk.

Research Note: The Federal Reserve's 2024 guidance on AI model risk management (SR 11-7 Supplemental Guidance) explicitly extended model risk management requirements to AI and machine learning models, including generative AI. The guidance emphasizes that AI models require more rigorous validation than traditional models, not less, precisely because of their complexity, data dependency, and potential for emergent behavior. While this guidance applies directly only to supervised financial institutions, its principles represent best practice for any organization using AI models in consequential decisions.


AI Policy Development

Governance frameworks and organizational structures provide the architecture. Policies provide the rules. An organization's AI policies translate its governance principles into specific, actionable requirements that guide day-to-day decisions.

The AI Policy Stack

Most organizations need several complementary policies rather than a single monolithic "AI policy." A typical policy stack includes:

1. AI Principles or Values Statement. A high-level statement of the organization's commitments regarding AI. This is a public-facing document that establishes intent: "We commit to developing and deploying AI that is fair, transparent, accountable, safe, and respectful of privacy." The principles statement is aspirational — it sets direction but does not specify requirements.

2. AI Acceptable Use Policy. A policy defining how employees may and may not use AI tools in their work. This addresses the shadow AI problem from Chapter 22 head-on: Which AI tools are approved for use? What data may and may not be entered into external AI systems? What types of outputs require human review before use? What are the consequences of policy violations?

3. AI Development Standards. Technical standards for teams building AI systems in-house. These cover data requirements (quality, provenance, consent), model development practices (testing, documentation, version control), fairness and bias testing requirements, and deployment criteria (what a model must demonstrate before it goes to production).

4. AI Procurement Standards. Requirements for purchasing AI systems from third parties. These address vendor due diligence (how the vendor manages data, bias, and security), contractual requirements (model transparency, audit rights, data ownership), and ongoing vendor monitoring.

5. AI Deployment Requirements. Requirements that apply at the point of deployment — the transition from development to production. These include sign-off requirements by risk tier, documentation requirements (model cards, data sheets), monitoring and alerting requirements, and rollback procedures.

6. AI Incident Response Policy. Procedures for responding to AI-related incidents — discovered biases, model failures, data breaches involving AI systems, or unintended consequences. This policy defines what constitutes an incident, who must be notified, what investigation is required, and what remediation is expected.

Try It: Draft a one-page AI Acceptable Use Policy for your organization (or a hypothetical one). At minimum, it should address: (1) Which AI tools are approved for use, (2) What data may not be entered into external AI systems, (3) What outputs require human review, and (4) Who to contact with questions. You will be surprised how many difficult questions this exercise surfaces.

Writing Effective AI Policies

AI policies that gather dust on a SharePoint site are worse than no policies at all — they create the illusion of governance without the reality. Effective policies share several characteristics:

Clarity. Policies should be written in plain language that any employee can understand. If a data scientist needs a law degree to interpret your AI development standards, the standards will not be followed.

Specificity. "We are committed to fairness" is a value, not a policy. "All customer-facing models must be tested for disparate impact across protected characteristics before deployment" is a policy. The difference is testability — can you determine, objectively, whether the policy has been followed?

Proportionality. The burden of compliance should be proportional to the risk. Requiring a full ethics board review for an internal reporting dashboard wastes everyone's time and erodes trust in the governance process. Requiring only documentation for a hiring algorithm is dangerous negligence. The risk tier system ensures that the governance burden matches the governance need.

Enforceability. A policy without consequences is a suggestion. Enforcement mechanisms can range from gentle (training and coaching for first violations) to severe (project suspension, performance impact), but they must exist. More importantly, the organization must be willing to enforce them — including against senior leaders and high-priority projects.

Adaptability. AI technology and regulation evolve rapidly. Policies should include review dates (typically annual) and mechanisms for updating them as the landscape changes. A policy that was appropriate in 2024 may be insufficient in 2026.


Governance Operating Models

How an organization structures its AI governance function determines how effectively governance translates from policy to practice. Three primary operating models exist, each with distinct advantages and limitations.

Centralized Governance

In a centralized model, a single governance function — often reporting to the CDO, CTO, or Chief Risk Officer — has authority over all AI governance activities across the organization. All AI impact assessments, risk classifications, policy decisions, and compliance monitoring flow through this central function.

Advantages: - Consistent standards across the entire organization - Clear accountability — one person or team owns AI governance - Efficient use of specialized governance expertise - Easier to maintain a comprehensive model registry - Stronger enforcement capability

Disadvantages: - Can become a bottleneck, particularly in large or fast-moving organizations - May lack deep understanding of business unit-specific contexts and needs - Risk of being perceived as a "police function" rather than a partner - Difficult to scale as the volume of AI systems grows

Best suited for: Small to medium organizations, organizations early in their AI governance journey, and organizations where AI use is concentrated in a few functions.

Federated Governance

In a federated model, each business unit or function has its own AI governance capability, operating within a common set of policies and standards set at the organizational level. A central function sets the rules; the business units implement them.

Advantages: - Governance is closer to the work — business unit governance teams understand the context - Faster review and approval cycles — no central bottleneck - Greater ownership by business units — governance feels less like an external imposition - Scales more naturally with organizational growth

Disadvantages: - Risk of inconsistency across business units - Requires significant investment in governance expertise across multiple teams - Harder to maintain a unified model registry - Potential for "governance shopping" — teams routing projects through the most permissive business unit

Best suited for: Large, diversified organizations where AI use is widespread and business units have significant autonomy.

Hybrid Governance

Most mature organizations converge on a hybrid model that combines elements of centralized and federated approaches. A central governance function (often called an AI Center of Excellence, AI Governance Office, or Responsible AI Office) sets policies, maintains the model registry, provides tools and templates, and handles escalations. Business units have embedded governance champions or liaisons who conduct initial assessments, facilitate reviews, and serve as the first point of contact for governance questions.

The hybrid model typically works as follows:

  • Central function responsibilities: Policy development, standard setting, model registry maintenance, ethics committee support, training and education, regulatory monitoring, reporting to senior leadership
  • Business unit responsibilities: Initial risk classification of AI projects, conducting impact assessments using centralized templates, ensuring compliance with development standards, flagging concerns and escalating to the central function
  • Shared responsibilities: Model validation (jointly conducted by central and business unit teams), incident response (led by central function with business unit participation), continuous improvement of governance practices

Athena Update: Athena adopted a hybrid model. Ravi's AI Governance Office (three people: a governance manager, a model risk analyst, and an administrative coordinator) sets policies, maintains the model registry, and staffs the AI Ethics Board. Each business unit that uses AI — marketing, supply chain, e-commerce, HR — has designated an AI governance liaison. These liaisons are not full-time governance roles; they are data scientists or product managers who serve as the governance point of contact for their teams, conduct initial risk classifications, and escalate to Ravi's office when needed. The structure is lean by design. "We don't need fifty people doing governance," Ravi told Grace Chen. "We need the right five people in the right places."

RACI Matrices for AI Governance

Regardless of the operating model, clear role definitions are essential. A RACI matrix — Responsible, Accountable, Consulted, Informed — provides a structured way to define who does what in AI governance.

Here is a simplified RACI matrix for key AI governance activities:

Activity Data Science Team AI Governance Office Ethics Committee Business Unit Leader Legal Executive Sponsor
Risk classification R A I C I I
Impact assessment R A C C C I
Fairness testing R A I I I I
Ethics review (high/critical) C R A C C I
Model deployment approval (low/med) R A I I I I
Model deployment approval (high/critical) C R A C C A
Incident response R A C C R I
Policy development C R C C R A
Annual governance review C R A C C I

R = Responsible (does the work), A = Accountable (owns the outcome), C = Consulted (provides input), I = Informed (kept aware)

Business Insight: The most important cell in the RACI matrix is not the "R" — it is the "A." Accountability without responsibility can work (the accountable person delegates the work). Responsibility without accountability does not (the person doing the work has no authority to ensure the outcome). For AI governance, the accountable party must have sufficient authority to enforce governance requirements — including the authority to delay or block deployments when governance criteria are not met.


Monitoring and Compliance

Governance is not a checkpoint — it is a continuous process. Deploying a model through a governance framework does not make it permanently compliant. Models degrade, data shifts, regulations change, and the world moves on. Ongoing monitoring and compliance are the mechanisms that ensure governance remains effective after the initial deployment decision.

Ongoing Oversight

Production AI systems require continuous monitoring across several dimensions:

Performance monitoring. Is the model still performing as expected? Accuracy, precision, recall, and other performance metrics should be tracked over time and compared to baseline values established during development. Significant degradation triggers re-evaluation.

Fairness monitoring. Are the model's outcomes still fair across demographic groups? Fairness metrics should be tracked over time and compared to baseline values. Bias can emerge after deployment — even in a model that was fair at the time of deployment — if the input data distribution changes.

Data drift detection. Has the distribution of input data changed significantly from the data the model was trained on? Statistical drift detection methods (Kolmogorov-Smirnov tests, Population Stability Index, Jensen-Shannon divergence) can identify when the model's training data no longer represents the real world.

Usage monitoring. Is the model being used as intended? Models designed for one purpose are sometimes repurposed for another — a recommendation engine designed for product discovery might be used for pricing decisions, for example. Usage monitoring ensures that models remain within their approved scope.

Compliance monitoring. Are governance policies and procedures being followed? This includes verifying that model documentation is up to date, that required reviews have occurred, and that incident response procedures are followed when issues arise.

Audit Trails

Governance is only as good as its documentation. Audit trails provide the evidence that governance activities occurred, that they were thorough, and that their conclusions were reasonable. Effective audit trails include:

  • Model registry records — a catalog of all production models with metadata including purpose, owner, risk tier, data sources, performance metrics, known limitations, deployment date, and review schedule
  • Impact assessment records — completed assessments for all models above the documentation-only tier
  • Review and approval records — evidence that required reviews occurred, including committee meeting minutes, decision rationales, and any conditions or restrictions imposed
  • Monitoring records — time-series data on model performance, fairness metrics, and data drift indicators
  • Incident records — documentation of AI-related incidents, investigations, root causes, remediation actions, and follow-up verification
  • Policy version history — a record of policy changes, including the rationale for each change and the approval process followed

Incident Reporting and Remediation

Even the best governance framework cannot prevent all AI-related incidents. What it can do is ensure that incidents are detected quickly, reported consistently, investigated thoroughly, and remediated effectively.

An AI incident response process should define:

What constitutes an incident. Not every model hiccup is an incident. An incident is an AI system behavior that causes or could cause significant harm — discriminatory outcomes, privacy violations, safety failures, financial losses above a defined threshold, or violations of law or regulation.

Reporting mechanisms. How are incidents reported? Who can report them? (Answer: anyone in the organization.) Is there an anonymous reporting channel? How quickly must incidents be reported after detection?

Investigation procedures. Who investigates? What is investigated? Technical root cause analysis should be complemented by organizational analysis — not just "what went wrong in the model" but "what went wrong in the governance process that allowed the model to cause harm."

Remediation requirements. What actions are required in response to a confirmed incident? These may include taking the model offline, notifying affected individuals, implementing technical fixes, updating governance policies, and reporting to regulators (if required by law).

Learning mechanisms. How does the organization learn from incidents? Post-incident reviews (sometimes called "AI post-mortems" or "learning reviews") should identify root causes, contributing factors, and systemic improvements that reduce the likelihood of recurrence. These reviews should be blameless — focused on system improvement rather than individual punishment — to encourage reporting and honest analysis.

Athena Update: Athena's incident response playbook defines four severity levels for AI incidents. Level 1 (Minor): a model performance issue with no individual impact — addressed within the data science team. Level 2 (Moderate): a model behavior that has or could have adverse effects on individuals — reported to the AI Governance Office and investigated within five business days. Level 3 (Significant): a confirmed bias, discrimination, privacy violation, or significant financial impact — reported to the AI Ethics Board and the General Counsel, with the model suspended pending investigation. Level 4 (Critical): an incident that requires external notification (regulatory, customer, or public) — escalated immediately to the CEO and the Board Risk Committee. The HR screening bias incident from Chapter 25 was classified as Level 4.


Building a Governance Culture

We have covered the structures: frameworks, standards, committees, policies, operating models, and monitoring systems. But structures alone do not create governance. A governance framework that is technically perfect but culturally rejected is worthless. The final — and arguably most important — component of AI governance is culture.

Beyond Compliance

There is a crucial distinction between compliance and culture. Compliance means following the rules because the rules exist. Culture means wanting to do the right thing because the organization values doing the right thing.

Compliance-driven organizations check boxes. They conduct impact assessments because they are required to. They submit models for review because the process demands it. They document their work because the policy says they must. When the boxes are checked, compliance is complete.

Culture-driven organizations ask questions that go beyond the boxes. "We've met the fairness threshold — but is this system actually fair for the people it affects?" "The impact assessment didn't flag any issues — but are we asking the right questions?" "The policy doesn't require review for this model — but should we review it anyway, given the context?"

The difference is the difference between governance as a constraint and governance as a value.

How to Build a Governance Culture

Building a governance culture requires deliberate, sustained investment across several dimensions:

Tone from the top. Senior leaders — the CEO, the board, the C-suite — must visibly champion AI governance. This means more than approving the budget. It means talking about governance in earnings calls, all-hands meetings, and strategic planning sessions. It means asking about governance in project reviews. It means celebrating governance successes (not just AI deployment successes). Grace Chen's public backing of Ravi's governance framework at the board meeting sent a signal that echoed through Athena's organization far more effectively than any policy document.

Training and education. Every person who builds, deploys, or uses AI should understand the governance framework and their role within it. This is not a one-time onboarding session; it is an ongoing education program that evolves as the framework matures and as new challenges emerge. Training should be practical, not theoretical — focused on scenarios that employees actually encounter.

Incentives and recognition. People do what they are rewarded for. If data scientists are rewarded only for deploying models quickly and governance is seen as a tax on their time, governance will be grudgingly tolerated at best. Organizations that recognize and reward governance contributions — including governance in performance reviews, celebrating teams that identify and mitigate risks, and elevating governance work as "real work" rather than overhead — create a culture where governance is valued.

Psychological safety. People must feel safe raising concerns about AI systems without fear of retaliation or career consequences. This is particularly important for junior employees — the data scientist who notices a concerning pattern in model outputs, the customer service representative who observes AI-driven recommendations that seem inappropriate, the HR coordinator who questions an AI-assisted hiring decision. If these people are afraid to speak up, the governance framework loses its most valuable sensors.

Embedding in workflow. Governance should be woven into the development process, not bolted on at the end. Impact assessments should begin at project inception, not at deployment. Risk classification should occur during planning, not during review. Fairness testing should be part of the development pipeline, not a final gate. When governance is embedded in workflow, it feels like part of the work rather than an interruption of it.

Continuous improvement. A governance culture treats the governance framework itself as a product — one that requires iteration, feedback, and improvement. Regular retrospectives ("What is working in our governance process? What is not? What should we change?") keep the framework relevant and effective.

Caution

Cultural transformation is slow. Expect resistance, particularly from experienced technical teams who have been building and deploying AI without governance oversight. Some resistance is legitimate — governance processes that are unnecessarily burdensome should be streamlined. Some resistance is not — the belief that governance is for lesser mortals who cannot be trusted with their own judgment. Distinguishing between the two requires listening, adaptation, and firmness in the right measure.

Tom's Shift

Tom Kowalski's reaction to AI governance is illustrative of a pattern Ravi has seen repeatedly.

In the first weeks of the governance framework rollout, Tom is skeptical. "This is going to slow everything down," he tells NK over coffee. "I've shipped production models at my last company in weeks. Now I need to fill out an impact assessment form?"

NK, who has been thinking about the HR screening crisis from Chapter 25, pushes back. "The model your last company shipped in weeks — did anyone check whether it was fair? Did anyone even define what 'fair' meant for that use case?"

Tom pauses. He wants to say yes, but he cannot. At his previous fintech startup, model deployment was fast, informal, and driven by a small team that trusted each other's judgment. The models were good — technically good. But no one ever asked whether they were good for everyone they affected.

"No," he admits. "We didn't."

"That's the point," NK says. "Speed without governance isn't velocity. It's recklessness that hasn't caught up with you yet."

By the end of the semester, Tom has internalized a reframing: governance is not the opposite of speed. Governance is the foundation of sustainable speed. It is what allows you to move quickly without breaking things that cannot be repaired.

Professor Okonkwo, overhearing their conversation, adds the observation that will stay with both of them: "Every high-performing system needs a governor — a regulating mechanism that prevents it from destroying itself. An engine needs a governor. A democracy needs a constitution. An AI organization needs a governance framework. The governor isn't there to slow you down. It's there to keep you running."


Governance in Global Context

AI governance does not operate in a vacuum. It exists within a broader ecosystem of national regulations, international standards, industry practices, and cultural expectations. While Chapter 28 will cover the regulatory landscape in detail, a few observations about governance in global context are essential here.

Regulatory convergence and divergence. Major regulatory frameworks — the EU AI Act, the NIST AI RMF, China's AI regulations, Singapore's Model AI Governance Framework — share common themes: risk-based classification, transparency requirements, human oversight, and accountability. But they diverge on specifics: what counts as "high risk," what transparency requires, and how accountability is enforced. Organizations operating internationally need governance frameworks flexible enough to accommodate these differences.

Cultural variation. Privacy expectations, attitudes toward automation, and the relative weight given to innovation versus precaution vary significantly across cultures and jurisdictions. A governance framework that works in San Francisco may not work in Stockholm, Tokyo, or Lagos. Effective global governance requires cultural sensitivity and local adaptation.

Regulatory trajectory. The direction of travel is clear: toward more regulation, not less. Organizations that build governance frameworks now are investing in infrastructure they will need. Organizations that wait for regulation to force governance will find themselves building emergency governance under regulatory pressure — a significantly more expensive and disruptive proposition.

Business Insight: Lena Park, the AI policy advisor who has been advising Athena on regulatory alignment, frames the strategic calculation this way: "You can design your governance framework proactively, on your own timeline, reflecting your own values and business context. Or you can wait until a regulator tells you what your governance framework must look like. Option A costs less, gives you more flexibility, and produces a better outcome. Option B is more common."


Putting It All Together: Athena's AI Governance Framework

Let us examine Athena's complete governance framework — the one Ravi presented to the board — as an integrated example of the concepts in this chapter.

Organizational Structure

AI Ethics Board: Six permanent members (VP Data & AI as chair, CHRO, General Counsel, CTO, external ethics advisor, employee representative) plus rotating business unit leaders. Meets monthly for regular reviews, with ad hoc meetings as needed. Binding authority over high-risk and critical-risk deployments.

AI Governance Office: Three staff (governance manager, model risk analyst, administrative coordinator) reporting to the VP Data & AI. Responsible for policy development, model registry maintenance, impact assessment review, monitoring coordination, and ethics board support.

Business Unit Liaisons: One designated liaison per AI-using business unit. Responsible for initial risk classification, impact assessment facilitation, and first-point-of-contact for governance questions.

Risk Classification

Tier Criteria Examples at Athena Governance Requirements
Low Internal use only, no individual impact Sales reporting dashboard, inventory analytics Documentation only
Medium Operational decisions, indirect individual impact Demand forecasting, supply chain optimization Abbreviated impact assessment, peer review
High Customer-facing, direct individual impact Product recommendations, marketing personalization, dynamic pricing Full impact assessment, ethics board review
Critical Employment, credit, or safety decisions Hiring screening, employee performance scoring, customer creditworthiness Full impact assessment, ethics board review, independent external audit

Policy Stack

  • Athena AI Principles: Six commitments — fairness, transparency, accountability, privacy, safety, and human oversight
  • Acceptable Use Policy: Approved AI tools, data handling requirements, human review mandates
  • Development Standards: Data quality requirements, fairness testing protocols, documentation templates, version control mandates
  • Procurement Standards: Vendor due diligence checklist, contractual AI transparency requirements
  • Deployment Requirements: Sign-off matrices by risk tier, model card requirements, monitoring dashboards
  • Incident Response Policy: Four-tier severity classification, reporting timelines, investigation procedures, escalation paths

Model Registry

All production AI models cataloged with: - Model name and unique identifier - Business purpose and intended use - Model owner and development team - Data sources and data lineage - Risk tier classification - Performance metrics (baseline and current) - Fairness metrics by demographic group - Known limitations and failure modes - Deployment date and last review date - Scheduled next review date

Monitoring and Compliance

  • Automated performance and fairness dashboards for all high-risk and critical-risk models
  • Weekly automated drift detection reports
  • Monthly model performance summaries to the AI Governance Office
  • Quarterly ethics board reviews of all critical-risk models
  • Annual comprehensive governance review

Incident Response

  • Anonymous "governance hotline" for reporting concerns
  • Four-tier severity classification (Minor, Moderate, Significant, Critical)
  • Defined response timelines by severity
  • Post-incident learning reviews for all Level 3 and Level 4 incidents
  • Quarterly incident trend reports to the AI Ethics Board and the executive team

Athena Update: Three months after the board presentation, Ravi reports initial results. The model registry has cataloged thirty-eight of forty-three known production models (five legacy models proved harder to document than expected). Two models have been reclassified from medium to high risk following impact assessments that revealed unanticipated customer impact. One model — a customer churn prediction used to determine service level priority — was voluntarily taken offline by the customer service business unit after an impact assessment revealed it systematically deprioritized customers in lower-income zip codes. "That model was optimizing exactly what we told it to optimize," the business unit leader told Ravi. "The governance process helped us realize we were optimizing the wrong thing." Ravi considered that statement the strongest validation of the framework's value.


Common Governance Pitfalls

Before closing, it is worth cataloging the failure modes that undermine even well-designed governance frameworks:

1. Governance theater. All the structures exist — committees, policies, forms — but they are performative. Reviews are rubber stamps. Impact assessments are copy-pasted from templates. The ethics committee meets but never says no. The organization can point to governance artifacts, but the artifacts do not change behavior.

2. One-size-fits-all. The governance framework applies the same requirements to every AI system regardless of risk. Internal analytics dashboards require the same impact assessment as hiring algorithms. The result is governance fatigue — teams spend so much time on low-value governance activities that they have no energy left for the high-value ones.

3. The last mile problem. Governance requirements are defined at the corporate level but not implemented at the team level. Policies exist but nobody reads them. Training is offered but nobody attends. The gap between governance on paper and governance in practice widens until it is indistinguishable from no governance at all.

4. Innovation antagonism. The governance function sees itself as a check on irresponsible innovation rather than an enabler of responsible innovation. Every interaction between governance and development teams is adversarial. Data scientists view governance as the enemy. Governance professionals view data scientists as reckless. Neither perspective is productive.

5. Static governance. The framework is designed, approved, and never updated. Meanwhile, technology evolves, regulations change, the organization's AI portfolio expands, and new risks emerge. A governance framework that was appropriate for ten models may be inadequate for a hundred.

6. Under-resourcing. The organization commits to governance in principle but does not commit the resources — budget, staff, tools — to make governance work in practice. A governance manager who is expected to oversee hundreds of models with no analytical support and no technology platform is being set up for failure.


Chapter Summary

This chapter has covered the organizational infrastructure for responsible AI:

  1. AI governance is the set of policies, processes, structures, and accountability mechanisms that ensure AI is used safely, ethically, and effectively. It is not bureaucracy — it is the immune system that protects the organization from AI-related harm.

  2. The governance gap — the distance between AI deployment and AI oversight — is one of the most significant organizational risks in the current business landscape. Closing it requires deliberate investment in governance infrastructure.

  3. The NIST AI Risk Management Framework provides a comprehensive, voluntary framework organized around four functions: Govern, Map, Measure, and Manage. These functions are iterative, not sequential.

  4. ISO/IEC 42001 provides a certifiable management system standard for AI, following the same structure as other ISO management system standards and enabling integration with existing quality and security management systems.

  5. The OECD AI Principles — inclusive growth, human-centered values, transparency, robustness, and accountability — provide the policy foundation that many national governance frameworks build upon.

  6. Ethics committees provide cross-functional oversight for consequential AI decisions. Their effectiveness depends on composition, authority, responsiveness, and integration into the development workflow.

  7. AI impact assessments are structured evaluations of AI systems' potential effects. Tied to risk tiers, they ensure that oversight is proportional to risk — lightweight for internal analytics, comprehensive for hiring and credit decisions.

  8. Model risk management applies the three lines of defense (development, validation, audit) to AI systems. The principles of SR 11-7, originally designed for financial models, are increasingly relevant across industries.

  9. AI policies translate governance principles into actionable requirements. An effective policy stack includes principles, acceptable use, development standards, procurement standards, deployment requirements, and incident response.

  10. Governance operating models — centralized, federated, and hybrid — determine how governance authority is distributed across the organization. Most mature organizations converge on hybrid models.

  11. Monitoring and compliance ensure that governance remains effective after deployment through performance monitoring, fairness monitoring, drift detection, audit trails, and incident response.

  12. Governance culture — built through leadership commitment, training, incentives, psychological safety, and workflow embedding — is what transforms governance from a compliance exercise into an organizational value.


Next chapter: Chapter 28: AI Regulation — Global Landscape, where we will map the regulatory terrain that organizations must navigate — the EU AI Act's risk tiers, the US sector-specific approach, China's AI regulations, and the compliance strategies that connect governance frameworks to legal requirements. Lena Park takes center stage as she guides Athena through the regulatory implications of its expanding AI portfolio.