Appendix B: Templates and Worksheets

This appendix provides fifteen ready-to-use templates and worksheets designed for immediate application in AI and machine learning initiatives. Each template has been refined through real-world deployment across industries ranging from financial services to healthcare to retail. You can copy, adapt, and extend every template to fit your organizational context.

How to use this appendix. Each template follows a consistent structure: a brief description of its purpose, instructions for effective use, the template itself in a format you can reproduce in any document or spreadsheet tool, and a short example showing how a completed version might look. Templates are sequenced to mirror the lifecycle of an AI initiative --- from strategic framing and project proposal through execution, governance, and incident response.

A note on customization. No template is one-size-fits-all. Where you see bracketed placeholders like [Company Name] or [Date], replace them with your own details. Where you see scoring scales, adjust the weights and criteria to reflect your organization's priorities. The goal is not rigid adherence to a form but disciplined thinking about the right questions.


Template 1: AI Project Proposal

Purpose

This one-page template provides a structured format for pitching an AI initiative to leadership. It forces proposers to articulate the business problem, data requirements, approach, resource needs, and risk profile in a concise format that decision-makers can evaluate quickly. Use it at the ideation stage to secure funding and sponsorship for exploratory or production AI projects.

Instructions for Use

  1. Complete every section. If you cannot fill in a section, that is a signal the project needs more definition before it is ready for proposal.
  2. Keep the entire proposal to a single page (or two pages maximum for complex initiatives). Brevity forces clarity.
  3. Attach supplementary detail (data dictionaries, technical architecture diagrams, vendor quotes) as appendices rather than embedding them in the proposal body.
  4. Circulate the draft to at least one technical stakeholder and one business stakeholder before submitting. Proposals that reflect only one perspective rarely survive scrutiny.
  5. Revisit and update the proposal quarterly if the project is approved. It should remain a living reference point.

Template


AI PROJECT PROPOSAL

Field Details
Project Title [Descriptive name]
Proposer [Name, title, department]
Date [Submission date]
Executive Sponsor [Name, title]
Version [e.g., 1.0]

1. Problem Statement

Describe the business problem or opportunity in 2--3 sentences. What is the current state? What pain does it cause? Who is affected?

[Write here]

2. Proposed AI/ML Approach

What type of AI/ML solution do you envision? (e.g., classification model, NLP pipeline, recommendation engine, generative AI application, computer vision system). Why is AI the right approach rather than a rules-based or manual solution?

[Write here]

3. Data Requirements

Data Source Owner Volume Quality Assessment Access Status
[Source 1] [Team/person] [Rows/GB] [High/Medium/Low] [Available/Restricted/TBD]
[Source 2] [Team/person] [Rows/GB] [High/Medium/Low] [Available/Restricted/TBD]
[Source 3] [Team/person] [Rows/GB] [High/Medium/Low] [Available/Restricted/TBD]

4. Team and Resources

Role Person/Status Time Commitment
Project Lead [Name or "To hire"] [% FTE]
Data Scientist [Name or "To hire"] [% FTE]
Data Engineer [Name or "To hire"] [% FTE]
Domain Expert [Name or "To hire"] [% FTE]
ML Engineer [Name or "To hire"] [% FTE]

5. Timeline

Phase Duration Key Deliverable
Discovery and Data Assessment [X weeks] Data readiness report
Proof of Concept [X weeks] Working prototype with baseline metrics
Development and Testing [X weeks] Production-ready model
Deployment and Integration [X weeks] Live system with monitoring
Evaluation and Optimization [X weeks] Performance report and iteration plan

6. Budget Estimate

Category One-Time Cost Annual Recurring Cost
Personnel [$]` | `[$]
Cloud / Compute [$]` | `[$]
Software / Licensing [$]` | `[$]
Data Acquisition [$]` | `[$]
Training and Change Management [$]` | `[$]
Total [$]`** | **`[$]

7. Success Metrics

Metric Current Baseline Target Measurement Method
[Business KPI 1] [Value] [Value] [How measured]
[Business KPI 2] [Value] [Value] [How measured]
[Technical Metric] [Value] [Value] [How measured]

8. Risks and Mitigations

Risk Likelihood Impact Mitigation
[Risk 1] [H/M/L] [H/M/L] [Action]
[Risk 2] [H/M/L] [H/M/L] [Action]
[Risk 3] [H/M/L] [H/M/L] [Action]

Example (Completed)

Project Title: Customer Churn Prediction for Enterprise Accounts Proposer: Maria Chen, Director of Customer Success Problem Statement: Enterprise accounts represent 68% of annual recurring revenue but churn has increased from 8% to 14% over the past 18 months. The customer success team currently relies on gut instinct and lagging indicators (support tickets, renewal date proximity) to identify at-risk accounts. By the time warning signs are visible, it is often too late to intervene effectively. Proposed Approach: A gradient-boosted classification model trained on 4 years of customer behavior data to predict churn probability 90 days in advance, enabling proactive intervention by account managers. Budget: $285K one-time, $95K annual recurring. Target: Reduce enterprise churn from 14% to 9% within 12 months of deployment, yielding an estimated $4.2M in retained ARR.


Template 2: AI Use Case Prioritization Matrix

Purpose

When an organization identifies multiple potential AI use cases --- and most do --- this scoring framework helps leadership compare them systematically rather than defaulting to the loudest advocate or the shiniest technology. The matrix evaluates each use case across five weighted criteria and produces a composite score that enables rank-ordering and portfolio-level decision-making.

Instructions for Use

  1. Assemble a cross-functional scoring team that includes business, technical, and risk perspectives. Individual scoring introduces bias; group calibration produces more reliable rankings.
  2. Before scoring, align the team on definitions. What constitutes "high" business impact in your organization? What does "data readiness" mean specifically? Use the scoring guide below as a starting point and customize it.
  3. Adjust the default weights to reflect your organization's strategic priorities. A company in a heavily regulated industry might increase the weight on Risk Profile; a startup might increase the weight on Business Impact.
  4. Score each criterion on a 1--5 scale. Multiply by the weight. Sum for the composite score.
  5. Do not treat the composite score as absolute truth. Use it to structure discussion, surface disagreements, and identify clusters of high-priority versus low-priority initiatives.

Scoring Guide

Score Business Impact Feasibility Data Readiness Strategic Alignment Risk Profile
5 >$5M annual value or critical competitive advantage Proven approach, team has deep experience Clean, labeled, accessible, sufficient volume Core to top 3 strategic priorities Minimal regulatory, ethical, or reputational risk
4 $1--5M annual value or significant efficiency gain Established approach, team has some experience Available but requires moderate preparation Directly supports a stated strategic goal Low risk with standard controls
3 $250K--1M annual value or notable improvement Feasible but requires new capabilities or tools Exists but requires significant cleaning, labeling, or integration Indirectly supports strategic goals Moderate risk requiring active mitigation
2 $50--250K annual value or incremental improvement Technically challenging, limited internal expertise Partially available, major gaps or quality issues Tangentially related to strategy Significant regulatory, bias, or privacy concerns
1 <$50K annual value or unclear benefit Cutting-edge research, high uncertainty Does not exist or is inaccessible No clear strategic connection Severe risk: high-stakes decisions, vulnerable populations, regulatory prohibition

Template

Default Weights: Business Impact (30%), Feasibility (25%), Data Readiness (20%), Strategic Alignment (15%), Risk Profile (10%)

Use Case Business Impact (x0.30) Feasibility (x0.25) Data Readiness (x0.20) Strategic Alignment (x0.15) Risk Profile (x0.10) Composite Score
[Use Case 1] [1-5] × 0.30 = [1-5] × 0.25 = [1-5] × 0.20 = [1-5] × 0.15 = [1-5] × 0.10 = [Sum]
[Use Case 2] [1-5] × 0.30 = [1-5] × 0.25 = [1-5] × 0.20 = [1-5] × 0.15 = [1-5] × 0.10 = [Sum]
[Use Case 3] [1-5] × 0.30 = [1-5] × 0.25 = [1-5] × 0.20 = [1-5] × 0.15 = [1-5] × 0.10 = [Sum]
[Use Case 4] [1-5] × 0.30 = [1-5] × 0.25 = [1-5] × 0.20 = [1-5] × 0.15 = [1-5] × 0.10 = [Sum]
[Use Case 5] [1-5] × 0.30 = [1-5] × 0.25 = [1-5] × 0.20 = [1-5] × 0.15 = [1-5] × 0.10 = [Sum]

Priority Tiers: - Tier 1 (Score 4.0--5.0): Pursue immediately. Allocate resources and assign executive sponsor. - Tier 2 (Score 3.0--3.9): Develop further. Invest in a proof of concept or data readiness initiative. - Tier 3 (Score 2.0--2.9): Monitor. Revisit when conditions improve (better data, lower risk, stronger alignment). - Tier 4 (Score <2.0): Deprioritize. Document rationale and archive.

Example (Completed)

Use Case Impact (x0.30) Feasibility (x0.25) Data (x0.20) Alignment (x0.15) Risk (x0.10) Score
Customer churn prediction 5 × 0.30 = 1.50 4 × 0.25 = 1.00 4 × 0.20 = 0.80 5 × 0.15 = 0.75 4 × 0.10 = 0.40 4.45
Invoice anomaly detection 3 × 0.30 = 0.90 4 × 0.25 = 1.00 3 × 0.20 = 0.60 3 × 0.15 = 0.45 5 × 0.10 = 0.50 3.45
Autonomous pricing engine 5 × 0.30 = 1.50 2 × 0.25 = 0.50 2 × 0.20 = 0.40 4 × 0.15 = 0.60 2 × 0.10 = 0.20 3.20
Resume screening automation 3 × 0.30 = 0.90 3 × 0.25 = 0.75 2 × 0.20 = 0.40 2 × 0.15 = 0.30 1 × 0.10 = 0.10 2.45

The scoring team placed customer churn prediction in Tier 1, invoice anomaly detection and autonomous pricing in Tier 2 (with pricing flagged for a data readiness initiative), and resume screening in Tier 3 due to elevated bias and regulatory risk.


Template 3: ML Project Canvas

Purpose

Adapted from the ML Canvas developed by Louis Dorard, this template provides a structured, one-page view of the core components of a machine learning project. It bridges the gap between business requirements and technical specification by forcing practitioners to define what the model predicts, what decisions it informs, what data it requires, how it will be evaluated, and how it will be monitored in production. Complete this canvas before writing any code.

Instructions for Use

  1. Start with the top row (Prediction Task, Decisions, Impact). These sections anchor the entire canvas in business value. If you cannot articulate the decision the model supports, stop and clarify with stakeholders before proceeding.
  2. Move to the middle row (Data Sources, Features, Offline Evaluation). These sections define the technical core of the project.
  3. Finish with the bottom row (Online Evaluation, Monitoring, Maintenance). These sections are the most frequently neglected and the most important for production success.
  4. Review the completed canvas with both business sponsors and technical leads. It should be understandable to both audiences.
  5. Post the canvas in a visible location (physical or digital) and update it as the project evolves.

Template


ML PROJECT CANVAS

Field Details
Project Name [Name]
Owner [Name, role]
Date [Date]
Version [e.g., 1.0]

ROW 1: BUSINESS CONTEXT

Prediction Task Decisions Supported Business Impact
What is the model predicting? (target variable, prediction type --- classification/regression/ranking, prediction horizon) What business decisions will this prediction inform? Who acts on the prediction and how? What is the expected business value? How will you measure impact on KPIs?
[Write here] [Write here] [Write here]

ROW 2: TECHNICAL CORE

Data Sources Features Offline Evaluation
What data do you need? Where does it live? What is its granularity, freshness, and volume? Any labeling requirements? What are the most important features? How will they be engineered? Any features that must be excluded (leakage, privacy, fairness)? What metrics will you use? (accuracy, precision, recall, AUC, RMSE, etc.) What is the minimum acceptable performance? What is the baseline to beat?
[Write here] [Write here] [Write here]

ROW 3: PRODUCTION AND OPERATIONS

Online Evaluation Monitoring Plan Maintenance and Retraining
How will you evaluate the model in production? (A/B test, shadow mode, canary deployment) What is the success criterion for full rollout? What will you monitor? (prediction drift, data drift, feature distributions, latency, error rates) What alerting thresholds will you set? How often will you retrain? What triggers retraining? Who is responsible for ongoing model health? What is the decommissioning plan?
[Write here] [Write here] [Write here]

CONSTRAINTS AND ASSUMPTIONS

Constraint Type Description
Latency [e.g., predictions must return in <200ms]
Fairness [e.g., model must not discriminate by protected attributes]
Interpretability [e.g., must provide feature-level explanations for each prediction]
Regulatory [e.g., must comply with GDPR right to explanation]
Data Privacy [e.g., no PII in feature set, data must remain in EU region]

Example (Completed)

Project Name: 90-Day Enterprise Churn Predictor Prediction Task: Binary classification --- will an enterprise customer churn within 90 days? Predicted weekly for each active enterprise account. Decisions Supported: Customer success managers use the churn probability score to prioritize outreach. Accounts with >60% churn probability are assigned to senior CSMs for intervention. VP of Customer Success uses aggregate churn forecasts for quarterly planning. Data Sources: CRM activity logs (Salesforce, 4 years), product usage telemetry (Mixpanel, 2 years), support ticket history (Zendesk, 3 years), billing and contract data (NetSuite, 5 years). Labels derived from historical churn events (non-renewal or mid-contract cancellation). Offline Evaluation: Primary metric: AUC-ROC (target >0.85, baseline logistic regression at 0.72). Secondary: precision at top decile (target >0.60). Evaluated on time-based train/test split to prevent temporal leakage. Monitoring: Weekly drift detection on top 20 features using Population Stability Index (alert if PSI > 0.2). Daily prediction distribution monitoring. Monthly model performance review against realized churn outcomes.


Template 4: Data Readiness Assessment

Purpose

Data readiness is the single most common point of failure for AI projects. This checklist provides a structured assessment of whether the data needed for a specific AI initiative actually exists, is accessible, is of sufficient quality, and is governed appropriately. Complete this assessment during the discovery phase of any AI project, before committing significant resources to model development.

Instructions for Use

  1. Complete one assessment per primary data source. If a project requires multiple data sources, assess each independently and then assess the integration challenge.
  2. For each item, mark the current status: Green (ready), Yellow (needs work but achievable in <4 weeks), Red (major gap requiring >4 weeks or significant investment), or N/A (not applicable).
  3. Aggregate the results into an overall data readiness score. Use the scoring rubric at the bottom.
  4. Share the completed assessment with the project sponsor. A project with multiple Red items should not proceed to model development until those gaps are addressed.

Template


DATA READINESS ASSESSMENT

Field Details
Project Name [Name]
Data Source [Source name and system]
Assessor [Name, role]
Date [Date]

SECTION A: Data Availability

# Question Status Notes
A1 Does the required data exist? [G/Y/R] [Details]
A2 Is the data in a queryable/extractable format? [G/Y/R] [Details]
A3 Is there sufficient historical depth for the use case? [G/Y/R] [Specify: need X months/years, have Y]
A4 Is the data volume sufficient for the intended modeling approach? [G/Y/R] [Specify: need X rows, have Y]
A5 Is labeled/annotated data available (if supervised learning)? [G/Y/R/N/A] [Specify labeling source and coverage]
A6 Can the data be refreshed at the required frequency for production? [G/Y/R] [Specify: need daily/weekly/real-time]

SECTION B: Data Quality

# Question Status Notes
B1 What is the rate of missing values in key fields? [G/Y/R] [Specify: field X has Y% missing]
B2 Are there known data entry errors or inconsistencies? [G/Y/R] [Details]
B3 Are there duplicate records that need resolution? [G/Y/R] [Estimated duplicate rate]
B4 Is the data consistent across time periods? (no schema changes, definition shifts) [G/Y/R] [Details]
B5 Are there known biases in data collection that could affect model fairness? [G/Y/R] [Details]
B6 Has the data been validated against a ground truth or external source? [G/Y/R] [Details]

SECTION C: Data Governance and Access

# Question Status Notes
C1 Is there a clear data owner who can authorize use for this project? [G/Y/R] [Owner name]
C2 Does the data contain PII or sensitive information? [G/Y/R] [Specify categories]
C3 Are there regulatory constraints on data use? (GDPR, HIPAA, CCPA, etc.) [G/Y/R] [Specify regulations]
C4 Has legal/compliance reviewed and approved the intended data use? [G/Y/R] [Approval status]
C5 Are data sharing agreements or contracts required? [G/Y/R] [Status]
C6 Is there a data retention policy that affects model training data? [G/Y/R] [Details]

SECTION D: Infrastructure and Integration

# Question Status Notes
D1 Can the data be accessed by the ML platform/environment? [G/Y/R] [Details]
D2 Is there an ETL/ELT pipeline for this data source? [G/Y/R] [Details]
D3 Can data from multiple sources be joined reliably? (common keys, entity resolution) [G/Y/R] [Details]
D4 Is the data environment (storage, compute) sufficient for the project? [G/Y/R] [Details]
D5 Is there version control or lineage tracking for the data? [G/Y/R] [Details]
D6 Is there a feature store or equivalent for feature management? [G/Y/R] [Details]

OVERALL READINESS SCORE

Rating Criteria
Ready (Go) No Red items. Fewer than 4 Yellow items. All Section C items are Green.
Conditional (Go with Plan) 1--2 Red items with a clear remediation plan. Fewer than 8 Yellow items.
Not Ready (Stop) 3 or more Red items. Any Red item in Section C (governance). No clear path to remediation.
Overall Assessment [Ready / Conditional / Not Ready]
Key Gaps [List the most critical gaps]
Remediation Plan [Actions, owners, and timelines for addressing gaps]
Estimated Time to Ready [Weeks/months]

Example (Completed)

Project: Customer Churn Prediction. Data Source: Salesforce CRM. Highlights: A1 (Green) --- 4 years of account activity data exists. B1 (Yellow) --- the "reason for contact" field is missing in 23% of records from 2022 and earlier; imputation or exclusion strategy needed. C2 (Yellow) --- data contains customer company names and contact names; pseudonymization required before loading into ML environment. D3 (Yellow) --- joining Salesforce data with Mixpanel usage data requires matching on account domain, which has inconsistent formatting (85% match rate currently). Overall: Conditional. Three Yellow items with clear remediation paths. Estimated 3 weeks to Ready.


Template 5: Vendor/Platform Evaluation Scorecard

Purpose

Choosing an AI platform or vendor is a high-stakes, long-lived decision. This scorecard provides a structured evaluation framework that prevents teams from being swayed by impressive demos while overlooking critical factors like security posture, vendor lock-in, and total cost of ownership. Use it when evaluating cloud ML platforms, AutoML tools, AI-as-a-service providers, or specialized AI vendors.

Instructions for Use

  1. Customize the criteria and weights before beginning the evaluation. The defaults below reflect a general enterprise perspective; adjust for your context.
  2. Assign 2--3 evaluators per vendor. Ensure at least one evaluator is technical (can assess architecture and APIs) and one is business-oriented (can assess pricing models and contract terms).
  3. Score each criterion on a 1--5 scale after hands-on testing, reference checks, and documentation review. Avoid scoring based solely on vendor presentations.
  4. Calculate the weighted composite score and use it to structure a recommendation, not as a mechanical decision.
  5. Document the rationale for each score in the Notes column. Future teams will need to understand why a particular vendor was selected or rejected.

Template


VENDOR/PLATFORM EVALUATION SCORECARD

Field Details
Evaluation Purpose [e.g., Select cloud ML platform for enterprise AI initiatives]
Evaluators [Names and roles]
Date [Date]
Vendors Evaluated [Vendor A, Vendor B, Vendor C]

Scoring Scale: 1 = Unacceptable, 2 = Below Requirements, 3 = Meets Requirements, 4 = Exceeds Requirements, 5 = Best in Class

# Criterion Weight Vendor A Vendor B Vendor C Notes
Functionality
1 Core ML/AI capabilities (training, deployment, monitoring) 15% [1-5] [1-5] [1-5]
2 Breadth of supported algorithms, frameworks, and model types 10% [1-5] [1-5] [1-5]
3 AutoML and low-code capabilities 5% [1-5] [1-5] [1-5]
4 GenAI / LLM support (fine-tuning, RAG, prompt management) 5% [1-5] [1-5] [1-5]
Integration and Architecture
5 API quality, SDK maturity, and developer experience 10% [1-5] [1-5] [1-5]
6 Integration with existing data infrastructure 10% [1-5] [1-5] [1-5]
7 Portability and avoidance of vendor lock-in 5% [1-5] [1-5] [1-5]
Security and Compliance
8 Security certifications and data protection 10% [1-5] [1-5] [1-5]
9 Compliance with relevant regulations (GDPR, HIPAA, SOC 2) 5% [1-5] [1-5] [1-5]
10 Data residency and sovereignty options 5% [1-5] [1-5] [1-5]
Commercial
11 Pricing transparency and predictability 5% [1-5] [1-5] [1-5]
12 Total cost of ownership (including hidden costs) 5% [1-5] [1-5] [1-5]
Support and Ecosystem
13 Technical support quality and responsiveness 5% [1-5] [1-5] [1-5]
14 Documentation, community, and training resources 3% [1-5] [1-5] [1-5]
15 Vendor viability and product roadmap 2% [1-5] [1-5] [1-5]
WEIGHTED TOTAL 100% [Sum] [Sum] [Sum]

Additional Considerations (not scored but documented):

  • [ ] Contract flexibility (term length, exit clauses, SLA guarantees)
  • [ ] Reference customers in our industry
  • [ ] Migration support and onboarding resources
  • [ ] Environmental sustainability commitments
  • [ ] AI ethics and responsible AI features (bias detection, explainability tools)

Recommendation: [Write recommendation with rationale]


Example (Completed)

Evaluation: Cloud ML platform for enterprise deployment. Vendors: AWS SageMaker, Azure ML, Google Vertex AI. SageMaker scored highest on integration (4.5 weighted) due to the company's existing AWS footprint. Vertex AI scored highest on core ML capabilities (4.8 weighted) and GenAI support (5.0). Azure ML scored highest on security and compliance (4.6 weighted) due to built-in HIPAA compliance for the healthcare division. Recommendation: Proceed with Azure ML as the primary platform given the healthcare division's regulatory requirements, with a secondary evaluation of Vertex AI for the marketing team's GenAI use cases.


Template 6: AI Ethics Impact Assessment

Purpose

This framework provides a structured process for identifying, evaluating, and mitigating the ethical risks of an AI system before and during deployment. It is designed to complement --- not replace --- technical testing by examining the broader human, social, and organizational impacts of AI. Complete this assessment during the design phase and revisit it at each major milestone.

Instructions for Use

  1. Convene a review panel that includes at least one person outside the project team. External perspectives catch blind spots that insiders miss.
  2. Complete the assessment collaboratively. Do not delegate it to a single person or treat it as a checkbox exercise.
  3. For each risk identified, assign a severity rating and define a concrete mitigation action with an owner and deadline.
  4. The assessment should be a living document. Update it when the system's scope, data, or deployment context changes.
  5. Escalate any High-severity risks to the AI governance body (if one exists) or to senior leadership before proceeding.

Template


AI ETHICS IMPACT ASSESSMENT

Field Details
System Name [Name]
System Description [1--2 sentence description of what the system does]
Assessment Lead [Name, role]
Review Panel [Names and roles]
Date [Date]
Assessment Version [e.g., 1.0]

SECTION 1: Stakeholder Analysis

Stakeholder Group How They Interact with the System Potential Benefits Potential Harms
[Direct users] [Description] [Benefits] [Harms]
[Subjects of predictions/decisions] [Description] [Benefits] [Harms]
[Downstream decision-makers] [Description] [Benefits] [Harms]
[Affected communities] [Description] [Benefits] [Harms]
[Organization / employees] [Description] [Benefits] [Harms]

SECTION 2: Fairness and Bias

# Question Assessment Severity
2.1 Does the system make or inform decisions about individuals? [Yes/No. If yes, describe.] [H/M/L/N/A]
2.2 Could the system produce different outcomes for different demographic groups? [Assessment] [H/M/L]
2.3 Does the training data reflect historical biases that could be perpetuated? [Assessment] [H/M/L]
2.4 Have you tested for disparate impact across protected characteristics? [Assessment] [H/M/L]
2.5 Are there proxy variables in the feature set that correlate with protected attributes? [Assessment] [H/M/L]

SECTION 3: Privacy and Data Rights

# Question Assessment Severity
3.1 Does the system process personal data? What categories? [Assessment] [H/M/L]
3.2 Were individuals informed that their data would be used for this purpose? [Assessment] [H/M/L]
3.3 Can individuals access, correct, or delete their data? [Assessment] [H/M/L]
3.4 Is there a risk of re-identification from model outputs or aggregated data? [Assessment] [H/M/L]
3.5 Does the system comply with applicable data protection regulations? [Assessment] [H/M/L]

SECTION 4: Transparency and Explainability

# Question Assessment Severity
4.1 Can the system's decisions be explained to affected individuals? [Assessment] [H/M/L]
4.2 Do users understand the system's capabilities and limitations? [Assessment] [H/M/L]
4.3 Is there a meaningful mechanism for individuals to contest a decision? [Assessment] [H/M/L]
4.4 Is the system's existence and use disclosed to those affected by it? [Assessment] [H/M/L]

SECTION 5: Safety and Reliability

# Question Assessment Severity
5.1 What happens when the system fails or produces incorrect outputs? [Assessment] [H/M/L]
5.2 Is there human oversight for high-stakes decisions? [Assessment] [H/M/L]
5.3 Could the system be manipulated through adversarial inputs? [Assessment] [H/M/L]
5.4 Are there adequate fallback mechanisms? [Assessment] [H/M/L]

SECTION 6: Societal and Environmental Impact

# Question Assessment Severity
6.1 Could this system contribute to job displacement? [Assessment] [H/M/L]
6.2 Could it concentrate power or create harmful dependencies? [Assessment] [H/M/L]
6.3 What is the environmental cost (compute, energy) of training and running this system? [Assessment] [H/M/L]
6.4 Could the system be repurposed for harmful uses? [Assessment] [H/M/L]

SECTION 7: Risk Mitigation Plan

Risk ID Risk Description Severity Mitigation Action Owner Deadline Status
[e.g., 2.3] [Description] [H/M/L] [Action] [Name] [Date] [Open/In Progress/Closed]
[e.g., 3.1] [Description] [H/M/L] [Action] [Name] [Date] [Open/In Progress/Closed]

OVERALL ETHICAL RISK RATING: [Low / Medium / High / Unacceptable]

Approval: [Approved / Approved with Conditions / Not Approved]

Conditions (if applicable): [List conditions that must be met before deployment]


Example (Completed)

System: Customer churn prediction model. Overall Risk Rating: Medium. Key findings: (1) The model uses zip code as a feature, which correlates with race and income (Severity: Medium, Risk 2.5). Mitigation: Test for disparate impact across demographic segments using proxy analysis; remove zip code if disparate impact exceeds the four-fifths rule threshold. Owner: Data Science Lead, Deadline: Pre-deployment. (2) Customer success managers may over-rely on model scores without understanding limitations (Severity: Medium, Risk 4.2). Mitigation: Add confidence intervals to the output, provide training on model limitations, and establish quarterly calibration reviews. Owner: CS Director, Deadline: Launch week.


Template 7: Model Card

Purpose

Based on the framework proposed by Mitchell et al. (2019) at Google, a model card is a standardized document that accompanies a trained machine learning model. It communicates what the model does, how it was built, how it performs, and what its limitations are. Model cards promote transparency, enable informed use, and support responsible deployment. Every model deployed to production should have a model card.

Instructions for Use

  1. Create the model card when the model is ready for deployment. Update it whenever the model is retrained, its scope changes, or new limitations are discovered.
  2. Write for multiple audiences: data scientists who will maintain the model, product managers who will make decisions based on its outputs, and compliance reviewers who need to assess risk.
  3. Be honest about limitations. A model card that claims no limitations is not credible and is therefore useless.
  4. Store the model card alongside the model artifact in your model registry or version control system.

Template


MODEL CARD

Model Details

Field Details
Model Name [Name and version, e.g., churn-predictor-v2.3]
Model Type [e.g., Gradient Boosted Trees (XGBoost)]
Framework / Library [e.g., XGBoost 2.0.3, Python 3.11]
Developed By [Team or individual]
Date [Training date]
License [e.g., Internal use only, Apache 2.0]
Contact [Email or team channel]
Model Registry Link [URL]

Intended Use

Field Details
Primary Intended Use [Describe the intended application]
Primary Intended Users [Who should use this model]
Out-of-Scope Uses [Uses the model is NOT designed for and should NOT be used for]

Training Data

Field Details
Dataset Description [What data was used, source, time range]
Dataset Size [Number of samples, features]
Preprocessing [Key preprocessing steps]
Label Definition [How the target variable was defined]
Known Gaps or Biases [Any known issues with the training data]

Evaluation Data

Field Details
Dataset Description [What data was used for evaluation, how it differs from training]
Dataset Size [Number of samples]
Split Strategy [e.g., time-based split, stratified k-fold]

Performance Metrics

Metric Overall Segment A Segment B Segment C
[Metric 1, e.g., AUC-ROC] [Value] [Value] [Value] [Value]
[Metric 2, e.g., Precision] [Value] [Value] [Value] [Value]
[Metric 3, e.g., Recall] [Value] [Value] [Value] [Value]
[Metric 4, e.g., F1] [Value] [Value] [Value] [Value]

Ethical Considerations

Consideration Assessment
Sensitive Attributes [What protected attributes were considered]
Fairness Testing [What fairness tests were conducted and results]
Potential Harms [What harms could arise from model errors]
Mitigation Steps [What was done to mitigate identified risks]

Limitations and Recommendations

Limitation Recommendation
[Limitation 1] [How to account for this limitation]
[Limitation 2] [How to account for this limitation]
[Limitation 3] [How to account for this limitation]

Caveats and Additional Notes

[Any additional context, known issues, or important caveats for users]


Example (Completed)

Model Name: churn-predictor-v2.3 (XGBoost). Intended Use: Predict 90-day churn probability for enterprise SaaS accounts to prioritize customer success outreach. Out-of-Scope Uses: This model should NOT be used for: individual employee performance evaluation, automated contract termination, or predictions for SMB accounts (trained only on enterprise data). Key Metrics: AUC-ROC 0.87 overall; 0.91 for accounts >$100K ARR; 0.78 for accounts $25--50K ARR. The model performs significantly worse on smaller enterprise accounts due to sparser behavioral data. Key Limitation: The model was trained on data from a period of macroeconomic stability (2020--2024). Performance during economic downturns is unknown and should be monitored closely if market conditions change.


Template 8: AI ROI Business Case

Purpose

This template provides a structured approach to building the financial case for an AI project. It forces discipline around cost estimation (including commonly overlooked categories like ongoing maintenance and change management), benefit quantification (distinguishing between hard savings and soft benefits), and financial analysis (NPV, payback period, sensitivity). Use it to secure budget approval and to establish the financial baseline against which the project will be evaluated.

Instructions for Use

  1. Be conservative in benefit estimates and generous in cost estimates. AI projects almost always cost more and deliver value later than initially projected.
  2. Distinguish between benefits you can measure in dollars (revenue increase, cost reduction) and benefits that are real but harder to quantify (improved customer experience, reduced risk, faster decisions). Both matter, but they require different treatment in the business case.
  3. Include a sensitivity analysis. What happens to the ROI if the model performs at 75% of the expected level? What if deployment takes 50% longer? Decision-makers need to understand how robust the case is.
  4. Plan for a 3-year horizon at minimum. Many AI projects do not break even until year 2.

Template


AI ROI BUSINESS CASE

Field Details
Project Name [Name]
Prepared By [Name, role]
Date [Date]
Analysis Period [e.g., 3 years]
Discount Rate [e.g., 10%]

SECTION 1: Cost Estimate

Cost Category Year 0 (Setup) Year 1 Year 2 Year 3 Total
Personnel
- Data scientists / ML engineers [$]` | `[$] [$]` | `[$] [$]
- Data engineers [$]` | `[$] [$]` | `[$] [$]
- Project management [$]` | `[$] [$]` | `[$] [$]
- External consultants [$]` | `[$] [$]` | `[$] [$]
Technology
- Cloud compute and storage [$]` | `[$] [$]` | `[$] [$]
- ML platform licensing [$]` | `[$] [$]` | `[$] [$]
- Data acquisition / labeling [$]` | `[$] [$]` | `[$] [$]
- Integration development [$]` | `[$] [$]` | `[$] [$]
Organizational
- Training and change management [$]` | `[$] [$]` | `[$] [$]
- Process redesign [$]` | `[$] [$]` | `[$] [$]
- Governance and compliance [$]` | `[$] [$]` | `[$] [$]
Ongoing Operations
- Model monitoring and maintenance [$]` | `[$] [$]` | `[$] [$]
- Retraining and updates [$]` | `[$] [$]` | `[$] [$]
- Incident response [$]` | `[$] [$]` | `[$] [$]
Contingency (15--20%) [$]` | `[$] [$]` | `[$] [$]
TOTAL COSTS [$]`** | **`[$] [$]`** | **`[$] [$]

SECTION 2: Benefit Estimate

Benefit Category Year 1 Year 2 Year 3 Total Confidence
Hard Benefits (Quantifiable)
- Revenue increase [$]` | `[$] [$]` | `[$] [H/M/L]
- Cost reduction [$]` | `[$] [$]` | `[$] [H/M/L]
- Productivity improvement [$]` | `[$] [$]` | `[$] [H/M/L]
- Error / waste reduction [$]` | `[$] [$]` | `[$] [H/M/L]
Soft Benefits (Describable)
- Improved customer experience [Description]
- Faster decision-making [Description]
- Reduced risk / improved compliance [Description]
- Competitive differentiation [Description]
TOTAL HARD BENEFITS [$]`** | **`[$] [$]`** | **`[$]

SECTION 3: Financial Analysis

Metric Value Notes
Total Investment (3-year) [$] Sum of all costs
Total Hard Benefits (3-year) [$] Sum of quantifiable benefits
Net Present Value (NPV) [$] Discounted cash flows at [X]%
Internal Rate of Return (IRR) [X]%
Payback Period [X] months When cumulative benefits exceed cumulative costs
ROI (3-year) [X]% (Total Benefits - Total Costs) / Total Costs
Benefit-to-Cost Ratio [X]:1 Total Benefits / Total Costs

SECTION 4: Sensitivity Analysis

Scenario Assumption Change Impact on NPV Impact on Payback
Base Case As projected [$] [X months]
Conservative Benefits 25% lower, costs 25% higher [$] [X months]
Optimistic Benefits 25% higher, costs on target [$] [X months]
Delayed Adoption Full benefits delayed 6 months [$] [X months]
Model Underperformance Model achieves 75% of target accuracy [$] [X months]

SECTION 5: Recommendation

[Write 2--3 paragraph recommendation including the financial case, key risks, and conditions for approval]


Example (Completed)

Project: Customer Churn Prediction. 3-Year Analysis at 10% Discount Rate. Total Investment: $665K (Year 0: $285K setup, Years 1--3: $95K/year operations, 15% contingency). Total Hard Benefits: $8.4M (retained ARR from reduced churn: $2.1M Year 1, $3.0M Year 2, $3.3M Year 3). NPV: $5.8M. Payback: 4.2 months. ROI: 1,163%. Sensitivity: Even in the conservative scenario (benefits 25% lower, costs 25% higher), NPV remains strongly positive at $3.9M with a 6.1-month payback. The case is robust.


Template 9: AI Governance Charter

Purpose

An AI governance charter establishes the mandate, membership, authority, and operating procedures of the body responsible for overseeing AI within an organization. Without a formal charter, governance defaults to ad hoc decision-making, inconsistent standards, and diffused accountability. This template provides a comprehensive framework that can be adapted to organizations of any size, from a three-person review committee at a startup to a multi-tiered governance structure at a Fortune 500 company.

Instructions for Use

  1. Draft the charter with input from legal, compliance, technology, business, HR, and (if applicable) external advisors. Governance that is designed by one function rarely earns the trust of others.
  2. Secure explicit executive sponsorship. A governance body without executive backing will be ignored when it makes inconvenient decisions.
  3. Start with a scope and authority level that is appropriate for your organization's AI maturity. Overly ambitious charters that claim authority over every algorithm often collapse under their own weight.
  4. Review and update the charter annually. AI capabilities, regulations, and organizational needs evolve rapidly.

Template


AI GOVERNANCE CHARTER

1. Purpose and Scope

Field Details
Purpose [Why this governance body exists --- e.g., "To ensure that AI systems developed and deployed by [Company Name] are safe, ethical, effective, and aligned with organizational values and regulatory requirements."]
Scope [What falls under this body's purview --- e.g., "All AI and ML systems that make or significantly inform decisions affecting customers, employees, or business operations."]
Exclusions [What is explicitly out of scope --- e.g., "Simple business rules, traditional statistical reporting, and basic automation that does not involve learned models."]
Effective Date [Date]
Review Date [Annual review date]

2. Governance Body Structure

Field Details
Name [e.g., AI Ethics and Governance Committee, AI Review Board]
Reporting Line [e.g., Reports to the Chief Technology Officer and the Board Risk Committee]
Executive Sponsor [Name, title]

3. Membership

Role Name Title Function Term
Chair [Name] [Title] [Function] [e.g., 2 years, renewable]
Vice Chair [Name] [Title] [Function] [Term]
Member [Name] [Title] Technology/AI [Term]
Member [Name] [Title] Legal/Compliance [Term]
Member [Name] [Title] Business Operations [Term]
Member [Name] [Title] HR/People [Term]
Member [Name] [Title] Data/Privacy [Term]
External Advisor [Name] [Title] [Domain] [Term]
Secretary [Name] [Title] [Function] [Term]

4. Authority and Responsibilities

The governance body shall have authority to:

  • [ ] Approve or reject deployment of AI systems classified as high-risk
  • [ ] Require modifications to AI systems that do not meet ethical, safety, or performance standards
  • [ ] Commission audits and reviews of deployed AI systems
  • [ ] Establish and enforce AI policies, standards, and guidelines
  • [ ] Escalate concerns to the executive team or board
  • [ ] Mandate incident response actions for AI-related failures
  • [ ] [Additional authority items]

The governance body is responsible for:

  • [ ] Maintaining the organization's AI risk classification framework
  • [ ] Reviewing AI ethics impact assessments for high-risk systems
  • [ ] Overseeing the AI project portfolio and prioritization
  • [ ] Monitoring regulatory developments and advising on compliance
  • [ ] Publishing an annual AI governance report
  • [ ] [Additional responsibilities]

5. Operating Procedures

Procedure Details
Meeting Cadence [e.g., Monthly, with ad hoc meetings for urgent matters]
Quorum [e.g., Majority of members including at least one from technology and one from legal/compliance]
Decision Process [e.g., Consensus preferred; majority vote when consensus cannot be reached; Chair breaks ties]
Documentation [e.g., Meeting minutes recorded and distributed within 5 business days; decisions logged in the AI governance register]
Escalation Path [e.g., Unresolved issues escalated to CTO within 10 business days; board-level issues escalated through the Risk Committee]

6. AI Risk Classification Framework

Risk Level Criteria Governance Requirement
Critical Decisions with legal, safety, or civil rights implications; affects vulnerable populations; autonomous decision-making with no human override Full governance review, ethics impact assessment, ongoing monitoring, board notification
High Significant financial or operational impact; uses sensitive personal data; affects large numbers of people Governance review, ethics impact assessment, quarterly monitoring
Medium Moderate business impact; uses non-sensitive data; human-in-the-loop decision process Lightweight review, self-assessment, annual monitoring
Low Internal tools, process automation, no personal data, minimal decision impact Registration only, standard development practices

7. Review and Amendment

This charter shall be reviewed annually by the governance body. Amendments require approval by [specify --- e.g., majority vote of the governance body and executive sponsor]. Material changes to scope or authority require [specify --- e.g., board approval].


Example (Completed)

Name: NovaCorp AI Governance Committee. Scope: All ML models and GenAI applications deployed in production or customer-facing environments. Chair: Chief Data Officer. Meeting Cadence: Monthly (first Thursday), with emergency sessions callable by any member with 48-hour notice. Decision Process: Consensus preferred. When consensus cannot be reached, the committee votes; the Chair votes only to break ties. All decisions and dissenting views are recorded in minutes. Risk Classification in Practice: The committee classified the customer churn model as "Medium" risk (business impact, no personal data in features after pseudonymization, human-in-the-loop) and the loan pre-approval model as "Critical" risk (financial impact on consumers, uses demographic-adjacent data, regulatory scrutiny under fair lending laws).


Template 10: Change Management Plan

Purpose

AI projects fail more often from organizational resistance than from technical shortcomings. This template uses the ADKAR framework (Awareness, Desire, Knowledge, Ability, Reinforcement) to structure a change management plan specifically tailored for AI initiatives. It addresses the unique challenges of AI adoption: fear of job displacement, distrust of algorithmic decisions, lack of AI literacy, and the need for new workflows.

Instructions for Use

  1. Begin this plan during the project design phase, not after deployment. Change management that starts at go-live is change management that starts too late.
  2. Identify stakeholders at every level: executives who sponsor, managers who implement, and frontline employees who use (or are affected by) the AI system.
  3. Tailor communication and training to each stakeholder group. The CTO and the call center agent need different messages.
  4. Plan for resistance. It is not a sign of failure; it is a predictable human response to uncertainty. Name it, understand its sources, and address it directly.
  5. Build in reinforcement mechanisms that sustain adoption beyond the initial launch excitement.

Template


AI CHANGE MANAGEMENT PLAN

Field Details
Project Name [Name]
Change Manager [Name, role]
Executive Sponsor [Name, role]
Date [Date]
Go-Live Date [Planned deployment date]

SECTION 1: Change Impact Assessment

Dimension Current State Future State Impact Level
Processes [How work is done today] [How work will be done with AI] [H/M/L]
Roles and Responsibilities [Current roles] [Changed roles] [H/M/L]
Skills and Competencies [Current skill set] [Required new skills] [H/M/L]
Tools and Technology [Current tools] [New tools/interfaces] [H/M/L]
Culture and Mindset [Current attitudes toward AI] [Required mindset shift] [H/M/L]
Organizational Structure [Current structure] [Any structural changes] [H/M/L]

SECTION 2: Stakeholder Analysis

Stakeholder Group # of People Impact Level Current Readiness Key Concerns ADKAR Gap
[Group 1, e.g., Customer Success Managers] [N] [H/M/L] [H/M/L] [Main concerns] [Which ADKAR element is weakest]
[Group 2, e.g., Sales Leadership] [N] [H/M/L] [H/M/L] [Main concerns] [Weakest element]
[Group 3, e.g., IT Operations] [N] [H/M/L] [H/M/L] [Main concerns] [Weakest element]
[Group 4, e.g., Customers] [N] [H/M/L] [H/M/L] [Main concerns] [Weakest element]

SECTION 3: ADKAR Plan

A --- Awareness (Why is this change happening?)

Activity Target Audience Timing Owner Channel
Executive announcement of AI initiative All staff [Date] [Sponsor] Town hall, email
Department-specific briefings on impact Affected departments [Date] [Dept. leads] Team meetings
FAQ document addressing common concerns All affected staff [Date] [Change Manager] Intranet, Slack
[Additional activity] [Audience] [Date] [Owner] [Channel]

D --- Desire (Building willingness to participate)

Activity Target Audience Timing Owner Channel
Demonstrate early wins and benefits Affected teams [Date] [Project Lead] Demo sessions
Address "What's in it for me" by role Each stakeholder group [Date] [Change Manager] 1:1 and small groups
Identify and empower change champions Selected early adopters [Date] [Change Manager] Direct engagement
Address job displacement concerns directly At-risk roles [Date] [HR + Sponsor] 1:1 meetings
[Additional activity] [Audience] [Date] [Owner] [Channel]

K --- Knowledge (How to work in the new way)

Training Topic Target Audience Format Duration Timing Owner
AI literacy fundamentals All affected staff Workshop [Hours] [Date] [Trainer]
New tool / interface training Direct users Hands-on lab [Hours] [Date] [Trainer]
Interpreting AI outputs and making decisions Decision-makers Case-based workshop [Hours] [Date] [Trainer]
New process / workflow training Affected teams Process walkthrough [Hours] [Date] [Trainer]
When to override or escalate Direct users Scenario training [Hours] [Date] [Trainer]

A --- Ability (Enabling people to perform in the new way)

Activity Target Audience Timing Owner
Supervised practice period with support Direct users [Weeks post-launch] [Team leads]
Help desk / support channel for AI tool questions All users [Ongoing] [IT support]
Job aids, quick reference guides, and cheat sheets Direct users [At launch] [Change Manager]
Performance coaching for struggling adopters Identified individuals [As needed] [Managers]

R --- Reinforcement (Sustaining the change)

Activity Target Audience Timing Owner
Celebrate and publicize early wins All staff [Monthly] [Sponsor]
Incorporate AI tool usage into performance metrics Direct users [Quarter post-launch] [HR + Managers]
Regular feedback collection and iteration All users [Monthly] [Change Manager]
Share impact metrics (time saved, results improved) All staff [Quarterly] [Project Lead]
Refresher training based on usage data Low-adoption users [Quarterly] [Trainer]

SECTION 4: Resistance Management

Anticipated Resistance Source Root Cause Response Strategy
"AI will replace my job" Frontline staff Fear of displacement Communicate that the AI augments their role; provide reskilling; show how AI handles tedious tasks so they can focus on high-value work
"I don't trust the algorithm" Experienced employees Loss of autonomy, expertise not valued Involve them in model validation; show them how their expertise informs the model; maintain human override authority
"This is just another tech fad" Skeptical managers Change fatigue Demonstrate concrete ROI from pilot; connect to strategic priorities; secure visible executive commitment
"The system doesn't work for my cases" Power users Edge cases, legitimate gaps Establish a feedback loop; commit to iterating on the model; acknowledge limitations openly
[Additional resistance] [Source] [Root cause] [Strategy]

Example (Completed)

Project: Customer Churn Prediction deployment for the Customer Success team (42 CSMs, 6 managers, 2 directors). Key Change Impact: CSMs shift from reactive (waiting for renewal date or support escalation) to proactive (acting on weekly churn scores). This requires new daily workflow, new dashboard, and a different skill set (interpreting probabilistic scores rather than binary signals). Biggest ADKAR Gap: Knowledge. CSMs are skilled relationship managers but have no experience interpreting ML model outputs. A 4-hour workshop on "Working with AI Predictions" was designed, including hands-on practice interpreting scores, understanding confidence levels, and knowing when to override the model. Primary Resistance: Two senior CSMs with 10+ years of experience expressed concern that "a computer can't understand my customers better than I do." Response: Invited them to participate in model validation, showed them cases where the model caught early warning signals they had missed, and positioned the tool as amplifying --- not replacing --- their intuition.


Template 11: AI Team Hiring Plan

Purpose

Building an effective AI team requires more than posting job descriptions for data scientists. This template provides a structured approach to defining roles, identifying skill requirements, planning team structure, and designing an interview process that evaluates both technical competence and the collaborative, communication, and ethical reasoning skills that distinguish effective AI practitioners from technically brilliant but organizationally ineffective ones.

Instructions for Use

  1. Start with the business problems the team will solve, not with the roles you think you need. The required team composition follows from the use case portfolio.
  2. Distinguish between skills you must hire for (hard to develop internally) and skills you can develop through training (easier to build).
  3. Design the interview process to assess what you actually need. A take-home modeling challenge reveals technical skill; a case study presentation reveals communication skill; a scenario discussion reveals ethical reasoning.
  4. Plan for the team you need in 12--18 months, not just today. AI teams tend to grow, and hiring takes longer than expected.

Template


AI TEAM HIRING PLAN

Field Details
Team Name [e.g., Enterprise AI Team, ML Platform Team]
Hiring Manager [Name, title]
Date [Date]
Planning Horizon [e.g., 12 months]
Team Mission [1--2 sentence mission statement]

SECTION 1: Team Structure

Role Current Headcount Target Headcount Gap Priority Hire/Develop
Head of AI / ML Lead [N] [N] [N] [H/M/L] [Hire/Develop]
Senior Data Scientist [N] [N] [N] [H/M/L] [Hire/Develop]
Data Scientist [N] [N] [N] [H/M/L] [Hire/Develop]
ML Engineer [N] [N] [N] [H/M/L] [Hire/Develop]
Data Engineer [N] [N] [N] [H/M/L] [Hire/Develop]
AI Product Manager [N] [N] [N] [H/M/L] [Hire/Develop]
AI Ethics / Responsible AI [N] [N] [N] [H/M/L] [Hire/Develop]
GenAI / LLM Specialist [N] [N] [N] [H/M/L] [Hire/Develop]
[Additional role] [N] [N] [N] [H/M/L] [Hire/Develop]

SECTION 2: Role Definitions

Complete one block per role to be filled.

Role: [Title]

Field Details
Reports To [Title]
Level [e.g., IC3, Manager, Senior Manager]
Location [e.g., Hybrid --- NYC office 3 days/week]
Compensation Range [$X -- $Y base + equity/bonus]
Skill Category Required Skills Nice-to-Have Skills
Technical [List specific technical requirements] [Preferred but not required]
Domain [Industry or business domain knowledge] [Preferred domains]
Tools / Platforms [Specific tools, languages, platforms] [Additional tools]
Soft Skills [Communication, collaboration, etc.] [Leadership, mentoring]
Education / Experience [Minimum requirements] [Preferred qualifications]

SECTION 3: Interview Process

Stage Format Duration Interviewer(s) What It Assesses
1. Resume Screen CV review 15 min Recruiter + Hiring Manager Baseline qualifications
2. Phone Screen Phone / video call 30 min Recruiter Culture fit, motivation, salary alignment
3. Technical Screen Coding / technical questions 60 min Senior Data Scientist Core technical competency
4. Take-Home Challenge Practical ML problem 4--6 hours Panel End-to-end problem solving, code quality
5. On-Site / Virtual Panel
- Technical Deep Dive Presentation of take-home 60 min Technical panel Communication, depth of understanding
- System Design ML system design question 45 min ML Engineer Architecture, scalability, production thinking
- Business Case Study AI business problem 45 min Product Manager + Business Stakeholder Business acumen, stakeholder communication
- Ethical Scenario AI ethics discussion 30 min Hiring Manager Ethical reasoning, judgment
- Culture / Values Behavioral interview 30 min Team members Collaboration, growth mindset
6. Reference Check Phone calls 20 min each Recruiter Verification

SECTION 4: Evaluation Scorecard

Competency Weight Score (1--5) Notes
Technical depth (ML/AI fundamentals) 25% [1-5] [Notes]
Applied problem solving 20% [1-5] [Notes]
Communication and stakeholder management 15% [1-5] [Notes]
Production / engineering mindset 15% [1-5] [Notes]
Business acumen 10% [1-5] [Notes]
Ethical reasoning and responsible AI 10% [1-5] [Notes]
Culture add and collaboration 5% [1-5] [Notes]
Weighted Total 100% [Score]

Hiring Decision Thresholds: - Strong Hire: Weighted score >= 4.0, no competency below 3 - Hire: Weighted score >= 3.5, no competency below 2 - No Hire: Weighted score < 3.5 or any competency at 1

SECTION 5: Hiring Timeline

Milestone Target Date Owner
Job descriptions finalized [Date] [Hiring Manager]
Requisitions approved [Date] [HR/Finance]
Sourcing begins [Date] [Recruiter]
First interviews [Date] [Panel]
Offers extended [Date] [Hiring Manager]
Start dates [Date] [HR]

Example (Completed)

Team: Enterprise AI Team at a mid-market SaaS company. Current: 1 data scientist, 1 data engineer. Target (12 months): 3 data scientists, 2 ML engineers, 2 data engineers, 1 AI product manager. First Hire Priority: Senior Data Scientist (to serve as technical lead while the team scales). Required: 5+ years ML experience, production deployment experience, strong communication skills. Nice-to-have: SaaS domain experience, experience mentoring junior team members. Interview Design Note: The ethical scenario stage asks candidates to discuss how they would handle a situation where a deployed model was discovered to have disparate impact across customer segments. There is no single right answer; the stage assesses whether the candidate identifies the stakeholders, considers multiple perspectives, and proposes a thoughtful course of action.


Template 12: AI Strategy One-Pager

Purpose

An AI strategy one-pager distills the organization's AI vision, priorities, and plan into a format that an executive can read in five minutes. It is not a substitute for a full AI strategy document but rather a communication tool that ensures alignment across leadership, provides a reference point for investment decisions, and enables every team to understand how their work connects to the broader AI agenda. Update it quarterly or when strategic priorities shift.

Instructions for Use

  1. Develop the one-pager after completing the full strategic planning process, not before. It is a summary, not a substitute.
  2. Limit it to one page (two at most). Every sentence must earn its place.
  3. Ensure each strategic pillar has at least one measurable outcome. Strategy without measurement is aspiration.
  4. Circulate for review to the CEO, CTO, CFO, and business unit leaders before finalizing. Misalignment at the top cascades into confusion at every level.
  5. Post it where people will see it. A strategy document that lives in a SharePoint folder no one opens is not a strategy.

Template


AI STRATEGY ONE-PAGER

Field Details
Organization [Company name]
Prepared By [Name, title]
Date [Date]
Time Horizon [e.g., 2026--2028]
Approved By [Name(s), title(s)]

VISION

[1--2 sentences describing the aspirational future state of AI in the organization. What will be true when this strategy is realized?]

CURRENT STATE

[2--3 sentences describing where the organization is today on its AI journey. Be honest. Reference the AI maturity assessment if completed.]

STRATEGIC PILLARS

Pillar Description Key Initiatives Success Metrics Owner
1. [Pillar Name] [1 sentence description] [2--3 initiatives] [Measurable outcomes] [Executive owner]
2. [Pillar Name] [1 sentence description] [2--3 initiatives] [Measurable outcomes] [Executive owner]
3. [Pillar Name] [1 sentence description] [2--3 initiatives] [Measurable outcomes] [Executive owner]
4. [Pillar Name] [1 sentence description] [2--3 initiatives] [Measurable outcomes] [Executive owner]

INVESTMENT SUMMARY

Category Year 1 Year 2 Year 3
People [$]` | `[$] [$]
Technology [$]` | `[$] [$]
Data [$]` | `[$] [$]
Training / Change Management [$]` | `[$] [$]
Total [$]`** | **`[$] [$]

EXPECTED OUTCOMES (3-YEAR)

  • [Outcome 1: e.g., "$15M in AI-driven revenue or cost savings"]
  • [Outcome 2: e.g., "10 AI models in production across 4 business units"]
  • [Outcome 3: e.g., "AI maturity level advanced from Level 2 to Level 4"]
  • [Outcome 4: e.g., "AI governance framework fully operational"]

GOVERNANCE

[1--2 sentences describing who oversees AI strategy execution and how progress is tracked.]

KEY RISKS

Risk Mitigation
[Risk 1] [Mitigation]
[Risk 2] [Mitigation]
[Risk 3] [Mitigation]

Example (Completed)

Vision: By 2028, AI will be embedded in every major decision process at NovaCorp, enabling faster, more accurate, and more equitable outcomes for customers and employees. Current State: NovaCorp is at AI maturity Level 2 (Exploratory). We have 3 ML models in production, a 6-person data science team, and no formal AI governance structure. AI efforts are siloed within the technology organization. Pillars: (1) AI-Powered Customer Experience --- deploy churn prediction, next-best-action recommendation, and intelligent routing (Owner: VP Customer Success). (2) Operational Intelligence --- automate invoice processing, demand forecasting, and quality inspection (Owner: COO). (3) AI Foundation --- build ML platform, data infrastructure, and feature store (Owner: CTO). (4) Responsible AI --- establish governance committee, ethics review process, and AI literacy program (Owner: Chief Legal Officer). 3-Year Investment: $8.2M ($2.1M Year 1, $3.0M Year 2, $3.1M Year 3). Expected Outcomes: $15M in value creation, 10 models in production, maturity Level 4.


Template 13: Prompt Engineering Brief

Purpose

As generative AI becomes a core business tool, organizations need a systematic way to document, share, iterate on, and govern the prompts that drive AI outputs. This template provides a structured format for capturing a prompt's objective, design rationale, expected output, evaluation criteria, and version history. It transforms prompt engineering from an individual craft into an organizational capability. Use it for any prompt that will be used repeatedly, shared across teams, or embedded in a production application.

Instructions for Use

  1. Complete a prompt engineering brief for every prompt that is used in production, shared with more than two people, or expected to be maintained over time. Ad hoc experimental prompts do not need this level of documentation.
  2. Include the full prompt text, including system instructions, few-shot examples, and any formatting specifications. A brief without the actual prompt is incomplete.
  3. Define evaluation criteria before deploying the prompt. How will you know if it is working well? What does a good output look like? What does a bad output look like?
  4. Maintain version history. Prompts evolve as models change, requirements shift, and edge cases emerge. Without version tracking, regression is invisible.
  5. Include at least one example of expected output and one example of unacceptable output.

Template


PROMPT ENGINEERING BRIEF

Field Details
Brief ID [e.g., PE-2026-042]
Prompt Name [Descriptive name]
Author [Name, role]
Date Created [Date]
Current Version [e.g., 3.1]
Status [Draft / In Review / Approved / Deployed / Deprecated]

SECTION 1: Objective and Context

Field Details
Business Objective [What business problem does this prompt solve?]
Target Audience [Who will use the outputs? Internal team? Customers?]
Use Case [Specific scenario where this prompt is used]
Target Model [e.g., GPT-4o, Claude Sonnet 4, Gemini 2.0 Pro]
Integration Point [e.g., Customer support chatbot, internal analysis tool, API endpoint]
Frequency of Use [e.g., 500 times/day, weekly batch, on-demand]

SECTION 2: Prompt Design

System Instructions / System Prompt:

[Full system prompt text here]

User Prompt Template:

[Full user prompt template with {{variables}} marked]

Variables:

Variable Description Type Example Value Required
{{variable_1}} [Description] [Text/Number/List] [Example] [Yes/No]
{{variable_2}} [Description] [Text/Number/List] [Example] [Yes/No]

Few-Shot Examples (if applicable):

# Input Expected Output
1 [Example input] [Example output]
2 [Example input] [Example output]

Configuration Parameters:

Parameter Value Rationale
Temperature [e.g., 0.3] [Why this setting]
Max Tokens [e.g., 1024] [Why this limit]
Top-p [e.g., 0.9] [Why this setting]
Stop Sequences [e.g., "###"] [Why these stops]
Response Format [e.g., JSON, Markdown, plain text] [Why this format]

SECTION 3: Evaluation

Criterion Definition Acceptable Threshold
Accuracy [How accuracy is defined for this use case] [e.g., >95% factual correctness]
Relevance [Output addresses the stated objective] [e.g., addresses all key points]
Tone / Style [Expected voice and register] [e.g., professional, concise]
Completeness [All required elements present] [e.g., all 5 sections included]
Safety [No harmful, biased, or inappropriate content] [Zero tolerance]
Format Compliance [Output matches required structure] [e.g., valid JSON, correct headers]

Example of Acceptable Output:

[Paste a representative good output here]

Example of Unacceptable Output:

[Paste an example of a bad output with annotation explaining why it fails]

SECTION 4: Guardrails and Constraints

  • [ ] Input validation: [How inputs are validated before reaching the prompt]
  • [ ] Output filtering: [Any post-processing or filtering applied to outputs]
  • [ ] Content moderation: [Moderation layer, if applicable]
  • [ ] PII handling: [How PII in inputs/outputs is managed]
  • [ ] Rate limiting: [Usage limits, if applicable]
  • [ ] Fallback behavior: [What happens when the prompt fails or output is filtered]

SECTION 5: Version History

Version Date Author Changes Rationale
[1.0] [Date] [Name] Initial version ---
[2.0] [Date] [Name] [What changed] [Why]
[3.0] [Date] [Name] [What changed] [Why]
[3.1] [Date] [Name] [What changed] [Why]

Example (Completed)

Prompt Name: Customer Email Response Generator. Model: Claude Sonnet 4. Use Case: Customer support agents use this prompt to draft personalized responses to customer inquiries. The agent reviews, edits, and sends the draft. System Prompt (excerpt): "You are a customer support assistant for NovaCorp, a B2B SaaS company. Draft a professional, empathetic response to the customer email below. Use the customer's name. Acknowledge their concern before providing the solution. Keep the response under 200 words. Do not make commitments about timelines or features without explicit approval." Key Evaluation Criteria: Accuracy (references correct product information), Empathy (acknowledges customer's frustration), Actionability (provides clear next steps), Safety (does not promise features or timelines). Version History: v1.0 used generic tone; v2.0 added empathy requirement after feedback that responses felt robotic; v3.0 added the constraint against committing to timelines after an agent inadvertently promised a feature delivery date; v3.1 adjusted temperature from 0.7 to 0.3 after observing inconsistent tone across responses.


Template 14: AI Incident Response Playbook

Purpose

AI systems fail in ways that traditional software does not: models drift silently, biases emerge unexpectedly, adversarial inputs cause harmful outputs, and hallucinated content can mislead users. This playbook provides a structured template for responding to AI-specific incidents --- from detection through resolution and post-incident review. Every AI system in production should have an incident response plan tailored to its risk profile.

Instructions for Use

  1. Customize this playbook for each production AI system. A chatbot and a fraud detection model have very different failure modes and very different response requirements.
  2. Identify the incident response team and ensure every member has read and understood the playbook before an incident occurs. Playbooks that are read for the first time during an incident are ineffective.
  3. Conduct at least one tabletop exercise per year to practice the response process.
  4. After every real incident, update the playbook based on lessons learned. The playbook should improve continuously.
  5. Define severity levels and escalation thresholds clearly. In the moment of an incident, ambiguity about "who decides" creates paralysis.

Template


AI INCIDENT RESPONSE PLAYBOOK

Field Details
System Name [Name]
Playbook Owner [Name, role]
Date [Date]
Version [e.g., 1.0]
Last Tabletop Exercise [Date]
Next Scheduled Review [Date]

SECTION 1: Incident Classification

Severity Definition Examples Response Time Escalation
SEV-1 (Critical) AI system causing active harm, regulatory violation, or major business impact Biased decisions affecting customers at scale; data breach via AI system; harmful content generation in customer-facing app; safety-critical system failure Immediate (within 15 min) CTO, Legal, and CEO notified immediately
SEV-2 (High) AI system producing significantly degraded or incorrect outputs with material business impact Model accuracy dropped below minimum threshold; systematic errors in a specific segment; data pipeline failure affecting model inputs Within 1 hour VP Engineering and AI Lead notified
SEV-3 (Medium) AI system experiencing degradation that is noticeable but not causing direct harm Gradual performance drift; intermittent errors; increased latency affecting user experience Within 4 hours AI Lead notified; team monitors
SEV-4 (Low) Minor issues that do not affect users or business outcomes Monitoring alert for a non-critical metric; minor UI issue in model dashboard; documentation gap Within 1 business day Logged for next sprint

SECTION 2: Incident Response Team

Role Primary Backup Contact
Incident Commander [Name] [Name] [Phone, email, Slack]
AI/ML Lead [Name] [Name] [Phone, email, Slack]
Data Engineer [Name] [Name] [Phone, email, Slack]
Product Owner [Name] [Name] [Phone, email, Slack]
Communications Lead [Name] [Name] [Phone, email, Slack]
Legal/Compliance [Name] [Name] [Phone, email, Slack]
Executive Sponsor [Name] [Name] [Phone, email, Slack]

SECTION 3: Response Procedures

Phase 1: Detection and Triage (Time: 0 to T+30 min)

  • [ ] Incident detected via: [monitoring alert / user report / internal discovery / external report]
  • [ ] Incident logged in tracking system with timestamp, reporter, and initial description
  • [ ] Incident Commander assigned
  • [ ] Severity level assessed using classification table above
  • [ ] Response team assembled based on severity level
  • [ ] Initial impact assessment: How many users/customers/decisions are affected?
  • [ ] War room / incident channel created (for SEV-1 and SEV-2)

Phase 2: Containment (Time: T+30 min to T+2 hours for SEV-1)

  • [ ] Decision: Can the system be safely paused or rolled back?
  • If yes: Initiate rollback to last known good model version or activate fallback system
  • If no: Implement interim controls (human review layer, output filtering, scope restriction)
  • [ ] Affected outputs identified and flagged (decisions made during incident window)
  • [ ] Users/stakeholders notified that an issue is being investigated
  • [ ] Data preserved for post-incident analysis (model inputs, outputs, logs, monitoring data)

Phase 3: Investigation (Time: varies by severity)

  • [ ] Root cause analysis initiated
  • [ ] Data issue? (drift, corruption, pipeline failure, labeling error)
  • [ ] Model issue? (overfitting, concept drift, adversarial input, training bug)
  • [ ] Infrastructure issue? (compute failure, latency, dependency outage)
  • [ ] Integration issue? (API change, upstream system change, feature store error)
  • [ ] Human issue? (incorrect configuration, unauthorized change, process failure)
  • [ ] Impact quantified: number of affected decisions, financial impact, customer impact
  • [ ] Regulatory notification requirements assessed (with Legal/Compliance)

Phase 4: Resolution (Time: varies)

  • [ ] Fix developed and tested
  • [ ] Fix reviewed and approved by Incident Commander and AI Lead
  • [ ] Fix deployed to production
  • [ ] System monitored for stability (minimum 24 hours for SEV-1/SEV-2)
  • [ ] Affected decisions reviewed and remediated where possible
  • [ ] Affected users/customers notified of resolution
  • [ ] Incident status updated to "Resolved"

Phase 5: Post-Incident Review (Time: within 5 business days of resolution)

  • [ ] Post-incident review meeting scheduled with full response team
  • [ ] Incident timeline documented
  • [ ] Root cause confirmed and documented
  • [ ] Contributing factors identified (not just the proximate cause)

Post-Incident Review Questions:

  1. What happened, and when?
  2. How was the incident detected? Could it have been detected earlier?
  3. What was the impact (users, decisions, revenue, reputation)?
  4. What was the root cause?
  5. Did the response process work as designed? Where did it break down?
  6. What monitoring, testing, or process changes would prevent recurrence?
  7. Are there similar risks in other AI systems that should be assessed?

Action Items from Review:

Action Owner Deadline Status
[Action 1] [Name] [Date] [Open]
[Action 2] [Name] [Date] [Open]
[Action 3] [Name] [Date] [Open]

SECTION 4: Communication Templates

Internal Notification (SEV-1/SEV-2):

AI INCIDENT ALERT --- [SEVERITY LEVEL] System: [System name] Detected: [Timestamp] Description: [Brief description of the issue] Impact: [Known or estimated impact] Current Status: [Investigating / Contained / Resolved] Incident Commander: [Name] Next Update: [Time]

External Customer Notification (if required):

We identified an issue with [system/feature] that may have affected [description of impact] between [start time] and [end time]. We have [contained/resolved] the issue and are [taking steps to prevent recurrence / reviewing affected outcomes]. If you believe you were affected, please contact [support channel]. We apologize for the inconvenience and are committed to transparency about this matter.


Example (Completed)

System: Customer-facing product recommendation engine. Incident: On March 3, a monitoring alert flagged that recommendation diversity had dropped by 73% --- the model was recommending the same 12 products to 89% of users. Severity: SEV-2 (significant degradation, no direct harm but material impact on customer experience and revenue). Root Cause: A feature pipeline update on March 2 introduced a schema change that caused the model to receive null values for 8 of 15 user preference features. The model defaulted to popularity-based recommendations, effectively eliminating personalization. Resolution: Rolled back to the previous feature pipeline version. Implemented schema validation checks that would have caught the breaking change. Added a "recommendation diversity" metric to the real-time monitoring dashboard with an alert threshold. Key Lesson: Feature pipeline changes should be treated with the same rigor as model deployments, including staging environment testing and canary rollout.


Template 15: AI Maturity Self-Assessment Questionnaire

Purpose

This self-assessment helps organizations understand where they stand on their AI journey and where they need to invest to advance. It evaluates AI maturity across six dimensions --- Strategy, Data, Technology, Talent, Governance, and Culture --- using 30 questions (5 per dimension). The results identify strengths to leverage, gaps to address, and a roadmap for advancing to the next maturity level. Administer this assessment annually to track progress and recalibrate priorities.

Instructions for Use

  1. Assemble a diverse assessment team (5--10 people) that includes leaders from technology, data, business, HR, legal, and operations. Individual assessments produce biased results; collective assessment produces a more accurate picture.
  2. Have each team member complete the questionnaire independently, then convene to discuss and calibrate scores. Disagreements are informative --- they often reveal organizational blind spots.
  3. Score each question on the 1--5 scale defined in the scoring guide. Be honest. Aspirational scoring defeats the purpose.
  4. Calculate dimension scores and the overall maturity score. Use the maturity level definitions to identify where the organization falls.
  5. Use the results to inform the AI strategy (Template 12) and investment priorities. Focus improvement efforts on the dimensions that are both low-scoring and strategically important.

Scoring Guide

Score Label Definition
1 Not Started No meaningful activity in this area. No plans or awareness.
2 Emerging Initial awareness and ad hoc efforts. No formal processes or dedicated resources.
3 Developing Formal processes being established. Some dedicated resources. Inconsistent execution.
4 Advanced Well-established processes, dedicated resources, consistent execution. Measurable results.
5 Leading Best-in-class practices. Continuous improvement. Competitive advantage. Organization-wide capability.

Template


AI MATURITY SELF-ASSESSMENT

Field Details
Organization [Company name]
Assessment Team [Names and roles]
Date [Date]
Previous Assessment Date [Date or "First assessment"]

DIMENSION 1: STRATEGY (Weight: 20%)

# Question Score (1--5) Evidence / Notes
1.1 Does the organization have a documented AI strategy aligned with business strategy? [ ] [Notes]
1.2 Is there executive sponsorship and accountability for AI initiatives? [ ] [Notes]
1.3 Is there a prioritized portfolio of AI use cases with clear business value? [ ] [Notes]
1.4 Are AI investments evaluated with rigorous business cases (ROI, NPV)? [ ] [Notes]
1.5 Is AI strategy regularly reviewed and updated based on results and market changes? [ ] [Notes]
Dimension Average [Avg]

DIMENSION 2: DATA (Weight: 20%)

# Question Score (1--5) Evidence / Notes
2.1 Is data treated as a strategic asset with clear ownership and stewardship? [ ] [Notes]
2.2 Is data quality measured, monitored, and actively managed? [ ] [Notes]
2.3 Are data pipelines automated, reliable, and scalable? [ ] [Notes]
2.4 Is there a data catalog or discovery mechanism that enables teams to find and access relevant data? [ ] [Notes]
2.5 Are data governance policies (privacy, retention, access control) defined and enforced? [ ] [Notes]
Dimension Average [Avg]

DIMENSION 3: TECHNOLOGY (Weight: 15%)

# Question Score (1--5) Evidence / Notes
3.1 Is there a dedicated ML/AI platform or infrastructure for model development and deployment? [ ] [Notes]
3.2 Are ML models deployed to production using automated, repeatable processes (MLOps)? [ ] [Notes]
3.3 Is there production monitoring for model performance, drift, and data quality? [ ] [Notes]
3.4 Can the organization experiment with and deploy GenAI/LLM solutions responsibly? [ ] [Notes]
3.5 Are AI/ML tools integrated with existing business systems and workflows? [ ] [Notes]
Dimension Average [Avg]

DIMENSION 4: TALENT (Weight: 15%)

# Question Score (1--5) Evidence / Notes
4.1 Does the organization have sufficient AI/ML technical talent (data scientists, ML engineers)? [ ] [Notes]
4.2 Are there defined career paths and development programs for AI practitioners? [ ] [Notes]
4.3 Do business teams have sufficient AI literacy to collaborate effectively with technical teams? [ ] [Notes]
4.4 Is there a structured onboarding program for new AI team members? [ ] [Notes]
4.5 Can the organization attract and retain top AI talent competitively? [ ] [Notes]
Dimension Average [Avg]

DIMENSION 5: GOVERNANCE (Weight: 15%)

# Question Score (1--5) Evidence / Notes
5.1 Is there a formal AI governance body with defined authority and accountability? [ ] [Notes]
5.2 Are AI ethics principles documented and applied to project decisions? [ ] [Notes]
5.3 Is there a risk classification framework for AI systems? [ ] [Notes]
5.4 Are deployed AI models inventoried, documented (model cards), and auditable? [ ] [Notes]
5.5 Is there an incident response process for AI-specific failures and issues? [ ] [Notes]
Dimension Average [Avg]

DIMENSION 6: CULTURE (Weight: 15%)

# Question Score (1--5) Evidence / Notes
6.1 Is there organizational enthusiasm and openness toward AI adoption? [ ] [Notes]
6.2 Do teams across the organization proactively identify AI opportunities? [ ] [Notes]
6.3 Is experimentation with AI encouraged, with tolerance for failure? [ ] [Notes]
6.4 Do business and technical teams collaborate effectively on AI projects? [ ] [Notes]
6.5 Is there a culture of data-driven decision-making (not just data-informed)? [ ] [Notes]
Dimension Average [Avg]

RESULTS SUMMARY

Dimension Weight Average Score Weighted Score Previous Score Change
Strategy 20% [Avg] [Avg × 0.20] [Prior or N/A] [+/-]
Data 20% [Avg] [Avg × 0.20] [Prior or N/A] [+/-]
Technology 15% [Avg] [Avg × 0.15] [Prior or N/A] [+/-]
Talent 15% [Avg] [Avg × 0.15] [Prior or N/A] [+/-]
Governance 15% [Avg] [Avg × 0.15] [Prior or N/A] [+/-]
Culture 15% [Avg] [Avg × 0.15] [Prior or N/A] [+/-]
OVERALL 100% [Sum] [Prior or N/A] [+/-]

MATURITY LEVEL DEFINITIONS

Level Score Range Label Description
1 1.0 -- 1.4 Ad Hoc No formal AI activity. Isolated experiments by individuals. No strategy, governance, or dedicated resources.
2 1.5 -- 2.4 Exploratory Initial AI experiments underway. Some executive interest. No formal strategy or governance. Data infrastructure immature. Talent limited to a few individuals.
3 2.5 -- 3.4 Operational Multiple AI projects in production. Formal strategy exists. Dedicated AI team in place. Data infrastructure being built. Governance emerging. Some business units adopting AI.
4 3.5 -- 4.4 Systematic AI embedded in multiple business processes. Mature MLOps and data infrastructure. Strong governance. AI talent pipeline established. Cross-functional collaboration is the norm. Measurable business impact.
5 4.5 -- 5.0 Transformative AI is a core organizational capability and competitive advantage. Continuous innovation. Leading-edge practices across all dimensions. AI-first culture. Industry-recognized leader.

Our Current Maturity Level: [Level and Label]

Target Maturity Level (12 months): [Level and Label]


DIMENSION GAP ANALYSIS AND ACTION PLAN

Dimension Current Score Target Score Gap Priority Actions Owner Timeline
[Weakest dimension] [Score] [Target] [Gap] [1--2 key actions] [Name] [Timeline]
[Second weakest] [Score] [Target] [Gap] [1--2 key actions] [Name] [Timeline]
[Third weakest] [Score] [Target] [Gap] [1--2 key actions] [Name] [Timeline]

Example (Completed)

Organization: NovaCorp (mid-market SaaS, 1,200 employees). Results: Strategy 2.8, Data 2.2, Technology 2.6, Talent 2.0, Governance 1.6, Culture 3.2. Overall: 2.4 (Exploratory). Interpretation: NovaCorp's strongest dimension is Culture (3.2) --- teams are enthusiastic about AI and willing to experiment. Its weakest dimensions are Governance (1.6) and Talent (2.0). The organization has energy and ambition but lacks the structural foundations to scale AI responsibly. Priority Actions: (1) Governance: Establish an AI governance committee and adopt a risk classification framework within 6 months (Owner: Chief Legal Officer). (2) Talent: Hire 2 senior data scientists and 1 ML engineer; launch an AI literacy program for business leaders (Owner: VP Engineering + HR). (3) Data: Appoint data stewards for the 5 highest-priority data domains; implement automated data quality monitoring (Owner: CTO). Target: Advance from Level 2 (Exploratory) to Level 3 (Operational) within 12 months.


How to Use These Templates Together

The fifteen templates in this appendix are not isolated documents. They form an interconnected toolkit that supports the full lifecycle of AI in an organization. Here is how they fit together:

Strategic Planning Phase - Start with the AI Maturity Self-Assessment (Template 15) to understand where you are. - Use the AI Strategy One-Pager (Template 12) to define where you are going. - Apply the AI Use Case Prioritization Matrix (Template 2) to decide which initiatives to pursue first.

Project Initiation Phase - Write an AI Project Proposal (Template 1) for each selected initiative. - Complete the Data Readiness Assessment (Template 4) to validate that the data foundation exists. - Build the AI ROI Business Case (Template 8) to secure funding. - Complete the ML Project Canvas (Template 3) to align business and technical teams on the project scope.

Build and Deploy Phase - Evaluate platforms using the Vendor/Platform Evaluation Scorecard (Template 5). - Staff the team using the AI Team Hiring Plan (Template 11). - Conduct the AI Ethics Impact Assessment (Template 6) before deployment. - Create a Model Card (Template 7) for each model entering production. - Document critical prompts using the Prompt Engineering Brief (Template 13).

Operate and Govern Phase - Establish oversight with the AI Governance Charter (Template 9). - Prepare for problems with the AI Incident Response Playbook (Template 14). - Manage organizational adoption with the Change Management Plan (Template 10).

Annual Review - Re-administer the AI Maturity Self-Assessment (Template 15) to measure progress. - Update the AI Strategy One-Pager (Template 12) based on results and evolving priorities.

Taken together, these templates transform AI from a collection of ad hoc experiments into a disciplined, governed, and strategically aligned organizational capability. The templates are the scaffolding; the thinking you put into them is the structure.