Chapter 26: Further Reading — Explainable AI (XAI) and Model Governance
Regulatory Frameworks and Primary Sources
Federal Reserve System — SR 11-7: Guidance on Model Risk Management (2011) Board of Governors of the Federal Reserve System and the Office of the Comptroller of the Currency. The foundational US regulatory text on model governance. Defines models broadly, establishes the three-pillar framework (conceptual soundness, ongoing monitoring, outcomes analysis), and specifies independent validation requirements. Required reading for anyone building or overseeing models at a US financial institution. Available at: federalreserve.gov/supervisionreg/srletters/sr1107.htm
European Parliament and Council of the European Union — Regulation (EU) 2024/1689: The EU Artificial Intelligence Act The primary legislative text of the EU AI Act. Annex III lists high-risk AI applications; Articles 9–15 specify requirements for high-risk AI systems including risk management, data governance, technical documentation, transparency, and human oversight. Articles 85–101 address penalties and enforcement. The full text is available through EUR-Lex. Understanding Annex III and the Chapter 2 requirements is essential for financial institutions operating in EU markets.
National Institute of Standards and Technology — AI Risk Management Framework (NIST AI RMF 1.0, 2023) The NIST AI RMF provides a voluntary, comprehensive framework organized around four functions: Govern, Map, Measure, and Manage. While voluntary for most US firms, it is increasingly referenced in supervisory guidance from the OCC and Federal Reserve and provides practical implementation guidance that complements SR 11-7. Accompanied by a detailed AI RMF Playbook with specific actions and outcomes. Available at: nist.gov/artificial-intelligence
Board of Governors of the Federal Reserve System — Regulation B (Equal Credit Opportunity Act Implementation) The implementing regulation for the Equal Credit Opportunity Act. Subpart B, Section 202.9 specifies adverse action notification requirements including the requirement to provide specific reasons for credit decisions. Essential reading for US consumer credit model governance. Available at: consumerfinance.gov/rules-policy/regulations/1002/
Information Commissioner's Office (UK) — Guidance on AI and Data Protection The UK ICO's guidance on AI and data protection covers automated decision-making under UK GDPR Article 22, meaningful explanations, and data protection impact assessments for high-risk processing. Particularly useful for UK firms navigating GDPR Article 22 obligations. Available at: ico.org.uk/for-organisations/uk-gdpr-guidance-and-resources/artificial-intelligence/
Financial Conduct Authority — Discussion Paper DP22/4: AI and Machine Learning (2022) The FCA's most substantive published treatment of AI and ML in financial services. Covers the FCA's concerns about model transparency, governance, and consumer outcomes. Though a discussion paper rather than binding rules, it signals regulatory expectations and has informed subsequent supervisory communications. Available at: fca.org.uk
Academic Literature — Explainability Methods
Lundberg, S. M., & Lee, S.-I. (2017). A Unified Approach to Interpreting Model Predictions. Advances in Neural Information Processing Systems, 30. The original paper introducing SHAP. Establishes the theoretical foundation (Shapley values from game theory), proves that SHAP satisfies the axioms of local accuracy, missingness, and consistency, and introduces TreeSHAP for efficient computation on tree-based models. This is the primary citation for SHAP in regulatory documentation and the correct reference for any validation report or technical methodology document that discusses SHAP-based explanations.
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). "Why Should I Trust You?": Explaining the Predictions of Any Classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. The original LIME paper. Describes the local linear approximation approach, discusses the model-agnostic property, and provides empirical demonstrations across multiple model types. The paper's title reflects the core governance motivation: trust in a classifier depends on the ability to understand and challenge its outputs.
Lundberg, S. M., Erion, G., Chen, H., DeGrave, A., Prutkin, J. M., Nair, B., Katz, R., Himmelfarb, J., Bansal, N., & Lee, S.-I. (2020). From Local Explanations to Global Understanding with Explainable AI for Trees. Nature Machine Intelligence, 2(1), 56–67. Extends SHAP to global explainability, introduces SHAP interaction values, and demonstrates applications in clinical decision support — a domain with governance requirements analogous to those in regulated financial services. Covers SHAP summary plots, dependence plots, and the relationship between local and global explanation.
Wachter, S., Mittelstadt, B., & Russell, C. (2017). Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR. Harvard Journal of Law & Technology, 31(2), 841–887. The foundational paper on counterfactual explanations as a response to GDPR Article 22. Argues that counterfactual explanations are both technically feasible and normatively appropriate as a form of explanation under the GDPR. Essential reading for compliance officers designing explanation frameworks for EU markets.
Mothilal, R. K., Sharma, A., & Tan, C. (2020). Explaining Machine Learning Classifiers through Diverse Counterfactual Explanations. Proceedings of the 2020 ACM FAccT Conference. Introduces the DICE (Diverse Counterfactual Explanations) library, which generates multiple diverse counterfactual explanations to provide applicants with a range of actionable paths to approval. The library is available as a Python package and is actively maintained.
Fairness and Bias
Hardt, M., Price, E., & Srebro, N. (2016). Equality of Opportunity in Supervised Learning. Advances in Neural Information Processing Systems, 29. Introduces and formalizes equalized odds and equal opportunity as fairness criteria. Demonstrates the mathematical incompatibility between fairness metrics when base rates differ across groups — the result discussed in Section 5 of the chapter. A foundational paper for anyone designing fairness testing protocols for financial models.
Chouldechova, A. (2017). Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments. Big Data, 5(2), 153–163. Demonstrates the mathematical impossibility of simultaneously achieving calibration and equal false positive/negative rates across groups with different base rates. Uses the COMPAS recidivism tool as a case study. The mathematical results generalize directly to credit scoring, fraud detection, and other financial classification problems.
Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2021). A Survey on Bias and Fairness in Machine Learning. ACM Computing Surveys, 54(6), 1–35. A comprehensive survey of sources of bias in machine learning (historical bias, representation bias, measurement bias, aggregation bias) and of the major fairness metrics and their properties. Useful as a reference text for building a bias testing framework. Covers both pre-processing (data debiasing), in-processing (constrained training), and post-processing (threshold adjustment) approaches to fairness improvement.
Model Governance — Practice and Implementation
Supervisory Policy Statement on Model Risk Management (SS1/23) — Prudential Regulation Authority (Bank of England) The PRA's supervisory statement on model risk management for UK-authorised deposit takers, insurers, and designated investment firms. Aligns closely with SR 11-7 in substance while incorporating UK-specific regulatory context. Required reading for UK financial institutions. Covers model definition, identification, risk tiering, validation, and ongoing monitoring requirements.
Office of the Comptroller of the Currency — Comptroller's Handbook: Model Risk Management (2021) The OCC's detailed guidance on implementing model risk management at national banks and federal savings associations. More operational than SR 11-7, with specific guidance on validation scope, documentation standards, and examiner expectations. The handbook is a practical supplement to the SR 11-7 guidance for US bank compliance teams.
Basel Committee on Banking Supervision — Principles for the Sound Management of Operational Risk (2011) While predating the modern ML era, these principles provide the conceptual foundation for treating model risk as a component of operational risk — which determines how it is capitalized, reported, and escalated within the risk management framework of a Basel-regulated financial institution.
Tools and Technical Resources
SHAP Library Documentation and GitHub Repository The primary technical reference for the SHAP Python library. Includes documentation for TreeSHAP, LinearSHAP, KernelSHAP, and DeepSHAP explainers; example notebooks covering credit scoring, fraud detection, and NLP applications; and the full API reference. Available at: shap.readthedocs.io and github.com/slundberg/shap
DICE Library (Diverse Counterfactual Explanations) The Python implementation of the DICE counterfactual explanation library from Microsoft Research. Supports tabular data with continuous and categorical features, includes feasibility constraints, and integrates with scikit-learn models. Available at: github.com/interpretml/DiCE
InterpretML — Microsoft's Open-Source Interpretability Toolkit InterpretML provides an integrated framework for both glassbox models (Explainable Boosting Machines) and blackbox explanation methods including SHAP, LIME, partial dependence plots, and Morris sensitivity analysis. The Explainable Boosting Machine (EBM) is particularly relevant for regulated financial services as it provides near-gradient-boosting performance with full model transparency. Available at: github.com/interpretml/interpret
Fairlearn — Python Library for Assessing and Improving Fairness Microsoft's open-source library for assessing and improving fairness in machine learning. Provides implementations of demographic parity, equalized odds, equal opportunity, and other fairness metrics; includes the Fairness Dashboard for interactive fairness assessment; and provides mitigation algorithms including reweighting, threshold optimization, and adversarial debiasing. Available at: github.com/fairlearn/fairlearn and fairlearn.org
Practitioner Guides
European Banking Authority — Report on Big Data and Advanced Analytics (EBA/REP/2020/01) The EBA's assessment of the use of big data and advanced analytics in the banking sector. Covers the governance challenges of ML models in credit, operational risk, and compliance contexts. Identifies the explainability gap as a primary concern and discusses proportionate supervisory expectations for different model risk levels.
Financial Stability Board — Artificial Intelligence and Machine Learning in Financial Services (2017) An early but still relevant assessment of AI/ML in financial services from the FSB's perspective, covering both the potential benefits and the systemic and governance risks. The governance section identifies model risk management, opacity of AI systems, and potential for correlated model errors across institutions as key concerns.
Institute of International Finance — Machine Learning in Credit Risk (2019) A practitioner-focused review of ML adoption in credit risk management, covering model development practices, validation challenges, and regulatory engagement strategies. Includes case studies from member institutions and practical guidance on building explainability into ML credit models.