Further Reading: Transparency, Explainability, and the Black Box Problem
The sources below provide deeper engagement with the themes introduced in Chapter 16. They are organized by topic and include foundational technical papers on explainability methods, legal analyses of the right to explanation, critical perspectives on transparency, and practical guides for implementing explainable AI in high-stakes domains.
Explainability Methods: LIME and SHAP
Ribeiro, Marco Tulio, Sameer Singh, and Carlos Guestrin. "'Why Should I Trust You?': Explaining the Predictions of Any Classifier." Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135-1144. ACM, 2016. The original LIME paper. Ribeiro et al. introduce the method of explaining individual predictions by fitting a simple interpretable model to the black box's local behavior. The paper is accessible, well-illustrated, and includes experiments on text classification and image recognition. Essential reading for understanding both the power and the limitations of local, model-agnostic explanation. The paper's opening question — "why should I trust you?" — frames explainability as fundamentally about warranted trust rather than mere technical disclosure.
Lundberg, Scott M., and Su-In Lee. "A Unified Approach to Interpreting Model Predictions." Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS), 4765-4774. 2017. The foundational SHAP paper, which unifies several existing explanation methods under the framework of Shapley values from cooperative game theory. Lundberg and Lee demonstrate that SHAP satisfies three desirable properties — local accuracy, missingness, and consistency — that other methods lack. The paper is more technically demanding than the LIME paper but rewards careful reading. Students interested in the mathematical foundations of feature attribution should start here.
Molnar, Christoph. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. 2nd ed. Self-published, 2022. Available at: https://christophm.github.io/interpretable-ml-book/ The most accessible and comprehensive guide to interpretable machine learning methods. Molnar covers LIME, SHAP, partial dependence plots, permutation feature importance, counterfactual explanations, and many other methods — each with clear explanations, worked examples, and honest assessments of strengths and weaknesses. This freely available book is the single best resource for students who want to move from conceptual understanding to practical application of XAI methods.
The Black Box Problem: Foundational Arguments
Rudin, Cynthia. "Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead." Nature Machine Intelligence 1, no. 5 (2019): 206-215. A provocative and influential argument that the field's emphasis on post-hoc explanation methods (LIME, SHAP) is misguided. Rudin contends that for high-stakes decisions — criminal justice, healthcare, child welfare — society should require inherently interpretable models rather than trying to explain opaque ones after the fact. She argues that the accuracy-interpretability trade-off is often overstated and that interpretable models can achieve comparable performance for many practical tasks. This paper challenges a central assumption of the chapter and is essential reading for any student who wants to engage critically with the XAI paradigm.
Burrell, Jenna. "How the Machine 'Thinks': Understanding Opacity in Machine Learning Algorithms." Big Data & Society 3, no. 1 (2016): 1-12. Burrell identifies three distinct sources of opacity in algorithmic systems: intentional corporate secrecy, technical illiteracy among non-specialists, and the fundamental mismatch between mathematical optimization and human-scale reasoning. Her taxonomy is more sociologically grounded than most computer science treatments and is particularly valuable for understanding why the black box problem is not solely a technical challenge. The paper connects directly to the chapter's distinction between the "locked room" and the "locked safe."
Pasquale, Frank. The Black Box Society: The Secret Algorithms That Control Money and Information. Cambridge, MA: Harvard University Press, 2015. The book that gave the "black box" metaphor its current prominence in policy discourse. Pasquale examines how opaque algorithms in finance, search, and reputation scoring concentrate power and evade accountability. Written for a general audience, the book is less technical than the papers above but provides essential context for understanding why algorithmic transparency became a major public concern. Students interested in the political and economic dimensions of opacity should start here.
The Right to Explanation and GDPR
Goodman, Bryce, and Seth Flaxman. "European Union Regulations on Algorithmic Decision-Making and a 'Right to Explanation.'" AI Magazine 38, no. 3 (2017): 50-57. The paper that launched the "right to explanation" debate. Goodman and Flaxman argue that the GDPR, read holistically, creates a meaningful right to understand how automated decisions are made. Their expansive interpretation has been influential in both academic and policy circles, though it has also been contested. Accessible and clearly written, this is the best starting point for the legal debate explored in Case Study 1.
Wachter, Sandra, Brent Mittelstadt, and Luciano Floridi. "Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation." International Data Privacy Law 7, no. 2 (2017): 76-99. The most rigorous counterargument to the "right to explanation" claim. Wachter, Mittelstadt, and Floridi conduct a detailed textual and structural analysis of the GDPR, concluding that the regulation creates only a right to general information about automated processing — not a right to a specific explanation of individual decisions. Reading this alongside Goodman and Flaxman provides students with both sides of a debate that remains unresolved in European law.
Selbst, Andrew D., and Julia Powles. "Meaningful Information and the Right to Explanation." International Data Privacy Law 7, no. 4 (2017): 233-242. A middle-ground contribution that argues the GDPR's requirement for "meaningful information about the logic involved" can be interpreted to require substantive explanations — if regulators and courts insist on meaningful rather than pro forma compliance. Selbst and Powles propose a functional approach: the adequacy of an explanation should be measured by whether it enables the data subject to exercise their rights effectively. This paper bridges the gap between the expansive and restrictive readings.
Transparency in Practice
Mitchell, Margaret, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. "Model Cards for Model Reporting." Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT)*, 220-229. ACM, 2019. The paper that introduced model cards — standardized documentation for machine learning models that includes intended uses, performance metrics across demographic groups, limitations, and ethical considerations. Model cards represent one of the most concrete proposals for operationalizing algorithmic transparency. The paper is essential for understanding both the promise and the limits of documentation-based transparency — and for evaluating whether model cards constitute meaningful transparency or risk becoming transparency theater.
Ananny, Mike, and Kate Crawford. "Seeing without Knowing: Limitations of the Transparency Ideal and Its Application to Algorithmic Accountability." New Media & Society 20, no. 3 (2018): 973-989. A critical examination of the assumption that transparency automatically produces accountability. Ananny and Crawford identify ten limitations of the transparency ideal, including the observation that complex systems may be visible but not understandable, that transparency can be selectively deployed to serve institutional interests, and that disclosure without power to act on what is disclosed is empty. This paper provides essential conceptual grounding for the chapter's discussion of transparency theater.
Healthcare AI and Explainability
Topol, Eric J. Deep Medicine: How Artificial Intelligence Can Make Healthcare Human Again. New York: Basic Books, 2019. Topol, a cardiologist and genomics researcher, makes the case that AI can improve healthcare not by replacing physicians but by freeing them to focus on the human dimensions of medicine. The book covers AI applications across radiology, pathology, dermatology, and clinical decision support, with thoughtful attention to the explainability challenges in each domain. Accessible to non-specialists and deeply relevant to Case Study 2's examination of XAI in healthcare.
Sendak, Mark P., Madeleine Clare Elish, Michael Gao, Joseph Futoma, William Ratliff, Marshall Nichols, Armando Bedoya, Suresh Balu, and Cara O'Brien. "'The Human Body Is a Black Box': Supporting Clinical Decision-Making with Deep Learning." Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT)*, 99-109. ACM, 2020. A detailed account of deploying Sepsis Watch at Duke University Hospital, examining how clinicians interact with a deep learning system's predictions and explanations in real time. The paper is unusual in its attention to the sociotechnical dimensions of deployment — the organizational changes, workflow integration, and trust-building required to make XAI work in practice. Essential reading for understanding why technical explainability is necessary but not sufficient for effective AI deployment.
Emerging Directions
European Commission. "Regulation of the European Parliament and of the Council Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act)." 2024. The EU AI Act represents the most comprehensive attempt to regulate AI transparency through law. It classifies AI systems by risk level and imposes transparency requirements that go beyond the GDPR, including requirements for human oversight, technical documentation, and post-market monitoring of high-risk systems. Students should read at minimum Articles 13-14 (transparency and information provision) and the Annex III classification of high-risk AI systems. The AI Act will shape the global regulatory landscape for years to come and provides the legal context for the chapter's arguments about mandatory transparency.
Kaur, Harmanpreet, Harsha Nori, Samuel Jenkins, Rich Caruana, Hanna Wallach, and Jennifer Wortman Vaughan. "Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning." Proceedings of the 2020 CHI Conference on Human Factors in Computing, 1-14. ACM, 2020. A sobering empirical study of how data scientists actually use explainability tools. Kaur et al. found that practitioners frequently over-trusted and misinterpreted SHAP and other XAI outputs — confirming that producing an explanation is not the same as producing understanding. The paper challenges the assumption that better explanation tools will automatically lead to better decisions and highlights the need for training and critical evaluation of XAI outputs.
These readings extend Chapter 16's analysis of transparency, explainability, and the black box problem. As Part 3 concludes and gives way to Chapter 17's treatment of accountability and audit, the question evolves once more: transparency enables scrutiny, but who is responsible for ensuring that scrutiny occurs — and who bears the consequences when it does not?