Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Machine learning models explanations as interpretations of evidence: a theoretical framework of explainability and its implications on high-stakes biomedical decision-making
0
Zitationen
8
Autoren
2025
Jahr
Abstract
Explainable Artificial Intelligence, or XAI, is a vibrant research topic in the artificial intelligence community. It is raising growing interest across methods and domains, especially those involving high-stakes decision-making, such as the biomedical sector. Much has been written about the subject, yet XAI still lacks shared terminology and a framework capable of providing structural soundness to explanations, a crucial need for decisions that impact healthcare. In our work, we address these issues by proposing a novel definition of explanation that synthesizes insights from the existing literature. We recognize that explanations are not atomic, but rather the combination of evidence stemming from the model and its input-output mapping, along with the human interpretation of this evidence. Furthermore, we fit explanations into the properties of faithfulness (i.e., the explanation is an accurate description of the model’s inner workings and decision-making process) and plausibility (i.e., how much the explanation seems convincing to the user). Our theoretical framework simplifies the operationalization of these properties and provides new insights into common explanation methods that we analyze through case studies. We explore the impact of our framework in the sensitive domain of biomedicine, where XAI can play a central role in generating trust by balancing faithfulness and plausibility.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.408 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.253 Zit.
"Why Should I Trust You?"
2016 · 14.286 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.132 Zit.