Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Assessing the Value of Explainable Artificial Intelligence for Magnetic Resonance Imaging
0
Zitationen
5
Autoren
2025
Jahr
Abstract
Abstract Recent advancements in Artificial Intelligence (AI) have often improved the accuracy of medical diagnostics in several fields, such as cancer detection and diagnosing cardiovascular or neuromuscular diseases. However, the opaque nature of AI decision-making can limit its adoption in the clinical setting, as physicians require clear and interpretable explanations to trust these tools. To address this issue, the field of eXplainable Artificial Intelligence (XAI) aims to clarify the rationale behind AI predictions while ensuring compliance with ethical standards and advanced regulations such as the GDPR and the AI Act. This study applies multiple explainability methods to a diagnostic support model for Distal Myopathies (DMs), a rare neuromuscular disorder marked by subtle, early-stage tissue alterations. Beyond classification, our approach generates detailed explanations for the model’s predictions. We propose novel techniques, including a hierarchical occlusion method and an ensemble framework that combines individual explanations to produce refined, interpretable visualizations. Feedback from expert radiologists is used to assess the effectiveness of these methods, highlighting their potential to enhance trust and usability in clinical practice. Our results show that pretrained convolutional networks achieve high classification accuracy, exceeding 88%, with perfect recall in identifying affected cases, while underscoring the need for adaptive and user-centric approaches to explainability in AI-driven diagnostic tools.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.644 Zit.
Generative Adversarial Nets
2023 · 19.894 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.313 Zit.
"Why Should I Trust You?"
2016 · 14.504 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.186 Zit.