Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Understanding Explainability in Medical Imaging
0
Zitationen
6
Autoren
2025
Jahr
Abstract
The integration of artificial intelligence (AI) into medical imaging has revolutionized healthcare offering unparalleled advancements in diagnosis, treatment planning, and patient care. However, alongside these advancements arises the critical need for transparency and accountability in AI-driven systems. This introduction explores the concept of explainability within the context of medical imaging emphasizing its vital role in fostering trust, acceptance, and widespread adoption of AI technologies in healthcare. Explainability in medical imaging refers to the ability of AI models or systems to provide understandable justifications for their decisions or predictions. It entails elucidating the underlying rationale behind AI outputs enabling clinicians and stakeholders to comprehend, trust, and validate the insights offered by these AI-enabled tools. By bridging the gap between the complex methodologies of AI and the interpretability requirements of clinical decision making, explainability serves as a crucial conduit for enhancing transparency and facilitating informed decision making in healthcare settings.
Ähnliche Arbeiten
New response evaluation criteria in solid tumours: Revised RECIST guideline (version 1.1)
2008 · 28.795 Zit.
TNM Classification of Malignant Tumours
1987 · 16.123 Zit.
A survey on deep learning in medical image analysis
2017 · 13.500 Zit.
Reduced Lung-Cancer Mortality with Low-Dose Computed Tomographic Screening
2011 · 10.736 Zit.
The American Joint Committee on Cancer: the 7th Edition of the AJCC Cancer Staging Manual and the Future of TNM
2010 · 9.101 Zit.