Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Explainable AI Models on Radiographic Images Integrated with Clinical Measurements: Prediction for Unstable Hips in Infants
0
Zitationen
10
Autoren
2024
Jahr
Abstract
<title>Abstract</title> Considering explainability is crucial in medical artificial intelligence, technologies to quantify Grad-CAM heatmaps and perform automatic integration based on domain knowledge remain lacking. Hence, we created an end-to-end model that produced CAM scores on regions of interest (CSoR), a measure of relative CAM activity, and feature importance scores by automatic algorithms for clinical measurement (aaCM) followed by LightGBM. In this multicenter research project, the diagnostic performance of the model was investigated with 813 radiographic hip images in infants at risk of unstable hips, with the ground truth defined by provocative examinations. The results indicated that the accuracy of aaCM was higher than that of specialists, and the model with ad hoc adoption of aaCM outperformed the image-only-based model. Subgroup analyses in positive cases indicated significant differences in CSoR between the unstable and contralateral sides despite containing only binary labels (positive or negative). In conclusion, aaCM reinforces the performance, and CSoR potentially indicates model reliability.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.260 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.116 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.493 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.438 Zit.