Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
A Comparison of Explainable AI Models on Numeric and Graph-Structured Data
8
Zitationen
6
Autoren
2024
Jahr
Abstract
An exponential growth of interest in the healthcare IoT over the past few years has increased the adoption of AI. However, healthcare analytics demands highly accurate and reliable systems that can not afford even a small amount of vulnerability. The black-box nature of AI models raises significant concerns about the computations involved in these predictions, leading to a lack of trust in these models. Explainable Artificial Intelligence (XAI) seeks to resolve this problem by providing a human-understandable explanation for AI decisions, bringing transparency, trust, and fairness to these AI models. This paper evaluates the existing XAI models on numeric and graph-structured healthcare data. The LIME and SHAP models are first used to explain the predictions in the numerical dataset of fetal health prediction. Second, graph-structured data is analyzed using GNN Explainer and PG Explainer. This evaluation suggests that XAI model interpretations benefit healthcare professionals and patients to trust AI predictions.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.929 Zit.
Generative Adversarial Nets
2023 · 19.896 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.356 Zit.
"Why Should I Trust You?"
2016 · 14.688 Zit.
Generative adversarial networks
2020 · 13.316 Zit.