OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 08.05.2026, 06:50

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

A Comparison of Explainable AI Models on Numeric and Graph-Structured Data

2024·8 Zitationen·Procedia Computer ScienceOpen Access
Volltext beim Verlag öffnen

8

Zitationen

6

Autoren

2024

Jahr

Abstract

An exponential growth of interest in the healthcare IoT over the past few years has increased the adoption of AI. However, healthcare analytics demands highly accurate and reliable systems that can not afford even a small amount of vulnerability. The black-box nature of AI models raises significant concerns about the computations involved in these predictions, leading to a lack of trust in these models. Explainable Artificial Intelligence (XAI) seeks to resolve this problem by providing a human-understandable explanation for AI decisions, bringing transparency, trust, and fairness to these AI models. This paper evaluates the existing XAI models on numeric and graph-structured healthcare data. The LIME and SHAP models are first used to explain the predictions in the numerical dataset of fetal health prediction. Second, graph-structured data is analyzed using GNN Explainer and PG Explainer. This evaluation suggests that XAI model interpretations benefit healthcare professionals and patients to trust AI predictions.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Explainable Artificial Intelligence (XAI)Machine Learning in HealthcareArtificial Intelligence in Healthcare and Education
Volltext beim Verlag öffnen