Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Explain, Embed, Retrieve, and Reason (E2R2): A SHAP-Informed LLM Framework for Decision Support
0
Zitationen
3
Autoren
2025
Jahr
Abstract
This paper introduces E2R2 (Explain, Embed, Retrieve, and Reason), a framework that combines SHAP-based feature attribution, case-based retrieval, and GPT-driven reasoning for explainable classification. Applied to student attrition prediction, E2R2 achieved 89.2% accuracy and 84.1 F1 on a 500-case holdout set, comparable to Decision Tree and Random Forest baselines while offering higher recall and balanced precision–recall. Validation showed 94% consistency across GPT sessions and resilience to incomplete data (accuracy = 86.6% under 10% feature dropout). Beyond predictive accuracy, E2R2 generates SHAP-grounded, peer-informed narratives that improve cognitive accessibility. Although demonstrated in higher education, the architecture is domain-agnostic and adaptable to fields such as healthcare or finance. By extending feature attributions into context-aware explanations, E2R2 exemplifies the design of next-generation decision support systems that combine analytic precision with interpretability.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.615 Zit.
Generative Adversarial Nets
2023 · 19.894 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.306 Zit.
"Why Should I Trust You?"
2016 · 14.446 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.171 Zit.