Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Enhancing AI Decision-Making with Explainable Large Language Models (LLMs) in Critical Applications
10
Zitationen
6
Autoren
2025
Jahr
Abstract
LLMs are being integrated into critical environments, such as healthcare, finance, and autonomous systems, which raise serious transparency, trust, and accountability concerns. Despite LLMs showing impressive capabilities with decision-making, the "black-box" nature of LLMs frequently limits LLM usage in high-stakes environments. Explainable AI (XAI) approaches seek to fill this gap to facilitate better understandability of decisions made by an LLM, which is an important aspect for end user trust, verification of compliance with regulations. Attention Mechanisms, Post-Hoc Explanations, Rule-based Reasons, etc. We examine existing problems, present a foundation for explainable LLM decision-making, and consider the implications of these approaches for deployment in the real world. Our research highlights the need to balance performance of model output with transparency for ethical and effective AI implementation across important industries.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.366 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.244 Zit.
"Why Should I Trust You?"
2016 · 14.255 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.122 Zit.