Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Explainable and Trustworthy Generative Al: A Framework for Interpretable Large Language Models in High-Stakes Decision Systems
0
Zitationen
1
Autoren
2026
Jahr
Abstract
The rapid adoption of large language models (LLMs) in high-stakes decision systems such as healthcare, finance, law, and public governance has raised critical concerns regarding transparency, reliability, and trustworthiness. While generative AI models demonstrate remarkable performance, their black-box nature limits interpretability and poses significant risks when decisions directly impact human lives. This paper proposes a comprehensive framework for Explainable and Trustworthy Generative AI, aimed at enhancing the interpretability, accountability, and robustness of large language models deployed in high-stakes environments.The proposed framework integrates intrinsic interpretability mechanisms, post-hoc explanation techniques, and uncertainty-aware reasoning to provide transparent model behavior at both global and local decision levels. Trustworthiness is further strengthened through bias detection, fairness auditing, robustness evaluation, and human-in-the-loop validation. Additionally, the framework incorporates governance-oriented components, including ethical compliance, auditability, and regulatory alignment, to support responsible AI deployment.Experimental evaluation across multiple high-stakes use cases demonstrates that the proposed approach significantly improves decision transparency and user trust while maintaining competitive predictive and generative performance. The results highlight the potential of interpretable generative AI systems to bridge the gap between model capability and real-world accountability. This work contributes toward the development of reliable, explainable, and ethically aligned large language models suitable for critical decision-making applications
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.311 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.238 Zit.
"Why Should I Trust You?"
2016 · 14.210 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.104 Zit.