Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Trust-Aware and Explainable Generative AI Frameworks for Interpretable Large Language Models in Critical Applications
0
Zitationen
2
Autoren
2026
Jahr
Abstract
The rapid adoption of Large Language Models (LLMs) across healthcare, finance, and governance has amplified concerns regarding explainability, accountability, and regulatory compliance. While generative AI systems demonstrate remarkable performance in language understanding and decision support, their opaque architectures and probabilistic reasoning processes hinder trust in high-stakes applications. This paper proposes a comprehensive framework titled ET-GEN (Explainable and Trustworthy Generative Network), designed to integrate interpretability mechanisms, uncertainty quantification, and governance-aligned evaluation metrics within LLM-driven decision systems. The framework incorporates attention attribution mapping, counterfactual reasoning modules, confidence calibration layers, and domain-specific rule alignment constraints. Experimental validation across healthcare diagnosis summarization, financial risk assessment, and public policy classification tasks demonstrates improved interpretability scores, calibrated confidence estimation, and regulatory alignment compared to baseline LLM systems. The results indicate that embedding explainability layers within generative architectures enhances transparency without significantly degrading predictive performance. The proposed model offers a scalable pathway for deploying trustworthy generative AI systems in regulated environments.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.299 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.236 Zit.
"Why Should I Trust You?"
2016 · 14.198 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.098 Zit.