Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
The Necessity of Explainable Artificial Intelligence: A Problem-Driven Perspective
0
Zitationen
2
Autoren
2025
Jahr
Abstract
As artificial intelligence systems are increasingly used in important areas such as healthcare, finance, and security, the need for transparency in their decision-making has become critical. Many modern AI models function as black boxes, producing accurate results but offering no insight into how those results were reached. This lack of clarity can create serious risks, including reduced trust, unfair outcomes, and difficulties in identifying errors. Explainable AI (XAI) addresses these concerns by making model behaviour understandable to humans. This paper explains why explainability is essential, focusing on real-world risks, accountability requirements, and the need for safe and reliable AI systems. Instead of reviewing specific methods, the paper presents a problem-driven perspective that highlights explainability as a foundational requirement for responsible AI adoption.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.463 Zit.
Generative Adversarial Nets
2023 · 19.843 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.259 Zit.
"Why Should I Trust You?"
2016 · 14.314 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.138 Zit.