Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Roadmap of Designing Cognitive Metrics for Explainable Artificial\n Intelligence (XAI)
15
Zitationen
5
Autoren
2021
Jahr
Abstract
More recently, Explainable Artificial Intelligence (XAI) research has shifted\nto focus on a more pragmatic or naturalistic account of understanding, that is,\nwhether the stakeholders understand the explanation. This point is especially\nimportant for research on evaluation methods for XAI systems. Thus, another\ndirection where XAI research can benefit significantly from cognitive science\nand psychology research is ways to measure understanding of users, responses\nand attitudes. These measures can be used to quantify explanation quality and\nas feedback to the XAI system to improve the explanations. The current report\naims to propose suitable metrics for evaluating XAI systems from the\nperspective of the cognitive states and processes of stakeholders. We elaborate\non 7 dimensions, i.e., goodness, satisfaction, user understanding, curiosity &\nengagement, trust & reliance, controllability & interactivity, and learning\ncurve & productivity, together with the recommended subjective and objective\npsychological measures. We then provide more details about how we can use the\nrecommended measures to evaluate a visual classification XAI system according\nto the recommended cognitive metrics.\n
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.336 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.241 Zit.
"Why Should I Trust You?"
2016 · 14.227 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.114 Zit.