Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluating the Clinical Utility of Artificial Intelligence Assistance and its Explanation on Glioma Grading Task
1
Zitationen
4
Autoren
2022
Jahr
Abstract
Abstract Background As a fast-advancing technology, artificial intelligence (AI) has considerable potential to assist physicians in various clinical tasks from disease identification to lesion segmentation. Despite much research, AI has not yet been applied to neurooncological imaging in a clinically meaningful way. To bridge the clinical implementation gap of AI in neuro-oncological settings, we conducted a clinical user-based evaluation, analogous to the phase II clinical trial, to evaluate the utility of AI for diagnostic predictions and the value of AI explanations on the glioma grading task. Method Using the publicly-available BraTS dataset, we trained an AI model of 88.0% accuracy on the glioma grading task. We selected the SmoothGrad explainable AI Weina Jin and Mostafa Fatehi are co-first authors. algorithm based on the computational evaluation regarding explanation truthfulness among a candidate of 16 commonly-used algorithms. SmoothGrad could explain the AI model’s prediction using a heatmap overlaid on the MRI to highlight important regions for AI prediction. The evaluation is an online survey wherein the AI prediction and explanation are embedded. Each of the 35 neurosurgeon participants read 25 brain MRI scans of patients with gliomas, and gave their judgment on the glioma grading without and with the assistance of AI’s prediction and explanation. Result Compared to the average accuracy of 82.5 ± 8.7% when physicians perform the task alone, physicians’ task performance increased to 87.7 ± 7.3% with statistical significance ( p -value = 0.002) when assisted by AI prediction, and remained at almost the same level of 88.5 ± 7.0% ( p -value = 0.35) with the additional AI explanation assistance. Conclusion The evaluation shows the clinical utility of AI to assist physicians on the glioma grading task. It also reveals the limitations of applying existing AI explanation techniques in clinical settings. Key points Phase II evaluation with 35 neurosurgeons on the clinical utility of AI and its explanation AI prediction assistance improved physicians’ performance on the glioma grading task Additional AI explanation assistance did not yield a performance boost Importance of the study This study is the first phase II AI clinical evaluation in neuro-oncology. Evaluating AI is a prerequisite for its clinical deployment. The four phases of AI clinical evaluation are analogous to the four phases of clinical trials. Prior works that apply AI in neurooncology utilize phase I algorithmic evaluation, which do not reflect how AI can be used in clinical settings to support physician decision making. To bridge the research gap, we conducted the first clinical evaluation to assess the joint neurosurgeon-AI task performance. The evaluation also includes AI explanation as an indispensable feature for AI clinical deployment. Results from quantitative and qualitative data analysis are presented for a detailed examination of the clinical utility of AI and its explanation.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.250 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.109 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.482 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.434 Zit.