Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Integrating AI in Clinical Education: Evaluating General Practice Residents’ Proficiency in Distinguishing AI-Generated Hallucinations and Its Impacting Factors
0
Zitationen
12
Autoren
2024
Jahr
Abstract
<title>Abstract</title> <bold>OBJECTIVE</bold> To evaluate the ability of general practice residents to detect AI-generated hallucinations and assess the influencing factors.<bold>METHODS</bold> This multi-center study involved 142 general practice residents, all of whom were undergoing standardized general practice training and volunteered to participate. The study evaluated AI’s accuracy and consistency, along with the residents’ response time, accuracy, sensitivity(d’), and standard tendencies (β). Binary regression analysis was used to explore factors affecting the residents' ability to identify AI-generated errors.<bold>RESULTS</bold> 137 participants ultimately included had an mean (SD) age 25.93 ± 2.10, with 46.72% male, 81.75% undergraduates, and 45.26% from Jiangsu. Regarding AI, 52.55% were unfamiliar with it, 35.04% had never used it. ChatGPT demonstrated 80.8% overall accuracy, including 57% in professional practice. 87 AI-generated hallucinations were identified, primarily in the level of application and evaluation. The mean (SD) accuracy was 55% ±4.3%, and the mean (SD) sensitivity (d') was 0.39 ± 0.33. The median response bias (β) was 0.74 (0.31). Regression analysis revealed that shorter response times (OR = 0.92, P = 0.02), higher self-assessed AI understanding (OR = 0.16, P = 0.04), and frequent AI use (OR = 10.43, P = 0.01) were associated with stricter error detection criteria.<bold>CONCLUSIONS</bold> The study concluded that residents struggled to identify AI errors, particularly in clinical cases, emphasizing the importance of improving AI literacy and critical thinking for effective integration into medical education.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.239 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.095 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.463 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.428 Zit.