Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
4128 Evaluating the impact of artificial intelligence-assisted image analysis on the diagnostic accuracy of front-line clinicians in detecting fractures on plain X-rays (FRACT-AI)
0
Zitationen
9
Autoren
2026
Jahr
Abstract
<h3>Aims and Objectives</h3> Incorrect fracture diagnosis is the most frequent diagnostic error in UK emergency departments, causing significant patient morbidity and cost to the NHS. This study aimed to test the impact of an artificial intelligence (AI)-assisted fracture detection tool on NHS clinicians’ diagnostic performance. <h3>Method and Design</h3> A dataset of 500 plain radiographs from Oxford University Hospitals was curated based on the frequency of fracture detection claims reported by NHS Resolution, with 250 images containing one or more fractures and 250 without. Ground truth was established through independent reporting by two senior musculoskeletal radiologists, with a third arbitrating disagreement. Images were subsequently inferenced by fracture detection software (Gleamer BoneView). A multicase multireader study was conducted including 18 clinicians of varying seniority from six clinical specialties. Readers interpreted all images, then repeated the reads after a four-week washout period with AI assistance. Changes in diagnostic performance, confidence, and reporting speed were compared, along with the diagnostic performance of the algorithm against ground truth. <h3>Results and Conclusion</h3> Pooled analyses for per-case reader performance demonstrated an increase in the area under the receiver operating curve from 0.883 (95% CI 0.858–0.907) without AI to 0.921 (95% CI 0.901–0.942) with AI (p < 0.001). Sensitivity improved from 0.828 (95% CI 0.788–0.868) to 0.867 (95% CI 0.827–0.906) with AI (p < 0.001), and specificity increased from 0.829 (95% CI 0.784–0.875) to 0.904 (95% CI 0.877–0.931) with AI (p < 0.001).AI-assisted fracture detection showed potential to improve clinician and radiologist accuracy on plain X-rays. Further real-world prospective studies are needed to demonstrate clinical efficacy in urgent care settings.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.545 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.436 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.935 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.589 Zit.