Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
How AI Responds to Obstetric Ultrasound Questions and Analyzes and Explains Obstetric Ultrasound Reports: ChatGPT-3.5 vs. Microsoft Copilot in Bing
0
Zitationen
7
Autoren
2024
Jahr
Abstract
<title>Abstract</title> <bold>Objectives: </bold>To evaluate and compare the accuracy and consistency of answers to obstetric ultrasound questions and analysis of obstetric ultrasound reports using publicly available ChatGPT-3.5 and Microsoft Copilot in Bing (Copilot). <bold>Methods: </bold>Twenty questions related to obstetric ultrasound were answered and 110 obstetric ultrasound reports were analyzed by both ChatGPT-3.5 and Copilot, with each question and report being posed three times to them at different times. The accuracy and consistency of each response to twenty questions and each analysis result in the report were evaluated and compared. <bold>Results: </bold>In answering twenty questions, ChatGPT-3.5 outperformed Copilot in both accuracy (95.0% vs. 80.0%) and consistency (90.0% vs. 75.0%). When analyzing obstetric ultrasound reports, two models performed similarly in accuracy and consistency, and can provide recommendations. The overall accuracy and consistency of ChatGPT-3.5 and Copilot were 83.86%, 87.30% vs 77.51%, 90.48%, respectively. However, in detecting abnormal amniotic fluid index, ChatGPT-3.5 was superior to Copilot (accuracy 87.50% vs. 66.67%, <italic>P</italic> < 0.05). <bold>Conclusion: </bold>While ChatGPT-3.5 and Copilot can provide valuable explanations on obstetric ultrasound and interpret most obstetric ultrasound reports accurately, neither model consistently answered all questions correctly or with complete consistency, the supervision of physician is crucial in the use of these models.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.200 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.051 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.416 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.410 Zit.