Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Testing the Reliability of ChatGPT Assistance for Surgical Choices in Challenging Glaucoma Cases
9
Zitationen
6
Autoren
2025
Jahr
Abstract
<b>Background:</b> This study's aim is to assess ChatGPT's capability to analyze detailed case descriptions in glaucomatous patients and suggest the best possible surgical treatment. <b>Methods:</b> We conducted a retrospective analysis of 60 medical records of surgical glaucoma cases, divided into "ordinary" cases (<i>n</i> = 40) and "challenging" cases (<i>n</i> = 20). We entered every case description into ChatGPT-3.5's interface and inquired "What kind of surgery would you perform?". The frequency of accurate surgical choices made by ChatGPT, compared to those reported in patients' files, was reported. Furthermore, we assessed the level of agreement with three senior glaucoma surgeons, asked to analyze the same 60 cases and outline their surgical choices. <b>Results:</b> Overall, ChatGPT surgical choices were consistent with those reported in patients' files in 47/60 cases (78%). When comparing ChatGPT choices with the three glaucoma specialists, levels of agreement were 75%, 70%, and 83%, respectively. In ordinary cases, we did not report any significant differences when comparing ChatGPT answers with those of the three glaucoma specialists, when both of them were matched with patients' files (<i>p</i> > 0.05 for all). ChatGPT's performances were lower in "challenging" cases: when compared to patients' files, the accuracy was 13/20 (65%); when compared to glaucoma specialists, the level of agreement was 50%, 40%, and 70%, respectively. <b>Conclusion:</b> In ordinary conditions, ChatGPT was able to propose coherent personalized treatment plans, and its performance was comparable to that of skilled glaucoma specialists but showed its limitations in the evaluation of more complex cases.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.239 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.095 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.463 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.428 Zit.