Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
S175 Evaluation of ChatGPT in Risk Stratification and Management of Choledocholithiasis
0
Zitationen
7
Autoren
2024
Jahr
Abstract
Introduction: Diagnosing choledocholithiasis often presents as a challenge requiring accurate risk stratification and management. Artificial Intelligence tools like ChatGPT have potential to assist in clinical decision-making, but their effectiveness remains underexplored. This study aims to evaluate the accuracy of ChatGPT in risk stratification and management of choledocholithiasis. We compared ChatGPT's recommendations, both with and without the provision of ASGE guidelines for choledocholithiasis, against the actual clinical decisions made by our GI team. Methods: We retrospectively analyzed 100 cases of choledocholithiasis, providing ChatGPT with de-identified clinical information. ChatGPT was tested in 2 scenarios: with and without access to the ASGE guidelines. We compared ChatGPT's risk stratification (low, intermediate, high) and management recommendations (Cholecystectomy vs MRCP vs ERCP) against the actual decisions made by the GI team. Results: ChatGPT trained with guidelines had higher precision (87%) in risk stratification compared to the untrained version (62%), closely aligning with the GI team's assessments (R=0.726, P< 0.001 vs R=0.471, P< 0.001). For management, trained ChatGPT's recommendations showed a lower precision (58%) compared to the untrained version (62%) (R=0.335, P< 0.001 vs R=0.199, P=0.047). Among intermediate-risk patients who underwent ERCP, the GI physician's decisions more frequently resulted in finding a CBD stone (62.1%) than the untrained ChatGPT (41.4%). Overall, while trained ChatGPT demonstrated better alignment in risk assessment, untrained ChatGPT demonstrated slightly better management. Conclusion: Although trained ChatGPT demonstrated high precision in risk stratification, it struggled to identify certain high-risk patients with ascending cholangitis. Both versions occasionally misread clinical information, impacting risk categorization. Untrained ChatGPT's management was slightly more aligned with the GI team, although its risk stratification appeared random and lacked clear rationale. Among intermediate-risk patients who underwent ERCP, the GI physician's decisions more frequently resulted in finding a CBD stone, indicative of the crucial role of clinical acumen and experience in patient evaluation. These findings suggest that while AI tools can augment clinical decision-making, their integration should be approached cautiously, ensuring thorough validation. Further research is necessary to optimize these tools for reliable clinical application.
Ähnliche Arbeiten
Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015: elaboration and explanation
2015 · 12.897 Zit.
CONSORT 2010 Explanation and Elaboration: updated guidelines for reporting parallel group randomised trials
2010 · 8.934 Zit.
Goodman and Gilman's the Pharmacological Basis of Therapeutics
1986 · 6.801 Zit.
1,500 scientists lift the lid on reproducibility
2016 · 4.170 Zit.
Heart Disease: A Textbook of Cardiovascular Medicine.
1984 · 4.135 Zit.