Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluating ChatGPT-4 for the Interpretation of Images from Several Diagnostic Techniques in Gastroenterology
8
Zitationen
20
Autoren
2025
Jahr
Abstract
<b>Background:</b> Several artificial intelligence systems based on large language models (LLMs) have been commercially developed, with recent interest in integrating them for clinical questions. Recent versions now include image analysis capacity, but their performance in gastroenterology remains untested. This study assesses ChatGPT-4's performance in interpreting gastroenterology images. <b>Methods:</b> A total of 740 images from five procedures-capsule endoscopy (CE), device-assisted enteroscopy (DAE), endoscopic ultrasound (EUS), digital single-operator cholangioscopy (DSOC), and high-resolution anoscopy (HRA)-were included and analyzed by ChatGPT-4 using a predefined prompt for each. ChatGPT-4 predictions were compared to gold standard diagnoses. Statistical analyses included accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and area under the curve (AUC). <b>Results:</b> For CE, ChatGPT-4 demonstrated accuracies ranging from 50.0% to 90.0%, with AUCs of 0.50-0.90. For DAE, the model demonstrated an accuracy of 67.0% (AUC 0.670). For EUS, the system showed AUCs of 0.488 and 0.550 for the differentiation between pancreatic cystic and solid lesions, respectively. The LLM differentiated benign from malignant biliary strictures with an AUC of 0.550. For HRA, ChatGPT-4 showed an overall accuracy between 47.5% and 67.5%. <b>Conclusions:</b> ChatGPT-4 demonstrated suboptimal diagnostic accuracies for image interpretation across several gastroenterology techniques, highlighting the need for continuous improvement before clinical adoption.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.197 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.047 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.410 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.410 Zit.
Autoren
- Miguel Mascarenhas
- Tiago Ribeiro
- Belén Agudo
- João Afonso
- Francisco Mendes
- Miguel Martins
- Pedro Cardoso
- Joana Mota
- Maria João Almeida
- António Costa
- Mariano González-Haba
- Jessica Widmer
- Eduardo Guimarães Hourneaux de Moura
- Ahsan Javed
- Thiago da Silveira Manzione
- Sidney Roberto Nadal
- Luis F. Barroso
- Vincent de Parades
- Jo�ão José Pinto Ferreira
- Guilherme Macedo
Institutionen
- Universidade do Porto(PT)
- Hospital de São João(PT)
- Hospital Universitario Puerta de Hierro Majadahonda(ES)
- Island Hospital(US)
- Hospital das Clínicas da Faculdade de Medicina da Universidade de São Paulo(BR)
- University of Liverpool(GB)
- Royal Liverpool University Hospital(GB)
- Instituto de Infectologia Emílio Ribas(BR)
- Wake Forest University(US)
- Hôpital Paris Saint-Joseph(FR)