Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Assessment of Pathology Domain-Specific Knowledge of ChatGPT and Comparison to Human Performance
42
Zitationen
20
Autoren
2024
Jahr
Abstract
CONTEXT.—: Artificial intelligence algorithms hold the potential to fundamentally change many aspects of society. Application of these tools, including the publicly available ChatGPT, has demonstrated impressive domain-specific knowledge in many areas, including medicine. OBJECTIVES.—: To understand the level of pathology domain-specific knowledge for ChatGPT using different underlying large language models, GPT-3.5 and the updated GPT-4. DESIGN.—: An international group of pathologists (n = 15) was recruited to generate pathology-specific questions at a similar level to those that could be seen on licensing (board) examinations. The questions (n = 15) were answered by GPT-3.5, GPT-4, and a staff pathologist who recently passed their Canadian pathology licensing exams. Participants were instructed to score answers on a 5-point scale and to predict which answer was written by ChatGPT. RESULTS.—: GPT-3.5 performed at a similar level to the staff pathologist, while GPT-4 outperformed both. The overall score for both GPT-3.5 and GPT-4 was within the range of meeting expectations for a trainee writing licensing examinations. In all but one question, the reviewers were able to correctly identify the answers generated by GPT-3.5. CONCLUSIONS.—: By demonstrating the ability of ChatGPT to answer pathology-specific questions at a level similar to (GPT-3.5) or exceeding (GPT-4) a trained pathologist, this study highlights the potential of large language models to be transformative in this space. In the future, more advanced iterations of these algorithms with increased domain-specific knowledge may have the potential to assist pathologists and enhance pathology resident training.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.545 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.436 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.935 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.589 Zit.
Autoren
- Andrew Y. Wang
- Sherman Lin
- Christopher Tran
- Robert Homer
- Dan Wilsdon
- Joanna C. Walsh
- Emily A. Goebel
- Irene Sansano
- Snehal Sonawane
- Vincent Cockenpot
- Sanjay Mukhopadhyay
- Toros Taşkın
- Nusrat Zahra
- Luca Cima
- Orhan Semerci
- Birsen Gizem Özamrak
- Pallavi Mishra
- Naga Sarika Vennavalli
- Po-Hsuan Cameron Chen
- Matthew J. Cecchini
Institutionen
- Western University(CA)
- London Health Sciences Centre(CA)
- Yale University(US)
- Vall d'Hebron Hospital Universitari(ES)
- University of Illinois Chicago(US)
- Université Paris Sciences et Lettres(FR)
- Institut Curie(FR)
- Cleveland Clinic(US)
- Ağrı İbrahim Çeçen University(TR)
- Ospedale Santa Chiara(IT)
- Sağlık Bilimleri Üniversitesi(TR)
- Izmir University(TR)
- Lewisham and Greenwich NHS Trust(GB)
- Queen Elizabeth Hospital(GB)
- Yashoda Hospital(IN)