Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
ChatGPT becomes an Oncologist: the performance of Artificial Intelligence in the American Society of Clinical Oncology Evaluation Program. (Preprint)
0
Zitationen
9
Autoren
2023
Jahr
Abstract
<sec> <title>BACKGROUND</title> Importance: ChatGPT is a state-of-the-art large language model that uses artificial intelligence (AI) to address questions across diverse topics. The American Society of Clinical Oncology Self-Evaluation (ASCO-SEP) program created a comprehensive educational program to help physicians keep up-to-date with the many rapid advances in the field. The question bank consists of multiple-choice questions (MCQs) addressing the many facets of cancer care, including diagnosis, treatment, and supportive care. </sec> <sec> <title>OBJECTIVE</title> As ChatGPT applications rapidly expand, we sought to investigate its performance in the field of medical oncology by using questions from ASCO-SEP. </sec> <sec> <title>METHODS</title> We conducted a systematic assessment of the performance of ChatGPT-3 on the American Society of Clinical Oncology Self-Evaluation Program (ASCO-SEP), the leading educational and assessment tool for medical oncologists in training and practice. Over 1000 multiple choice questions covering the spectrum of cancer care where extracted. Questions were categorized by cancer type/discipline, with sub-categorization as treatment, diagnosis or other. Answers were scored as correct if ChatGPT selected the answer as defined by ASCO-SEP. </sec> <sec> <title>RESULTS</title> Overall, ChatGPT achieved a score of 56% for correct answers provided (583/1040). The program demonstrated varying levels of accuracy across cancer types/disciplines. The highest accuracy was observed in questions related to developmental therapeutics (8/10; 80% correct), while the lowest accuracy was observed in questions related to gastrointestinal cancer (102/209; 49% correct). There was no significant difference in the program’s performance across the pre-defined sub-categories of diagnosis, treatment and other (p value = .16 > 0.05). </sec> <sec> <title>CONCLUSIONS</title> Although below the required passing rate, ChatGPT’s performance on the ASCO-SEP showed promise for future applications in cancer care and medical education. Current limitations of the technology include training data that does not extend beyond 2021, and the inability to process or interpret data tables or images. However, as the technology continues to evolve, it is expected that these limitations will be overcome, allowing for improved capabilities. </sec> <sec> <title>CLINICALTRIAL</title> Non applicable </sec>
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.312 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.169 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.564 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.466 Zit.