Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Large language models for diabetes training: a prospective study
15
Zitationen
38
Autoren
2025
Jahr
Abstract
Diabetes poses a considerable global health challenge, with varying levels of diabetes knowledge among healthcare professionals, highlighting the importance of diabetes training. Large Language Models (LLMs) provide new insights into diabetes training, but their performance in diabetes-related queries remains uncertain, especially outside the English language like Chinese. We first evaluated the performance of ten LLMs: ChatGPT-3.5, ChatGPT-4.0, Google Bard, LlaMA-7B, LlaMA2-7B, Baidu ERNIE Bot, Ali Tongyi Qianwen, MedGPT, HuatuoGPT, and Chinese LlaMA2-7B on diabetes-related queries, based on the Chinese National Certificate Examination for Primary Diabetes Care in China (NCE-CPDC) and the English Specialty Certificate Examination in Endocrinology and Diabetes of Membership of the Royal College of Physicians of the United Kingdom. Second, we assessed the training of primary care physicians (PCPs) without and with the assistance of ChatGPT-4.0 in the NCE-CPDC examination to ascertain the reliability of LLMs as medical assistants. We found that ChatGPT-4.0 outperformed other LLMs in the English examination, achieving a passing accuracy of 62.50%, which was significantly higher than that of Google Bard, LlaMA-7B, and LlaMA2-7B. For the NCE-CPFC examination, ChatGPT-4.0, Ali Tongyi Qianwen, Baidu ERNIE Bot, Google Bard, MedGPT, and ChatGPT-3.5 successfully passed, whereas LlaMA2-7B, HuatuoGPT, Chinese LLaMA2-7B, and LlaMA-7B failed. ChatGPT-4.0 (84.82%) surpassed all PCPs and assisted most PCPs in the NCE-CPDC examination (improving by 1 %-6.13%). In summary, LLMs demonstrated outstanding competence for diabetes-related questions in both the Chinese and English language, and hold great potential to assist future diabetes training for physicians globally.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.200 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.051 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.416 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.410 Zit.
Autoren
- Haoxuan Li
- Zehua Jiang
- Zhouyu Guan
- Yuqian Bao
- Yuexing Liu
- Tingting Hu
- Jiajia Li
- Ruhan Liu
- Liang Wu
- Di Cheng
- Hongwei Ji
- Yong Wang
- Ya Xing Wang
- Carol Y. Cheung
- Yingfeng Zheng
- Jihong Wang
- Zhen Li
- Weibing Wu
- Cynthia Ciwei Lim
- Yong Mong Bee
- Hong Chang Tan
- Elif I. Ekinci
- David C. Klonoff
- Justin B. Echouffo‐Tcheugui
- Nestoras Mathioudakis
- Leonor Corsino
- Rafael Simó
- Charumathi Sabanayagam
- Gavin Siew Wei Tan
- Ching‐Yu Cheng
- Tien Yin Wong
- Huating Li
- Chun Cai
- Lijuan Mao
- Lee‐Ling Lim
- Yih Chung Tham
- Bin Sheng
- Weiping Jia
Institutionen
- Shanghai University of Sport(CN)
- Affiliated Hospital of Qingdao University(CN)
- Qingdao University(CN)
- Chinese University of Hong Kong(CN)
- Sun Yat-sen University(CN)
- Singapore General Hospital(SG)
- Mills Peninsula Health Services(US)
- Duke University(US)
- Singapore Eye Research Institute(SG)
- Singapore National Eye Center(SG)