Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
STAGER checklist: Standardized testing and assessment guidelines for evaluating generative artificial intelligence reliability
0
Zitationen
35
Autoren
2024
Jahr
Abstract
Generative artificial intelligence (AI) holds immense potential for medical applications, but the lack of a comprehensive evaluation framework and methodological deficiencies in existing studies hinder its effective implementation. Standardized assessment guidelines are crucial for ensuring reliable and consistent evaluation of generative AI in healthcare. Our objective is to develop robust, standardized guidelines tailored for evaluating generative AI performance in medical contexts. Through a rigorous literature review utilizing the Web of Sciences, Cochrane Library, PubMed, and Google Scholar, we focused on research testing generative AI capabilities in medicine. Our multidisciplinary team of experts conducted discussion sessions to develop a comprehensive 32‐item checklist. This checklist encompasses critical evaluation aspects of generative AI in medical applications, addressing key dimensions such as question collection, querying methodologies, and assessment techniques. The checklist and its broader assessment framework provide a holistic evaluation of AI systems, delineating a clear pathway from question gathering to result assessment. It guides researchers through potential challenges and pitfalls, enhancing research quality and reporting and aiding the evolution of generative AI in medicine and life sciences. Our framework furnishes a standardized, systematic approach for testing generative AI's applicability in medicine. For a concise checklist, please refer to Table S or visit GenAIMed.org . Highlights This work formulates the standardized testing and assessment guidelines for evaluating generative artificial intelligence (AI) reliability (STAGER) checklist, a 32‐item framework offering standardized assessment guidelines tailored for evaluating generative AI systems in medical and life science contexts. It consists of key aspects, including question collection, querying approaches, and assessment techniques. It enhances research quality and facilitates advances in this emerging field.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.214 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.071 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.429 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.418 Zit.
Autoren
- Jinghong Chen
- Lingxuan Zhu
- Weiming Mou
- Anqi Lin
- Dongqiang Zeng
- Chang Qi
- Zaoqu Liu
- Aimin Jiang
- Bufu Tang
- Wenjie Shi
- Ulf D. Kahlert
- Jian‐Guo Zhou
- Shipeng Guo
- Xiaofan Lu
- Xu Sun
- Trunghieu Ngo
- Zhongji Pu
- Baolei Jia
- Che Ok Jeon
- Yongbin He
- Haiyang Wu
- Shuqin Gu
- Wisit Cheungpasitporn
- Haojie Huang
- Weipu Mao
- Shixiang Wang
- Xin Chen
- Loïc Cabannes
- Gerald Sng Gui Ren
- Iain S. Whitaker
- Stephen R Ali
- Cheng Quan
- Kai Miao
- Shuofeng Yuan
- Peng Luo