Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Unveiling large multimodal models in pulmonary CT: A comparative assessment of generative AI performance in lung cancer diagnostics
1
Zitationen
18
Autoren
2025
Jahr
Abstract
Abstract Introduction: The emerging generative artificial intelligence (Gen‐AI) is increasingly recognized for its potential in healthcare, particularly in complex radiological interpretations. However, the clinical utility of Gen‐AI requires thorough validation with real‐world data. Method: This retrospective study analyzed chest computed tomography (CT) scans from 404 patients with lung conditions with lung neoplasms ( n = 184) and non‐malignancy ( n = 210), incorporating The Cancer Genome Atlas ( n = 106) and Medical Imaging and Data Resource Center ( n = 110) datasets as external validation. We evaluated diagnostic performance of three Gen‐AI models (GPT‐4‐turbo, Gemini‐pro‐vision, and Claude‐3‐opus) using receiver operating characteristic (ROC) analysis and chi‐square tests across various clinical scenarios. Likert scale scoring combined with response rate and variance analysis were employed to evaluate internal diagnostic tendencies, while Lasso and stepwise regression were externally introduced to optimize model performance. Results: In single‐image CT diagnostics, Gemini and Claude demonstrated superior accuracy compared to GPT. However, when additional CT slices or clinical histories were incorporated, the diagnostic accuracy of all models declined. ROC analysis indicated that Gen‐AI performance was limited but improved in simplified prompting environments or integration with machine learning methods. Feature analysis revealed that Gen‐AI primarily relied on morphology and margins for malignancy predictions, but struggled to recognize critical imaging features and occasionally fabricated data. Conclusions: Gen‐AI demonstrated variable potential for pulmonary CT imaging diagnosis across prompts and diagnostic environments of differing complexity. However, their limitations and risks in processing complex multimodal information highlight significant challenges in the integration of clinical information by existing models. Ongoing efforts to improve the robustness and reliability of these models are crucial for their successful adoption in healthcare.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.214 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.071 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.429 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.418 Zit.
Autoren
Institutionen
- Zhujiang Hospital(CN)
- Nanjing Medical University(CN)
- Zhuhai People's Hospital(CN)
- Jiangnan University(CN)
- Kangda College of Nanjing Medical University
- Southern Medical University(CN)
- Nantong University(CN)
- Taizhou People's Hospital(CN)
- Changhai Hospital(CN)
- Second Military Medical University(CN)
- University of Hong Kong(HK)
- Jinan University(CN)
- Shanghai Jiao Tong University(CN)
- Shanghai First People's Hospital(CN)
- Sixth Affiliated Hospital of Sun Yat-sen University(CN)
- South China University of Technology(CN)
- Nanfang Hospital(CN)
- Sun Yat-sen University(CN)
- Zhongshan Hospital(CN)
- The First Affiliated Hospital, Sun Yat-sen University(CN)
- Fudan University(CN)
- Quzhou City People's Hospital(CN)
- Quzhou University(CN)
- Wenzhou Medical University(CN)
- Qingdao University(CN)
- Affiliated Hospital of Qingdao University(CN)
- Central South University(CN)
- Xiangya Hospital Central South University(CN)