Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Clinical Verification of AI Models for Radiology Report Annotation Using a Physician-Annotated MIMIC-CXR
0
Zitationen
6
Autoren
2025
Jahr
Abstract
Accurate evaluation of AI-generated radiology reports is crucial for clinical applications. This study investigates the use of large language models (LLMs) - specifically Google Gemini and OpenAI GPT-4o - for this task. We compare their performance against CheXpert in annotating both raw radiology reports and reports generated by image-to-report systems (MedRAX, RGRG), using a physician-validated subset of MIMIC-CXR as ground truth. Our evaluation demonstrates Gemini’s strong performance in this context. This work provides a clinical validation of LLMs against expert annotations, highlighting their potential for medical text processing and the challenges in accurately identifying rare findings.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.291 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.143 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.535 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.452 Zit.