Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Generative Large Language Models for Question Answering from Electronic Health Records: A Systematic Review
0
Zitationen
3
Autoren
2025
Jahr
Abstract
This systematic review synthesises evidence on the performance, methodological quality, and clinical applicability of generative large language models for question-answering tasks using electronic health record data. We searched PubMed, Scopus, Web of Science, IEEE Xplore, and ACM Digital Library for original empirical studies evaluating decoder-only or encoder-decoder transformer architectures applied to clinical question-answering over EHR data. Risk of bias was assessed using an adapted QUADAS-2 framework with signalling questions modified to address LLM-specific methodological concerns including data contamination, LLM-as-judge bias, and reproducibility. Due to anticipated heterogeneity in task definitions and evaluation metrics, findings are synthesised narratively rather than through meta-analysis. A preliminary clinical implementation decision framework is developed to translate research findings into actionable guidance for healthcare institutions evaluating these systems for deployment.
Ähnliche Arbeiten
"Why Should I Trust You?"
2016 · 14.478 Zit.
A Comprehensive Survey on Graph Neural Networks
2020 · 8.784 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.332 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.779 Zit.
Artificial intelligence in healthcare: past, present and future
2017 · 4.459 Zit.