OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 14.03.2026, 13:38

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Towards Leveraging Large Language Models for Automated Medical Q&A Evaluation

2024·0 Zitationen·arXiv (Cornell University)Open Access
Volltext beim Verlag öffnen

0

Zitationen

5

Autoren

2024

Jahr

Abstract

This paper explores the potential of using Large Language Models (LLMs) to automate the evaluation of responses in medical Question and Answer (Q\&A) systems, a crucial form of Natural Language Processing. Traditionally, human evaluation has been indispensable for assessing the quality of these responses. However, manual evaluation by medical professionals is time-consuming and costly. Our study examines whether LLMs can reliably replicate human evaluations by using questions derived from patient data, thereby saving valuable time for medical experts. While the findings suggest promising results, further research is needed to address more specific or complex questions that were beyond the scope of this initial investigation.

Ähnliche Arbeiten

Autoren

Themen

Topic ModelingExpert finding and Q&A systemsArtificial Intelligence in Healthcare and Education
Volltext beim Verlag öffnen