OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 20.03.2026, 13:37

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Don’t be my Doctor! Recognizing Healthcare Advice in Large Language Models

2024·1 ZitationenOpen Access
Volltext beim Verlag öffnen

1

Zitationen

5

Autoren

2024

Jahr

Abstract

Large language models (LLMs) have seen increasing popularity in daily use, with their widespread adoption by many corporations as virtual assistants, chatbots, predictors, and many more.Their growing influence raises the need for safeguards and guardrails to ensure that the outputs from LLMs do not mislead or harm users.This is especially true for highly regulated domains such as healthcare, where misleading advice may influence users to unknowingly commit malpractice.Despite this vulnerability, the majority of guardrail benchmarking datasets do not focus enough on medical advice specifically.In this paper, we present the HeAL benchmark (HEalth Advice in LLMs) 1 , a health-advice benchmark dataset that has been manually curated and annotated to evaluate LLMs' capability in recognizing health-advice -which we use to safeguard LLMs deployed in industrial settings.We use HeAL to assess several models and report a detailed analysis of the findings.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Artificial Intelligence in Healthcare and EducationElectronic Health Records SystemsChronic Disease Management Strategies
Volltext beim Verlag öffnen