OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 04.04.2026, 08:35

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Errors in Generative AI-Based Patient-Centered Mental Health Documentation: Psychiatrists’ Qualitative Pre-Post Comparison (Preprint)

2025·1 Zitationen
Volltext beim Verlag öffnen

1

Zitationen

7

Autoren

2025

Jahr

Abstract

<sec> <title>BACKGROUND</title> Patients' digital access to their personal health data is becoming increasingly common worldwide. However, medical documentation often contains technical language and sensitive information, which can lead to potential misunderstandings and distress among patients. These issues may be particularly impactful in mental health contexts. Large language models (LLMs) offer a promising approach by transforming clinician-generated health notes into language that is more patient-centered, non medicalized, and empathetic. However, risks related to accuracy and clinical safety have not been adequately investigated in psychiatry. </sec> <sec> <title>OBJECTIVE</title> To qualitatively analyze the errors introduced by LLMs when transforming notes written by psychiatrists into patient-facing formats. The study will also highlight the implications for clinical communication and patient safety. </sec> <sec> <title>METHODS</title> Clinical notes (n = 63) written by 19 psychiatrists in an outpatient treatment setting were collected, anonymized, and translated from German to English by humans. OpenAI's GPT-3.5 Turbo was used to develop a pre-prompt that transformed these notes into a patient-centered, lay-readable form through an iterative process. Three psychiatrists qualitatively analyzed the LLM-revised documentation using Kuckartz content analysis. They compared the pre- and post-conversion notes to systematically identify and categorize LLM-induced errors. </sec> <sec> <title>RESULTS</title> Five categories of clinically relevant errors were identified: (1) clinical misinterpretations, particularly in critical assessments such as suicidality, where nuanced terminology was oversimplified or inaccurately represented; (2) attribution errors, where behaviors or roles within family dynamics or interactions were incorrectly attributed to different individuals; (3) content distortion errors were characterized by speculative additions, emotional exaggerations, and inappropriate contextual assumptions; (4) abbreviation and terminology errors resulted from inaccurate expansions of medical abbreviations and terms; and (5) structural and syntax errors resulted in ambiguity, particularly when the original notes were brief or bulleted. Despite significant improvements in the readability and overall linguistic fluency of the converted notes, these errors occurred. </sec> <sec> <title>CONCLUSIONS</title> LLMs have the potential to transform psychiatric notes into patient-friendly formats. However, critical errors remain prevalent and can impair clinical judgment, understanding of patient circumstances, clarity of medication regimens, and interpretation of clinical observations. To safely integrate AI-generated documentation into psychiatric care, clinician oversight and targeted model refinement are essential. Future research should explore strategies to mitigate these errors, assess their comprehensive clinical impact, and incorporate patient and provider perspectives to ensure robust implementation. </sec> <sec> <title>CLINICALTRIAL</title> German Clinical Trial Register, Registiration Number: DRKS00030188, URL: https://drks.de/search/en/trial/DRKS00030188 </sec>

Ähnliche Arbeiten