Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Ein externer Link zum Volltext ist derzeit nicht verfügbar.
Large Language Models (LLMS) for Clinical Note Generation: International Classification of Disease (ICD) Code, Knowledge Graph (KG) and Prompt Evaluation
0
Zitationen
1
Autoren
2026
Jahr
Abstract
In the past decade, a surge in the amount of electronic health record (EHR) data in the United States occurred, driven by a favorable policy environment created by the Health Information Technology for Economic and Clinical Health (HITECH) Act of 2009 and the 21st Century Cures Act of 2016. Clinical notes for patients’ assessments, diagnoses, and treatments are captured in these EHRs in free-form text by physicians, who spend a considerable amount of time entering them. Manually writing these notes is time-consuming, increasing patient waiting times and potentially delaying diagnoses. Large language models (LLMs), such as GPT-4o, possess the ability to generate news articles that closely resemble human-written ones. In this work, we present several Chain-of-Thought (CoT) prompt engineering strategies that improve the LLM’s response in clinical note generation. In our prompts, we incorporate International Classification of Diseases (ICD) codes and basic patient information along with similar clinical case examples which effectively enhance the LLMs to formulate clinical notes. We evaluated our CoT prompt strategies on six clinical cases from the CodiEsp test dataset against several LLMs and our results show that it outperformed the standard one-shot prompt.
Ähnliche Arbeiten
"Why Should I Trust You?"
2016 · 14.688 Zit.
Coding Algorithms for Defining Comorbidities in ICD-9-CM and ICD-10 Administrative Data
2005 · 10.544 Zit.
A Comprehensive Survey on Graph Neural Networks
2020 · 8.925 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.504 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.025 Zit.