Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Hierarchical Reinforcement Learning for Detecting Safety and Reliability Vulnerabilities in Large Language Model-Assisted Healthcare Systems
0
Zitationen
5
Autoren
2025
Jahr
Abstract
The integration of large language models into the telemedicine industry has revolutionized digital health by promoting effective clinical decision-making and patient interactions. However, the deployment of LLMS in healthcare imposes critical risks to health data. Moreover, the existing standards of vulnerability detection lack the practical approach to dynamic and enhanced interactions in context-dependent healthcare systems. To overcome these potential hurdles, this paper presents a Hierarchical Reinforcement Learning (HRL) framework explicitly designed to detect and mitigate vulnerabilities in LLM-assisted healthcare applications. The proposed HRL model decomposes clinical interactions into hierarchical tasks, enabling efficient modeling of complex, multi-step clinical reasoning and nuanced conversational dynamics. Our approach integrates safety-aware reward engineering and policy trajectory analysis to systematically identify risky LLM behaviors. Experimental validation using realistic clinical scenarios demonstrates that our HRLbased approach significantly outperforms conventional methods, providing robust detection of safety and reliability vulnerabilities in healthcare LLM systems.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.338 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.418 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.303 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.301 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.499 Zit.