Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Enhancing Clinical Decision Support: A Hybrid Framework Integrating Markov Decision Processes and Explainable Neurosymbolic Reasoning
0
Zitationen
1
Autoren
2026
Jahr
Abstract
The integration of Artificial Intelligence (AI) into healthcare promises to revolutionize diagnostic accuracy and treatment optimization. However, the deployment of these systems faces significant hurdles related to trust, explainability, and the ability to reason through complex, sequential clinical scenarios. Current "black-box" deep learning models often fail to provide the transparency required for medical and legal accountability, while traditional rulebased systems lack the flexibility to learn from vast biomedical datasets. This paper proposes a hybrid framework that combines Markov Decision Processes (MDPs) for temporal treatment planning with Neurosymbolic AI to integrate logical reasoning with data-driven learning. By prioritizing explainability and the rigorous management of uncertainty, this approach aims to support clinicians in making optimal, evidence-based decisions. We discuss the theoretical underpinnings of this method, analyse its relation to existing literature, and outline a robust evaluation strategy to measure both performance and clinical helpfulness.
Ähnliche Arbeiten
"Why Should I Trust You?"
2016 · 14.643 Zit.
Coding Algorithms for Defining Comorbidities in ICD-9-CM and ICD-10 Administrative Data
2005 · 10.535 Zit.
A Comprehensive Survey on Graph Neural Networks
2020 · 8.902 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.470 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.984 Zit.