Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Enabling Responsible, Secure and Sustainable Healthcare AI - A Strategic Framework for Clinical and Operational Impact
0
Zitationen
1
Autoren
2025
Jahr
Abstract
Healthcare AI will realize long-lasting clinical and operational improvements only if technical innovation can be combined with organizational strategy, governance, security, and continued adoption. We describe a five-pillar strategic model - Leadership & Strategy, MLOps & Technical Infrastructure, Governance & Ethics, Education & Workforce Development, and Change Management & Adoption - for responsible AI in healthcare. Unlike other guidance, our approach includes AI-specific security and lifecycle MLOps combined with cross-disciplinary governance and user-centered change management - aligning a “compliance-by-design” philosophy with emerging regulatory expectations. We demonstrate the framework in two practical applications: (A) a hospital service for inpatient length-of-stay prediction and (B) an AI-assisted radiology second-reader for lung nodules. The LOS model achieved R² ≈ 0.4–0.6 in pilot evaluation across cohorts and was used by >75% of case managers, with targeted units implementing 5–10% reductions in average LOS for complex-discharge patients. A radiology tool (sensitivity ≈ 95%) was embedded in PACS with thresholding and explanation overlays to decrease alert fatigue, resulting in an 8% increase in detection of sub-centimeter actionable findings and no significant read slowdown. Overall, AI services were run under monitored and auditable pipelines in both cases, with no reported security incidents, and user-facing model cards informed trust and appropriate use. These findings show that robust MLOps and security, when coupled with governance, education, and change management, significantly enhance both acceptance and impact of AI at the bedside as well as in operations. We conclude by considering limits, generalization, and a roadmap for scaling. LOS pilot (n=3,184 encounters across 4 adult units at a single U.S. hospital, June–August 2025) achieved R² = 0.41–0.58; case-manager adoption reached 78% by week 6; targeted units observed 5–10% relative reductions in mean LOS for complex discharges versus each unit’s pre-pilot baseline. The radiology second-reader pilot (n=1,126 chest CTs over 8 weeks) showed +8.0 percentage-points in sub-centimeter actionable nodule detection (95% CI [2.1, 13.9]; χ² p=0.008), with median report turnaround time unchanged at 23 min (Wilcoxon p=0.64).
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.380 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.243 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.671 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.496 Zit.