Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Comparative Analysis of Artificial Intelligence Methods in Clinical Implementation: A Review of Techniques, Validation Strategies, and Success Metrics
2
Zitationen
4
Autoren
2025
Jahr
Abstract
Artificial intelligence (AI) and machine learning (ML) are increasingly integrated into clinical workflows, yet evidence comparing their real-world effectiveness remains fragmented. This review systematically evaluates AI/ML methods deployed in healthcare, focusing on implementation strategies, validation rigor, and performance metrics. To identify the most frequently implemented AI/ML techniques, assess their clinical success rates, and analyze workflow integration challenges across specialties. We reviewed PubMed articles (2019–2024) describing AI/ML clinical applications with quantitative outcomes. Ten studies met inclusion criteria, covering radiology, oncology, and pediatrics. Data were extracted on AI methods, validation types, performance metrics (e.g., sensitivity, AUC), and workflow integration. Descriptive statistics summarized findings. Logistic regression and deep learning (e.g., atlas-matching) were the most specified methods. Logistic regression achieved 71% sensitivity and 77% PPV in epilepsy screening, matching clinician performance. Deep learning models showed >90% retrospective acceptability in radiotherapy planning but lacked prospective metrics. Only 40% of studies reported quantitative outcomes; others emphasized usability or frameworks. Workflow integration (e.g., EHR embedding) was critical but inconsistently detailed. While both traditional and advanced AI methods demonstrate clinical utility, heterogeneous reporting and limited head-to-head comparisons hinder definitive conclusions. Future research should prioritize standardized performance metrics and prospective multi-method evaluations to guide evidence-based adoption.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.245 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.100 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.466 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.429 Zit.