Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Enhancing Cardiovascular Disease Classification with Routine Blood Tests Using an Explainable AI Approach
0
Zitationen
8
Autoren
2025
Jahr
Abstract
Background: While machine learning (ML) is widely applied in cardiology, a critical research gap persists. The incremental diagnostic value of routine blood tests for classifying cardiovascular disease (CVD) remains largely unquantified, and many models operate as non-interpretable “black boxes,” limiting their clinical adoption. This study aims to address these gaps by quantifying the contribution of readily available laboratory panels and demonstrating the utility of transparent diagnostic modeling within a real-world clinical cohort. Methods: We conducted a retrospective study on the clinical data of 896 adult patients from a hospital database. A baseline feature set (demographics, vital signs) was compared against an enhanced set that additionally included results from routine hematology and biochemistry panels. Five machine learning classifiers were trained and evaluated. To ensure transparency, SHAP (SHapley Additive exPlanations) analysis, a key component of explainable AI (XAI), was used to interpret the predictions of the top-performing model. Results: The inclusion of routine blood tests consistently and significantly improved the performance of all classifiers. The XGBoost model demonstrated the best performance (accuracy 91.62%, precision 95.00%, recall 87.36%). Critically, SHAP analysis identified aspartate aminotransferase (AST), glucose, and creatinine as the most significant biomarkers, providing clear, interpretable insights into the biochemical drivers of the model’s predictions. Conclusion: Routine laboratory markers contain a strong, interpretable signal indicative of CVD that is crucial for accurate risk stratification. These findings underscore the diagnostic relevance of common blood biomarkers and demonstrate how explainable AI can transform routine clinical data into transparent and actionable cardiovascular insights. Further validation in larger and demographically diverse cohorts is warranted.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.292 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.143 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.539 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.452 Zit.