Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Extracting Knowledge from Data in Lightweight Digital Twin Construction
0
Zitationen
9
Autoren
2025
Jahr
Abstract
In the medical domain, the early detection and monitoring of specific diseases require both accuracy and interpretability to support clinical decisions. Human digital twin systems are increasingly used in this context, but their adoption often requires data-intensive process for the learning tasks and a high degree of explainability to ensure clinical reliability. To address these challenges, we propose a pipeline that prioritises lightweight and explainability, to extract actionable knowledge from patient data in terms of rules. The approach follows a traditional Machine Learning methodology, including data pre-processing, followed by the construction and validation of a classification model. A rule extraction phase is then introduced to make the classifier’s decision process interpretable. The reliability of the pipeline was evaluated by extracting decision rules for the detection of kidney damage in patients with Congenital Solitary Functioning Kidney. Through the analysis of patient data and the use of a Random Forest classifier, key clinical parameters (e.g., creatinine levels, Holter monitor measurements, and kidney volume) were identified as fundamental to support accurate and reliable diagnoses in clinical practice.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.292 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.143 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.539 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.452 Zit.