Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Artificial Intelligence in PSE: A Scoping Review of Imaging-Based Models in African Settings and a Pathway to Practice (Preprint)
0
Zitationen
4
Autoren
2026
Jahr
Abstract
<sec> <title>BACKGROUND</title> Post-stroke seizures and post-stroke epilepsy are important complications that affect recovery and long-term quality of life after stroke. Predicting which patients are at higher risk remains challenging, and recent studies have explored the use of artificial intelligence to improve prediction. </sec> <sec> <title>OBJECTIVE</title> This scoping review aimed to map how artificial intelligence and related computational methods have been applied to predict post-stroke seizures and epilepsy, to describe the role of neuroimaging in these models, and to assess representation of African settings in the existing literature. </sec> <sec> <title>METHODS</title> A systematic search identified studies that used statistical, machine learning, radiomics, or deep learning approaches to predict seizure-related outcomes after stroke. Traditional statistical models and clinical risk scores were the most commonly used methods. A smaller number of studies applied machine learning techniques, while radiomics and deep learning approaches were reported in only a few studies. Neuroimaging was mainly used to predict long-term epilepsy rather than early seizures. Most models were evaluated using internal validation only, with limited external validation. </sec> <sec> <title>RESULTS</title> Studies from African settings were underrepresented and primarily focused on describing seizure frequency and associated factors using regression analysis. No studies were identified that developed or validated artificial intelligence-based prediction models using African datasets. </sec> <sec> <title>CONCLUSIONS</title> Overall, the findings show that while interest in AI-based prediction is growing, advanced methods remain limited, and major gaps exist in model validation and geographic representation. These findings highlight the need for context-specific, well-validated prediction models, particularly in African healthcare settings. </sec>
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.490 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.376 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.832 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.553 Zit.