Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Fairness-related performance and explainability effects in deep learning models for brain image analysis
37
Zitationen
4
Autoren
2022
Jahr
Abstract
<b>Purpose:</b> Explainability and fairness are two key factors for the effective and ethical clinical implementation of deep learning-based machine learning models in healthcare settings. However, there has been limited work on investigating how unfair performance manifests in explainable artificial intelligence (XAI) methods, and how XAI can be used to investigate potential reasons for unfairness. Thus, the aim of this work was to analyze the effects of previously established sociodemographic-related confounders on classifier performance and explainability methods. <b>Approach:</b> A convolutional neural network (CNN) was trained to predict biological sex from T1-weighted brain MRI datasets of 4547 9- to 10-year-old adolescents from the Adolescent Brain Cognitive Development study. Performance disparities of the trained CNN between White and Black subjects were analyzed and saliency maps were generated for each subgroup at the intersection of sex and race. <b>Results:</b> The classification model demonstrated a significant difference in the percentage of correctly classified White male ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mn>90.3</mml:mn> <mml:mo>%</mml:mo> <mml:mo>±</mml:mo> <mml:mn>1.7</mml:mn> <mml:mo>%</mml:mo></mml:mrow> </mml:math> ) and Black male ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mn>81.1</mml:mn> <mml:mo>%</mml:mo> <mml:mo>±</mml:mo> <mml:mn>4.5</mml:mn> <mml:mo>%</mml:mo></mml:mrow> </mml:math> ) children. Conversely, slightly higher performance was found for Black female ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mn>89.3</mml:mn> <mml:mo>%</mml:mo> <mml:mo>±</mml:mo> <mml:mn>4.8</mml:mn> <mml:mo>%</mml:mo></mml:mrow> </mml:math> ) compared with White female ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mn>86.5</mml:mn> <mml:mo>%</mml:mo> <mml:mo>±</mml:mo> <mml:mn>2.0</mml:mn> <mml:mo>%</mml:mo></mml:mrow> </mml:math> ) children. Saliency maps showed subgroup-specific differences, corresponding to brain regions previously associated with pubertal development. In line with this finding, average pubertal development scores of subjects used in this study were significantly different between Black and White females ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mi>p</mml:mi> <mml:mo><</mml:mo> <mml:mn>0.001</mml:mn></mml:mrow> </mml:math> ) and males ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mrow><mml:mi>p</mml:mi> <mml:mo><</mml:mo> <mml:mn>0.001</mml:mn></mml:mrow> </mml:math> ). <b>Conclusions:</b> We demonstrate that a CNN with significantly different sex classification performance between Black and White adolescents can identify different important brain regions when comparing subgroup saliency maps. Importance scores vary substantially between subgroups within brain structures associated with pubertal development, a race-associated confounder for predicting sex. We illustrate that unfair models can produce different XAI results between subgroups and that these results may explain potential reasons for biased performance.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.284 Zit.
Generative Adversarial Nets
2023 · 19.841 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.233 Zit.
"Why Should I Trust You?"
2016 · 14.179 Zit.
On a Method to Measure Supervised Multiclass Model’s Interpretability: Application to Degradation Diagnosis (Short Paper)
2024 · 13.096 Zit.