Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Artificial intelligence in breast cancer screening: primary care provider preferences
50
Zitationen
5
Autoren
2020
Jahr
Abstract
BACKGROUND: Artificial intelligence (AI) is increasingly being proposed for use in medicine, including breast cancer screening (BCS). Little is known, however, about referring primary care providers' (PCPs') preferences for this technology. METHODS: We identified the most important attributes of AI BCS for ordering PCPs using qualitative interviews: sensitivity, specificity, radiologist involvement, understandability of AI decision-making, supporting evidence, and diversity of training data. We invited US-based PCPs to participate in an internet-based experiment designed to force participants to trade off among the attributes of hypothetical AI BCS products. Responses were analyzed with random parameters logit and latent class models to assess how different attributes affect the choice to recommend AI-enhanced screening. RESULTS: Ninety-one PCPs participated. Sensitivity was most important, and most PCPs viewed radiologist participation in mammography interpretation as important. Other important attributes were specificity, understandability of AI decision-making, and diversity of data. We identified 3 classes of respondents: "Sensitivity First" (41%) found sensitivity to be more than twice as important as other attributes; "Against AI Autonomy" (24%) wanted radiologists to confirm every image; "Uncertain Trade-Offs" (35%) viewed most attributes as having similar importance. A majority (76%) accepted the use of AI in a "triage" role that would allow it to filter out likely negatives without radiologist confirmation. CONCLUSIONS AND RELEVANCE: Sensitivity was the most important attribute overall, but other key attributes should be addressed to produce clinically acceptable products. We also found that most PCPs accept the use of AI to make determinations about likely negative mammograms without radiologist confirmation.
Ähnliche Arbeiten
A survey on deep learning in medical image analysis
2017 · 13.880 Zit.
pROC: an open-source package for R and S+ to analyze and compare ROC curves
2011 · 13.750 Zit.
Dermatologist-level classification of skin cancer with deep neural networks
2017 · 13.439 Zit.
A survey on Image Data Augmentation for Deep Learning
2019 · 12.033 Zit.
QuPath: Open source software for digital pathology image analysis
2017 · 8.379 Zit.