Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Expert-quality Dataset Labeling via Gamified Crowdsourcing on Point-of-Care Lung Ultrasound Data
0
Zitationen
12
Autoren
2024
Jahr
Abstract
Machine learning tools can aid medical imaging data interpretation.Building such tools requires labeled training datasets.We tested whether a gamified crowdsourcing approach can produce clinical expert-quality lung ultrasound clip labels.2,384 lung ultrasound clips were retrospectively collected.Six lung ultrasound experts classified 393 of these clips as having no B-lines, one or more discrete B-lines, or confluent B-lines to create two sets of reference standard labels: a training and test set.Sets trained users on a gamified crowdsourcing platform, and compared concordance of the resulting crowd labels to the concordance of individual experts to reference standards, respectively.99,238 crowdsourced opinions were collected from 426 unique users over 8 days.Mean labeling concordance of individual experts relative to the reference standard was 85.0% 2.0 (SEM), compared to 87.9% crowdsourced label concordance (p=0.15).Scalable, high-quality labeling approaches such as crowdsourcing may streamline training dataset creation for machine learning model development.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.260 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.116 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.493 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.438 Zit.