OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 19.03.2026, 18:21

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Expert-quality Dataset Labeling via Gamified Crowdsourcing on Point-of-Care Lung Ultrasound Data

2024·0 Zitationen·Proceedings of the ... Annual Hawaii International Conference on System Sciences/Proceedings of the Annual Hawaii International Conference on System SciencesOpen Access
Volltext beim Verlag öffnen

0

Zitationen

12

Autoren

2024

Jahr

Abstract

Machine learning tools can aid medical imaging data interpretation.Building such tools requires labeled training datasets.We tested whether a gamified crowdsourcing approach can produce clinical expert-quality lung ultrasound clip labels.2,384 lung ultrasound clips were retrospectively collected.Six lung ultrasound experts classified 393 of these clips as having no B-lines, one or more discrete B-lines, or confluent B-lines to create two sets of reference standard labels: a training and test set.Sets trained users on a gamified crowdsourcing platform, and compared concordance of the resulting crowd labels to the concordance of individual experts to reference standards, respectively.99,238 crowdsourced opinions were collected from 426 unique users over 8 days.Mean labeling concordance of individual experts relative to the reference standard was 85.0% 2.0 (SEM), compared to 87.9% crowdsourced label concordance (p=0.15).Scalable, high-quality labeling approaches such as crowdsourcing may streamline training dataset creation for machine learning model development.

Ähnliche Arbeiten