Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Improving crowdsourcing for AI through cognitive-inspired data engineering
0
Zitationen
6
Autoren
2026
Jahr
Abstract
Crowdsourcing offers a fast and cost-efficient approach to obtaining human labeled datasets. However, crowdsourced datasets and the models trained on them can inherit the cognitive constraints and biases of their annotators. In a process we refer to as cognitive-inspired data engineering, we investigate whether ideas from cognitive science can be applied to mitigate the presence of cognitive constraints and cognitive biases in crowdsourced datasets and, as a result, improve the performance of models trained on these datasets. We evaluate our approach by crowdsourcing labels for medical image diagnostic tasks using two different crowdsourcing platforms across two experiments. In Experiment 1, we collect subjective probability judgments from novice annotators through Amazon Mechanical Turk and, in Experiment 2, we collect subjective probability judgments and binary classifications from skilled annotators though DiagnosUs, a crowdsourcing platform specializing in medical and scientific data annotation. In both experiments, we find that de-biasing subjective probability judgments via recalibration leads to more accurate crowdsourced datasets and more accurate models trained on these datasets. Our results suggest that cognitive-inspired data engineering offers a promising avenue to improve the quality of crowdsourced datasets\textcolor{black}{, with consistent downstream benefits for machine learning models
Ähnliche Arbeiten
Internet of Things (IoT): A vision, architectural elements, and future directions
2013 · 11.801 Zit.
Fog computing and its role in the internet of things
2012 · 5.902 Zit.
From Louvain to Leiden: guaranteeing well-connected communities
2019 · 4.865 Zit.
Advances and Open Problems in Federated Learning
2020 · 4.339 Zit.
Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk
2012 · 4.049 Zit.