Back to Search Start Over

LabelAId: Just-in-time AI Interventions for Improving Human Labeling Quality and Domain Knowledge in Crowdsourcing Systems

Authors :
Li, Chu
Zhang, Zhihan
Saugstad, Michael
Safranchik, Esteban
Kulkarni, Minchu
Huang, Xiaoyu
Patel, Shwetak
Iyer, Vikram
Althoff, Tim
Froehlich, Jon E.
Publication Year :
2024

Abstract

Crowdsourcing platforms have transformed distributed problem-solving, yet quality control remains a persistent challenge. Traditional quality control measures, such as prescreening workers and refining instructions, often focus solely on optimizing economic output. This paper explores just-in-time AI interventions to enhance both labeling quality and domain-specific knowledge among crowdworkers. We introduce LabelAId, an advanced inference model combining Programmatic Weak Supervision (PWS) with FT-Transformers to infer label correctness based on user behavior and domain knowledge. Our technical evaluation shows that our LabelAId pipeline consistently outperforms state-of-the-art ML baselines, improving mistake inference accuracy by 36.7% with 50 downstream samples. We then implemented LabelAId into Project Sidewalk, an open-source crowdsourcing platform for urban accessibility. A between-subjects study with 34 participants demonstrates that LabelAId significantly enhances label precision without compromising efficiency while also increasing labeler confidence. We discuss LabelAId's success factors, limitations, and its generalizability to other crowdsourced science domains.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2403.09810
Document Type :
Working Paper
Full Text :
https://doi.org/10.1145/3613904.3642089