CrowdSelect: Increasing Accuracy of Crowdsourcing Tasks through Behavior Prediction and User Selection

作者: Chenxi Qiu , Anna C. Squicciarini , Barbara Carminati , James Caverlee , Dev Rishi Khare

DOI: 10.1145/2983323.2983830

关键词: Term (time)Machine learningArtificial intelligenceTask (project management)Data miningComputer scienceQuality (business)Assignment problemCrowdsourcing

摘要: Crowdsourcing allows many people to complete tasks of various difficulty with minimal recruitment and administration costs. However, the lack participant accountability may entice as possible without fully engaging in them, jeopardizing quality responses. In this paper, we present a dynamic time efficient solution task assignment problem crowdsourcing platforms. Our proposed approach, CrowdSelect, offers theoretically proven algorithm assign workers cost manner, while ensuring high accuracy overall task. contrast existing works, our approach makes assumptions on probability error for workers, completely removes that such is known apriori it remains consistent over time. Through experiments real Amazon Mechanical Turk traces synthetic data, find CrowdSelect has significant gain term compared state-of-the-art algorithms, can provide 17.5\% answers' previous methods, even when there are 50\% malicious workers.

参考文章(37)
A. P. Dawid, A. M. Skene, Maximum Likelihood Estimation of Observer Error‐Rates Using the EM Algorithm Journal of The Royal Statistical Society Series C-applied Statistics. ,vol. 28, pp. 20- 28 ,(1979) , 10.2307/2346806
Vaughn Hester, David Oleson, Alexander Sorokin, John Le, Lukas Biewald, Greg Laughlin, Programmatic gold: targeted and scalable quality assurance in crowdsourcing national conference on artificial intelligence. pp. 43- 48 ,(2011)
Chien-Ju Ho, Jennifer Wortman Vaughan, Shahin Jabbari, Adaptive Task Assignment for Crowdsourced Classification international conference on machine learning. pp. 534- 542 ,(2013)
Ari Kobren, Chun How Tan, Panagiotis Ipeirotis, Evgeniy Gabrilovich, Getting More for Less: Optimized Crowdsourcing with Dynamic Tasks and Goals the web conference. pp. 592- 602 ,(2015) , 10.1145/2736277.2741681
Daniel Berend, Aryeh Kontorovich, A finite sample analysis of the Naive Bayes classifier Journal of Machine Learning Research. ,vol. 16, pp. 1519- 1545 ,(2015)
Rion Snow, Brendan O'Connor, Daniel Jurafsky, Andrew Y. Ng, Cheap and fast---but is it good? Proceedings of the Conference on Empirical Methods in Natural Language Processing - EMNLP '08. pp. 254- 263 ,(2008) , 10.3115/1613715.1613751
Arpita Ghosh, Satyen Kale, Preston McAfee, Who moderates the moderators?: crowdsourcing abuse detection in user-generated content electronic commerce. pp. 167- 176 ,(2011) , 10.1145/1993574.1993599