Title
CrowdED: Guideline for Optimal Crowdsourcing Experimental Design.
Abstract
Crowdsourcing involves the creating of HITs (Human Intelligent Tasks), submitting them to a crowdsourcing platform and providing a monetary reward for each HIT. One of the advantages of using crowdsourcing is that the tasks can be highly parallelized, that is, the work is performed by a high number of workers in a decentralized setting. The design also offers a means to cross-check the accuracy of the answers by assigning each task to more than one person and thus relying on majority consensus as well as reward the workers according to their performance and productivity. Since each worker is paid per task, the costs can significantly increase, irrespective of the overall accuracy of the results. Thus, one important question when designing such crowdsourcing tasks that arise is how many workers to employ and how many tasks to assign to each worker when dealing with large amounts of tasks. That is, the main research questions we aim to answer is: 'Can we a-priori estimate optimal workers and tasks' assignment to obtain maximum accuracy on all tasks'. Thus, we introduce a two-staged statistical guideline, CrowdED, for optimal crowdsourcing experimental design in order to a-priori estimate optimal workers and tasks' assignment to obtain maximum accuracy on all tasks. We describe the algorithm and present preliminary results and discussions. We implement the algorithm in Python and make it openly available on Github, provide a Jupyter Notebook and a R Shiny app for users to re-use, interact and apply in their own crowdsourcing experiments.
Year
DOI
Venue
2018
10.1145/3184558.3191543
WWW '18: The Web Conference 2018 Lyon France April, 2018
DocType
ISBN
Citations 
Conference
978-1-4503-5640-4
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Amrapali Zaveri136824.37
Pedro Hernandez Serrano200.34
Manisha Desai381.38
Michel Dumontier489893.35