Abstract
We study the causal effects of financial incentives on the quality of crowdwork. We focus on performance-based payments (PBPs), bonus payments awarded to workers for producing high quality work. We design and run randomized behavioral experiments on the popular crowdsourcing platform Amazon Mechanical Turk with the goal of understanding when, where, and why PBPs help, identifying properties of the payment, payment structure, and the task itself that make them most effective. We provide examples of tasks for which PBPs do improve quality. For such tasks, the effectiveness of PBPs is not too sensitive to the threshold for quality required to receive the bonus, while the magnitude of the bonus must be large enough to make the reward salient. We also present examples of tasks for which PBPs do not improve quality. Our results suggest that for PBPs to improve quality, the task must be effort-responsive: the task must allow workers to produce higher quality work by exerting more effort. We also give a simple method to determine if a task is effort-responsive a priori. Furthermore, our experiments suggest that all payments on Mechanical Turk are, to some degree, implicitly performance-based in that workers believe their work may be rejected if their performance is sufficiently poor. In the full version of this paper, we propose a new model of worker behavior that extends the standard principal-agent model from economics to include a worker's subjective beliefs about his likelihood of being paid, and show that the predictions of this model are in line with our experimental findings. This model may be useful as a foundation for theoretical studies of incentives in crowdsourcing markets.
- Alonso, O. 2013. Implementing crowdsourcing-based relevance experimentation: An industrial perspective. Information Retrieval 16, 2, 101--120. Google ScholarDigital Library
- Buhrmester, M., Kwang, T., and Gosling, S. D. 2011. Amazon's Mechanical Turk: A new source of inexpensive, yet high-quality, data? Perspectives on Psychological Science.Google Scholar
- Camerer, C. F. and Hogarth, R. 1999. The effects of financial incentives in economics experiments: A review and capital-labor-production framework. Journal of Risk and Uncertainty 19, 1, 7--42.Google ScholarCross Ref
- Gilchrist, D., Luca, M., and Malhotra, D. 2014. When 3+1 > 4: Gift structure and reciprocity in the field. Tech. rep. Working Paper.Google Scholar
- Harris, C. G. 2011. You're hired! An examination of crowdsourcing incentive models in human resource tasks. In WSDM 2011 Workshop on Crwdsourcing for Search and Data Mining.Google Scholar
- Ho, C.-J., Jabbari, S., and Vaughan, J. W. 2013. Adaptive task assignment for crowdsourced classification. In ICML.Google Scholar
- Ho, C.-J. and Vaughan, J. W. 2012. Online task assignment in crowdsourcing markets. In AAAI. Google ScholarDigital Library
- Horton, J. J. and Chilton, L. B. 2010. The labor economics of paid crowdsourcing. In ACM EC. Google ScholarDigital Library
- Horton, J. J., Rand, D., and Zeckhauser, R. 2011. The online laboratory: Conducting experiments in a real labor market. Experimental Economics 14, 3, 399--425.Google ScholarCross Ref
- Ipeirotis, P. G., Provost, F., and Wang, J. 2010. Quality management on Amazon Mechanical Turk. In HCOMP. Google ScholarDigital Library
- Karger, D., Oh, S., and Shah, D. 2011. Iterative learning for reliable crowdsourcing systems. In NIPS.Google Scholar
- Kittur, A., Chi, E., and Suh, B. 2008. Crowdsourcing user studies with Mechanical Turk. In CHI. Google ScholarDigital Library
- Liu, Q., Peng, J., and Ihler, A. 2012. Variational inference for crowdsourcing. In NIPS.Google Scholar
- Mason, W. and Suri, S. 2012. Conducting behavioral research on Amazon's Mechanical Turk. Behavior Research Methods 44, 1, 1--23.Google ScholarCross Ref
- Mason, W. and Watts, D. J. 2009. Financial incentives and the "performance of crowds". In HCOMP. Google ScholarDigital Library
- Rogstadius, J., Kostakos, V., Kittur, A., Smus, B., Laredo, J., and Vukovic, M. 2011. An assessment of intrinsic and extrinsic motivation on task performance in crowdsourcing markets. In ICWSM.Google Scholar
- Schall, D. 2012. Service-Oriented Crowdsourcing - Architecture, Protocols and Algorithms. Springer Briefs in Computer Science. Springer. Google ScholarDigital Library
- Seltman, H. J. 2014. Experimental design and analysis. http://www.stat.cmu.edu/~hseltman/309/Book/Book.pdf.Google Scholar
- Shaw, A. D., Horton, J. J., and Chen, D. L. 2011. Designing incentives for inexpert human raters. In CSCW. Google ScholarDigital Library
- Sheng, V., Provost, F., and Ipeirotis, P. 2008. Get another label? Improving data quality using multiple, noisy labelers. In KDD. Google ScholarDigital Library
- Wah, C., Branson, S., Welinder, P., Perona, P., and Belongie, S. 2011. The Caltech-UCSD Birds-200-2011 Dataset. Tech. Rep. CNS-TR-2011-001, California Institute of Technology.Google Scholar
- Yin, M., Chen, Y., and Sun, Y.-A. 2013. The effects of performance-contingent financial incentives in online labor markets. In AAAI. Google ScholarDigital Library
- Yin, M., Chen, Y., and Sun, Y.-A. 2014. Monetary interventions in crowdsourcing task switching. In HCOMP.Google Scholar
Index Terms
- Incentivizing high quality crowdwork
Recommendations
Curiosity Killed the Cat, but Makes Crowdwork Better
CHI '16: Proceedings of the 2016 CHI Conference on Human Factors in Computing SystemsCrowdsourcing systems are designed to elicit help from humans to accomplish tasks that are still difficult for computers. How to motivate workers to stay longer and/or perform better in crowdsourcing systems is a critical question for designers. ...
Incentivizing High Quality Crowdwork
WWW '15: Proceedings of the 24th International Conference on World Wide WebWe study the causal effects of financial incentives on the quality of crowdwork. We focus on performance-based payments (PBPs), bonus payments awarded to workers for producing high quality work. We design and run randomized behavioral experiments on the ...
Incentivizing Distributive Fairness for Crowdsourcing Workers
AAMAS '19: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent SystemsIn a crowd market such as Amazon Mechanical Turk, the remuneration of Human Intelligence Tasks is determined by the requester, for which they are not given many cues to ascertain how to "fairly'' pay their workers. Furthermore, the current methods for ...
Comments