skip to main content
10.1145/2566486.2568033acmotherconferencesArticle/Chapter ViewAbstractPublication PageswwwConference Proceedingsconference-collections
research-article

The wisdom of minority: discovering and targeting the right group of workers for crowdsourcing

Authors Info & Claims
Published:07 April 2014Publication History

ABSTRACT

Worker reliability is a longstanding issue in crowdsourcing, and the automatic discovery of high quality workers is an important practical problem. Most previous work on this problem mainly focuses on estimating the quality of each individual worker jointly with the true answer of each task. However, in practice, for some tasks, worker quality could be associated with some explicit characteristics of the worker, such as education level, major and age. So the following question arises: how do we automatically discover related worker attributes for a given task, and further utilize the findings to improve data quality? In this paper, we propose a general crowd targeting framework that can automatically discover, for a given task, if any group of workers based on their attributes have higher quality on average; and target such groups, if they exist, for future work on the same task. Our crowd targeting framework is complementary to traditional worker quality estimation approaches. Furthermore, an advantage of our framework is that it is more budget efficient because we are able to target potentially good workers before they actually do the task. Experiments on real datasets show that the accuracy of final prediction can be improved significantly for the same budget (or even less budget in some cases). Our framework can be applied to many real word tasks and can be easily integrated in current crowdsourcing platforms.

References

  1. Y. Bachrach, T. Graepel, T. Minka, and J. Guiver. How to grade a test without knowing the answers-a bayesian graphical model for adaptive crowdsourcing and aptitude testing. arXiv preprint arXiv:1206.6386, 2012.Google ScholarGoogle Scholar
  2. A. P. Dawid and A. M. Skene. Maximum Likelihood Estimation of Observer Error-Rates Using the EM Algorithm. Journal of the Royal Statistical Society., 28(1):20--28, 1979.Google ScholarGoogle Scholar
  3. A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal Statistical Society. Series B (Methodological), pages 1--38, 1977.Google ScholarGoogle ScholarCross RefCross Ref
  4. S. Ertekin, H. Hirsh, and C. Rudin. Learning to predict the wisdom of crowds. arXiv preprint arXiv:1204.3611, 2012.Google ScholarGoogle Scholar
  5. C.-J. Ho, S. Jabbari, and J. W. Vaughan. Adaptive task assignment for crowdsourced classification. In ICML, pages 534--542, 2013.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. P. G. Ipeirotis. Analyzing the amazon mechanical turk marketplace. XRDS: Crossroads, The ACM Magazine for Students, 17(2):16--21, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. P. G. Ipeirotis. Demographics of mechanical turk. In NYU Digital Working Paper CeDER-10-01, 2010.Google ScholarGoogle Scholar
  8. P. G. Ipeirotis, F. Provost, and J. Wang. Quality management on amazon mechanical turk. In Proceedings of the ACM SIGKDD workshop on human computation, pages 64--67. ACM, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. D. R. Karger, S. Oh, and D. Shah. Budget-optimal crowdsourcing using low-rank matrix approximations. In Communication, Control, and Computing (Allerton), 2011 49th Annual Allerton Conference on, pages 284--291. IEEE, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  10. D. R. Karger, S. Oh, and D. Shah. Iterative learning for reliable crowdsourcing systems. In NIPS, pages 1953--1961, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. G. Kazai, J. Kamps, and N. Milic-Frayling. The face of quality in crowdsourcing relevance labels: Demographics, personality and labeling accuracy. In Proceedings of the 21st ACM Conference on Information and Knowledge Management (CIKM 2012). ACM Press, New York NY, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. G. Kazai, J. Kamps, and N. Milic-Frayling. An analysis of human factors and label accuracy in crowdsourcing relevance judgments. Information Retrieval, 16:138--178, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. H. Li, B. Yu, and D. Zhou. Error Rate Analysis of Labeling by Crowdsourcing. In ICML Workshop: Machine Learning Meets Crowdsourcing. Atalanta, Georgia, USA., 2013.Google ScholarGoogle Scholar
  14. H. Li, B. Yu, and D. Zhou. Error rate bounds in crowdsourcing models. arXiv preprint arXiv:1307.2674, 2013.Google ScholarGoogle Scholar
  15. Q. Liu, J. Peng, and A. Ihler. Variational inference for crowdsourcing. NIPS, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. R. M. Mickey, O. J. Dunn, and V. Clark. Applied statistics: analysis of variance and regression. Wiley-Interscience, 2004.Google ScholarGoogle Scholar
  17. T. Pfeier, X. A. Gao, Y. Chen, A. Mao, and D. G. Rand. Adaptive polling for information aggregation. In AAAI, 2012.Google ScholarGoogle Scholar
  18. V. C. Raykar, S. Yu, L. H. Zhao, C. Florin, L. Bogoni, and L. Moy. Learning From Crowds. Journal of Machine Learning Research, 11:1297--1322, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. J. Ross, L. Irani, M. Silberman, A. Zaldivar, and B. Tomlinson. Who are the crowdworkers?: shifting demographics in mechanical turk. In CHI, pages 2863--2872. ACM, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. L. A. Schmidt. Crowdsourcing for human subjects research. Proceedings of CrowdConf, 2010.Google ScholarGoogle Scholar
  21. P. Smyth, U. Fayyad, M. Burl, P. Perona, and P. Baldi. Inferring Ground Truth from Subjective Labelling of Venus Images. In NIPS, 1995.Google ScholarGoogle Scholar
  22. R. Snow, B. O. Connor, D. Jurafsky, A. Y. Ng, D. Labs, and C. St. Cheap and Fast - But is it Good -- Evaluating Non-Expert Annotations for Natural Language Tasks. EMNLP, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. L. STAHLE and S. Wold. Analysis of variance (anova). Chemometrics and intelligent laboratory systems, 6(4):259--272, 1989.Google ScholarGoogle Scholar
  24. P. Welinder, S. Branson, S. Belongie, and P. Perona. The Multidimensional Wisdom of Crowds. In NIPS, 2010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. J. Whitehill, P. Ruvolo, T. Wu, J. Bergsma, and J. Movellan. Whose Vote Should Count More : Optimal Integration of Labels from Labelers of Unknown Expertise. In NIPS, pages 2035--2043, 2009.Google ScholarGoogle Scholar
  26. Y. Yan, G. M. Fung, R. Rosales, and J. G. Dy. Active learning from crowds. In ICML, pages 1161--1168, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. D. Zhou, J. Platt, S. Basu, and Y. Mao. Learning from the Wisdom of Crowds by Minimax Entropy. In NIPS, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. The wisdom of minority: discovering and targeting the right group of workers for crowdsourcing

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        WWW '14: Proceedings of the 23rd international conference on World wide web
        April 2014
        926 pages
        ISBN:9781450327442
        DOI:10.1145/2566486

        Copyright © 2014 Copyright is held by the International World Wide Web Conference Committee (IW3C2).

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 7 April 2014

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        WWW '14 Paper Acceptance Rate84of645submissions,13%Overall Acceptance Rate1,899of8,196submissions,23%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader