skip to main content
10.1145/2531602.2531733acmconferencesArticle/Chapter ViewAbstractPublication PagescscwConference Proceedingsconference-collections
research-article

Information extraction and manipulation threats in crowd-powered systems

Published:15 February 2014Publication History

ABSTRACT

Crowd-powered systems have become a popular way to augment the capabilities of automated systems in real-world settings. Many of these systems rely on human workers to process potentially sensitive data or make important decisions. This puts these systems at risk of unintentionally releasing sensitive data or having their outcomes maliciously manipulated. While almost all crowd-powered approaches account for errors made by individual workers, few factor in active attacks on the system. In this paper, we analyze different forms of threats from individuals and groups of workers extracting information from crowd-powered systems or manipulating these systems' outcomes. Via a set of studies performed on Amazon's Mechanical Turk platform and involving 1,140 unique workers, we demonstrate the viability of these threats. We show that the current system is vulnerable to coordinated attacks on a task based on the requests of another task and that a significant portion of Mechanical Turk workers are willing to contribute to an attack. We propose several possible approaches to mitigating these threats, including leveraging workers who are willing to go above and beyond to help, automatically flagging sensitive content, and using workflows that conceal information from each individual, while still allowing the group to complete a task. Our findings enable the crowd to continue to play an important part in automated systems, even as the data they use and the decisions they support become increasingly important.

References

  1. Bernstein, M.S., Brandt, J.R., Miller, R.C. and Karger, D.R. Crowds in two seconds: Enabling realtime crowdpowered interfaces. In Proceedings of UIST 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Bernstein, M.S., Little, G., Miller, R.C., Hartmann, B., Ackerman, M.S., Karger, D.R., Crowell, D. and Panovich, K. Soylent: A word processor with a crowd inside. In Proceedings of UIST 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Bigham, J.P., Jayant, C., Ji, H., Little, G., Miller, A., Miller, R.C., Miller, R., Tatarowicz, A., White, B., White, S. and Yeh, T. VizWiz: nearly real-time answers to visual questions. In Proceedings of UIST 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Chen, X., Bennett, P.N., Collins-Thompson, K. and Horvitz, E. Pairwise ranking aggregation in a crowdsourced setting. In Proceedings of WSDM 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Deng, J., Krause, J., and Fei-Fei, L. Fine-Grained Crowdsourcing for Fine-Grained Recognition. In Proceesings of CVPR 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Evanini, K. and Zechner, K. Using crowdsourcing to provide prosodic annotations for non-native speech. In Proceedings of Interspeech 2011.Google ScholarGoogle ScholarCross RefCross Ref
  7. Featured requester: Discover how comScore benefits from mechanical Turk. The Mechanical Turk Blog, March 2013. http://bit.ly/Ye64SbGoogle ScholarGoogle Scholar
  8. Harris, Christopher G. Dirty Deeds Done Dirt Cheap: A Darker Side to Crowdsourcing. In SocialCom 2011.Google ScholarGoogle ScholarCross RefCross Ref
  9. Kamar, E., Hacker, S., Lintott, C. and Horvitz, E. Combining human and machine learning intelligence in large-scale crowdsourcing: Principles, methods, and studies. MSR-TR-2012-58, 2012.Google ScholarGoogle Scholar
  10. Kittur, A. and Kraut, R.E. Harnessing the wisdom of crowds in Wikipedia: Quality through coordination. In Proceedings of CSCW 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Kittur, A., Nickerson, J.V., Bernstein, M., Gerber, E., Shaw, A., Zimmerman, J., Lease, M., and Horton, J. The future of crowd work. In Proceedings of CSCW 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Kokkalis, N., Köhn, T., Pfeiffer, C., Chornyi, D., Bernstein, M.S. and Klemmer, S.R. EmailValet: Managing email overload through private, accountable crowdsourcing. In Proceedings of CSCW 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Le, J., Edmonds, A., Hester,V., and Biewald, L. Ensuring quality in crowdsourced search relevance evaluation. In Proceedings of SIGIR 2010 Workshop on Crowdsourcing for Search Evaluation.Google ScholarGoogle Scholar
  14. Little, G., Chilton, L.B., Goldman, M. and Miller, R.C. TurKit: Human computation algorithms on mechanical turk. In Proceedings of UIST 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Little, G. and Sun, Y-A. Human OCR. CHI 2011 Workshop on Crowdsourcing and Human Computation.Google ScholarGoogle Scholar
  16. Lasecki, W.S., Miller, C., Sadilek, A., Abumoussa, A., Borrello, D., Kushalnagar, R. and Bigham, J.P. Realtime captioning by groups of non-experts. In Proceedings of UIST 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Lasecki, W.S., Murray, K.I., White, S., Miller, R.C. and Bigham, J.P. Real-time crowd control of existing interfaces. In Proceedings of UIST 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Lasecki, W.S., Song, Y.C., Kautz, H. and Bigham, J.P. Real-time crowd labeling for deployable activity recognition. In Proceedings CSCW 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Lasecki, W.S., Thiha, P., Zhong, Y., Brady, E. and Bigham, J.P. Answering Visual Questions with Conversational Crowd Assistants. In Proceedings of ASSETS 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Lasecki, W.S., Wesley, R., Nichols, J., Kulkari, A., Allen, J.F. and Bigham, J.P. Chorus: A Crowd-Powered Personal Assistant. In Proceedings of UIST 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Lasecki, W.S., White, S.C., Murray, K.I. and Bigham, J.P. Crowd Memory: Learning in the collective. In Proceedings of Collective Intelligence 2012.Google ScholarGoogle Scholar
  22. Mason, W. and Watts, D.J. Financial incentives and the performance of crowds. In Proceedings of HComp 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Massively multiplayer pong. http://collisiondetection.net, 2006.Google ScholarGoogle Scholar
  24. Menton, C. and Singh, P. Manipulation can be hard in tractable voting systems even for constant-sized coalitions. CoRR abs/1108.4439.Google ScholarGoogle Scholar
  25. Sun, Y-A. Roy, S. and Little, G. Beyond independent agreement: A tournament selection approach for quality assurance of human computation tasks. In HComp 2011.Google ScholarGoogle Scholar
  26. Sweeney, L. Uniqueness of simple demographics in the U.S. population. In LIDAP-WP4, Carnegie Mellon University. 2000.Google ScholarGoogle Scholar
  27. Quinn, A.J. and Bederson, B.B. Human computation: A survey and taxonomy of a growing field. In Proceedings of CHI 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Van Boskrik, S. US interactive marketing forecast, 2011 to 2016. Forrester. August 2011.Google ScholarGoogle Scholar
  29. Von Ahn, L. Games with a purpose. Comp. 39(6), 2006.Google ScholarGoogle Scholar
  30. Zaidan, O.F. and Callison-Burch, C. Crowdsourcing translation: professional quality from non-professionals. In Proceedings of ACL-HLT 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Information extraction and manipulation threats in crowd-powered systems

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        CSCW '14: Proceedings of the 17th ACM conference on Computer supported cooperative work & social computing
        February 2014
        1600 pages
        ISBN:9781450325400
        DOI:10.1145/2531602

        Copyright © 2014 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 15 February 2014

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        CSCW '14 Paper Acceptance Rate134of497submissions,27%Overall Acceptance Rate2,235of8,521submissions,26%

        Upcoming Conference

        CSCW '24

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader