Abstract
Crowdsourcing promises to expand organizational knowledge and “sensor” networks dramatically, making it possible to engage ordinary people in large-scale data collection, often at much lower cost than that of traditional approaches to gathering data. A major challenge in crowdsourcing is ensuring that the data that crowds provide is of sufficient quality to be usable in organizational decision-making and analysis. We refer to this challenge as the Problem of Crowd Information Quality (Crowd IQ). We need to increase quality while giving contributors the flexibility to contribute data based on their individual perceptions. The design science research project produced several artifacts, including a citizen science information system (NLNature), design principles (guidelines) for the development of crowdsourcing projects, and an instance-based crowdsourcing design theory. We also made several methodological contributions related to the process of design science research and behavioral research in information systems. Over the course of the project, we addressed several challenges in designing crowdsourcing systems, formulating design principles, and conducting rigorous design science research. Specifically, we showed that: design choices can have a sizable impact in the real world; it can be unclear how to implement design principles; and design features that are unrelated to design principles can confound efforts to evaluate artifacts. During the project, we also experienced challenges for which no adequate solution was found, reaffirming that design is an iterative process.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
While basic-level categories are among the most researched notions in cognitive psychology (Lassaline, Wisniewski, & Medin, 1992), our realization that this notion lacks operational precision precipitated development of a separate DSR project to formalize the notion of basic-level categories, provide clear and unambiguous principles for discovering and identifying these classes in a domain, and provide practitioners with principles for using these classes in the design, development and use of information systems (Castellanos, Castillo, Lukyanenko, & Tremblay, 2017; Castellanos, Lukyanenko, Samuel, & Tremblay, 2016; Lukyanenko & Samuel, 2017).
- 3.
This is a simplified implementation. In the real schema, additional attributes were included in each table, including additional semantic attributes and a variety of system attributes like time stamp, IP address, system properties of the record creator, security, validation, and monitoring keys.
References
Angles, R., & Gutierrez, C. (2008). Survey of graph database models. ACM Computing Surveys, 40(1), 1–39.
Baskerville, R. L., Kaul, M., & Storey, V. C. (2015). Genres of inquiry in design-science research: Justification and evaluation of knowledge production. MIS Quarterly, 39(3), 541–564.
Bauer, M. W., Petkova, K., & Boyadjieva, P. (2000). Public knowledge of and attitudes to science: Alternative measures that may end the “science war”. Science, Technology and Human Values, 25(1), 30–51.
Bergholtz, M., & Eriksson, O. (2015). Towards a socio-institutional ontology for conceptual modelling of information systems. In Advances in Conceptual Modeling (pp. 225–235). Springer.
Bonney, R. (1996). Citizen science: A lab tradition. Living Bird, 15(4), 7–15.
Bonney, R., Cooper, C. B., Dickinson, J., Kelling, S., Phillips, T., Rosenberg, K. V., et al. (2009). Citizen science: A developing tool for expanding science knowledge and scientific literacy. BioScience, 59(11), 977–984.
Brabham, D. C. (2013). Crowdsourcing. Cambridge, MA: MIT Press.
Brossard, D., Lewenstein, B., & Bonney, R. (2005). Scientific knowledge and attitude change: The impact of a citizen science project. International Journal of Science Education, 27(9), 1099–1121.
Bunge, M. (1977). Treatise on basic philosophy: Ontology I: the furniture of the world. Boston, MA: Reidel.
Burgess, H., DeBey, L., Froehlich, H., Schmidt, N., Theobald, E., Ettinger, A., … Parrish, J. (2017). The science of citizen science: Exploring barriers to use as a primary research tool. Biological Conservation, 1–8.
Castellanos, A., Castillo, A., Lukyanenko, R., & Tremblay, M. C. (2017). Understanding Benefits and Limitations of Unstructured Data Collection for Repurposing Organizational Data. In S. Wrycza & J. Maślankowski (Eds.), Information Systems: Research, Development, Applications, Education: 10th SIGSAND/PLAIS EuroSymposium 2017, Gdansk, Poland, September 22, 2017, Proceedings (pp. 13–24). Cham: Springer International Publishing.
Castellanos, A., Lukyanenko, R., Samuel, B. M., & Tremblay, M. C. (2016). Conceptual modeling in open information environments. In AIS SIGSAND Symposium (pp. 1–7). Lubbock, TX.
Cattell, R. (2011). Scalable SQL and NoSQL data stores. ACM SIGMOD Record, 39(4), 12–27.
Chandra Kruse, L., & Seidel, S. (2017). Tensions in design principle formulation and reuse. In DESRIST 2017.
Chandra Kruse, L., Seidel, S., & Purao, S. (2016). Making Use of Design Principles. LNCS 9661 (pp. 37–51). Berlin, Heidelberg: Springer.
Chandra, L., Seidel, S., & Gregor, S. (2015). prescriptive knowledge in is research: conceptualizing design principles in terms of materiality, action, and boundary conditions. Hawaii International Conference on System Sciences, pp. 4039–4047.
Chang, F., Dean, J., Ghemawat, S., Hsieh, W. C., Wallach, D. A., Burrows, M., … & Gruber, R. E. (2008). Bigtable: A distributed storage system for structured data. ACM Transactions on Computer Systems, 26(2), 4–23.
Davis, E., & Marcus, G. (2015). Commonsense reasoning and commonsense knowledge in artificial intelligence. Communications of the ACM, 58(9), 92–103.
DeCandia, G., Hastorun, D., Jampani, M., Kakulapati, G., Lakshman, A., Pilchin, A., … & Vogels, W. (2007). Dynamo: amazon’s highly available key-value store. In ACM SIGOPS Operating Systems Review (Vol. 41, pp. 205–220). ACM.
Doan, A., Ramakrishnan, R., & Halevy, A. Y. (2011). Crowdsourcing systems on the World-Wide Web. Communications of the ACM, 54(4), 86–96.
Eriksson, O., & Agerfalk, P. J. (2010). Rethinking the meaning of identifiers in information infrastructures. Journal of the Association for Information Systems, 11(8), 433–454.
Faraj, S., Jarvenpaa, S. L., & Majchrzak, A. (2011). Knowledge collaboration in online communities. Organization Science, 22(5), 1224–1239.
Feyerabend, P. (1980). Science in a free society. Philosophical Quarterly, 30(119), 172–174.
Fielden, M. A., Chaulk, A. C., Bassett, K., Wiersma, Y. F., Erbland, M., Whitney, H., et al. (2015). Aedes japonicus japonicus (Diptera: Culicidae) arrives at the most easterly point in North America. The Canadian Entomologist, 147(06), 737–740.
Fortson, L., Masters, K., Nichol, R., Borne, K., Edmondson, E., Lintott, C., … & Wallin, J. (2011). Galaxy Zoo: Morphological classification and citizen science. Advances in Machine Learning and Data Mining for Astronomy, 1–11.
Franklin, M. J., Kossmann, D., Kraska, T., Ramesh, S., & Xin, R. (2011). CrowdDB: Answering queries with crowdsourcing. In Proceedings of the 2011 ACM SIGMOD International Conference on Management of data (pp. 61–72). Athens, Greece: ACM.
Garcia-Molina, H., Joglekar, M., Marcus, A., Parameswaran, A., & Verroios, V. (2016). Challenges in data crowdsourcing. IEEE Transactions on Knowledge and Data Engineering, 28(4), 901–911.
Goodchild, M. (2007). Citizens as sensors: The world of volunteered geography. GeoJournal, 69(4), 211–221.
Gregor, S., & Jones, D. (2007). The Anatomy of design theory. Journal of the Association for Information Systems, 8(5), 312–335.
Haklay, M. (2013). Citizen science and volunteered geographic information: Overview and typology of participation. In Crowdsourcing geographic knowledge (pp. 105–122). Springer.
Hanseth, O., & Lyytinen, K. (2004). Theorizing about the design of Information Infrastructures: design kernel theories and principles. Sprouts: Working papers on information environments, systems and organizations, 4(4), 207–241.
Hevner, A., March, S., Park, J., & Ram, S. (2004). Design science in information systems research. MIS Quarterly, 28(1), 75–105.
Hosseini, M., Phalp, K., Taylor, J., & Ali, R. (2014). The four pillars of crowdsourcing: A reference model (pp. 1–12). Presented at the 2014 IEEE Eighth International Conference on Research Challenges in Information Science (RCIS), IEEE.
Howe, J. (2006). The rise of crowdsourcing. Wired, 14(6).
Iivari, J. (2007). A paradigmatic analysis of information systems as a design science. Scandinavian Journal of Information Systems, 39–64.
Ipeirotis, P. G., & Gabrilovich, E. (2014). Quizz: Targeted crowdsourcing with a billion (potential) users (pp. 143–154). Presented at the Proceedings of the 23rd international conference on World wide web, International World Wide Web Conferences Steering Committee.
Ipeirotis, P. G., & Paritosh, P. K. (2011). Managing crowdsourced human computation: a tutorial. In Proceedings of the 20th international conference companion on World wide web. ACM.
Ipeirotis, P. G., Provost, F., & Wang, J. (2010). Quality management on amazon mechanical turk. In Proceedings of the ACM SIGKDD workshop on human computation (pp. 64–67). ACM.
Irwin, A. (1995). Citizen science: A study of people, expertise and sustainable development. Psychology Press.
Kallinikos, J., & Tempini, N. (2014). Patient data as medical facts: Social media practices as a foundation for medical knowledge creation. Information Systems Research, 25(4), 817–833.
Kilian, D. (2008). When it absolutely has to be accurate, don’t trust the crowd. ECT News Network.
Kleek, M. G. V., Styke, W., Schraefel, M., & Karger, D. (2011). Finders/Keepers: A longitudinal study of people managing information scraps in a micro-note tool. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (pp. 2907–2916). Vancouver, BC, Canada: ACM.
Kosmala, M., Wiggins, A., Swanson, A., & Simmons, B. (2016). Assessing data quality in citizen science. Frontiers in Ecology and the Environment, 14(10), 551–560.
Kullenberg, C., & Kasperowski, D. (2016). What is citizen science?–A scientometric meta-analysis. PLoS ONE, 11(1), e0147152.
Lassaline, M. E., Wisniewski, E. J., & Medin, D. L. (1992). Basic levels in artificial and natural categories: Are all basic levels created equal? In B. Barbara (Ed.), Percepts, Concepts and Categories: The Representation and Processing of Information (Vol. Volume 93, pp. 328–378). North-Holland.
Levy, M., & Germonprez, M. (2017). The potential for citizen science in information systems research. Communications of the Association for Information Systems, 40(1), 2.
Lewandowski, E., & Specht, H. (2015). Influence of volunteer and project characteristics on data quality of biological surveys. Conservation Biology, 29(3), 713–723.
Li, G., Wang, J., Zheng, Y., & Franklin, M. (2016). Crowdsourced data management: A survey. IEEE Transactions on Knowledge and Data Engineering, 1–23.
Light, A., & Miskelly, C. (2014). Design for sharing. Northumbria University/The Sustainable Society Network.
Lintott, C. J., Schawinski, K., Keel, W., Arkel, H. V., Bennert, N., Edmondson, E., … & Vandenberg, J. (2009). Galaxy Zoo: Hanny’s Voorwerp, a quasar light echo? Monthly Notices of the Royal Astronomical Society, 399(1), 129–140.
Loos, P., Nebel, W., Gómez, J. M., Hasan, H., Watson, R. T., vom Brocke, J., … & Recker, J. (2011). Green IT: a matter of business and information systems engineering? Business & Information Systems Engineering, 3(4), 245–252.
Louv, R., Dickinson, J. L., & Bonney, R. (2012). Citizen science: Public participation in environmental research. Ithaca, NY: Cornell University Press.
Lukyanenko, R., Evermann, J., & Parsons, J. (2014). Instantiation validity in IS design research. In DESRIST 2014, LNCS 8463 (pp. 321–328). Springer.
Lukyanenko, R., Evermann, J., & Parsons, J. (2015). Guidelines for Establishing Instantiation Validity in IT Artifacts: A Survey of IS Research. In DESRIST 2015, LNCS 9073. Berlin/ Heidelberg: Springer.
Lukyanenko, R., & Parsons, J. (2011a). Rethinking data quality as an outcome of conceptual modeling choices (pp. 1–16). Adelaide: Australia.
Lukyanenko, R., & Parsons, J. (2011b). Unintended consequences of class-based ontological commitment. In O. De Troyer, C. Bauzer Medeiros, R. Billen, P. Hallot, A. Simitsis, & H. Van Mingroot (Eds.), (Vol. 6999, pp. 220–229). Berlin, Heidelberg: Springer.
Lukyanenko, R., & Parsons, J. (2013a). Is traditional conceptual modeling becoming obsolete? In Conceptual Modeling (pp. 1–14).
Lukyanenko, R., & Parsons, J. (2013b). Reconciling theories with design choices in design science research. In DESRIST 2013, LNCS 7939 (pp. 165–180). Springer Berlin, Heidelberg.
Lukyanenko, R., & Parsons, J. (2015). Information quality research challenge: Adapting information quality principles to user-generated content. ACM Journal of Data and Information Quality, 6(1), 1–3.
Lukyanenko, R., Parsons, J., & Samuel, B. M. (2018). Artifact sampling: using multiple information technology artifacts to increase research rigor. In Proceedings of the 51st Hawaii International Conference on System Sciences (HICSS 2018) (pp. 1–12). Big Island, Hawaii.
Lukyanenko, R., Parsons, J., & Samuel, B. M. (2019a). Representing instances: The case for reengineering conceptual modeling grammars. European Journal of Information Systems, 28(1), 68–90.
Lukyanenko, R., Parsons, J., & Wiersma, Y. (2011). Citizen Science 2.0: Data management principles to harness the power of the crowd. In H. Jain, A. Sinha, & P. Vitharana (Eds.), (Vol. 6629, pp. 465–473). Springer Berlin, Heidelberg.
Lukyanenko, R., Parsons, J., & Wiersma, Y. (2014b). The IQ of the crowd: Understanding and improving information quality in structured user-generated content. Information systems Research, 25(4), 669–689.
Lukyanenko, R., Parsons, J., & Wiersma, Y. (2016a). Emerging problems of data quality in citizen science. Conservation Biology, 30(3), 447–449.
Lukyanenko, R., Parsons, J., Wiersma, Y. F., Wachinger, G., Huber, B., & Meldt, R. (2017). Representing crowd knowledge: Guidelines for conceptual modeling of user-generated content. Journal of the Association for Information Systems, 18(4), 297–339.
Lukyanenko, R., Parsons, J., Wiersma, Y., & Maddah, M. (2019b). Expecting the unexpected: effects of data collection design choices on the quality of crowdsourced user-generated content. MIS Quarterly, 43(2), 634–647.
Lukyanenko, R., Parsons, J., Wiersma, Y., Sieber, R., & Maddah, M. (2016b). Participatory design for user-generated content: understanding the challenges and moving forward. Scandinavian Journal of Information Systems, 28(1), 37–70.
Lukyanenko, R., & Samuel, B. M. (2017). Are all classes created equal? Increasing precision of conceptual modeling grammars. ACM Transactions on Management Information Systems (TMIS), 40(2), 1–25.
Lukyanenko, R., Wiggins, A., & Rosser, H. K. (2019). Citizen science: An information quality research frontier. Information Systems Frontiers, 1–23.
March, S., & Allen, G. (2012). Toward a social ontology for conceptual modeling. In 11th Symposium on Research in Systems Analysis and Design (pp. 57–62). Vancouver, Canada.
March, S. T., & Allen, G. N. (2014). Toward a social ontology for conceptual modeling. Communications of the AIS, 34.
McKinley, D. C., Miller-Rushing, A. J., Ballard, H. L., Bonney, R., Brown, H., Cook-Patton, S. C., … & Phillips, T. B. (2016). Citizen science can improve conservation science, natural resource management, and environmental protection. Biological Conservation.
Murphy, G. (2004). The big book of concepts. Cambridge, MA: MIT Press.
Nov, O., Arazy, O., & Anderson, D. (2011). Dusting for science: Motivation and participation of digital citizen science volunteers. 2011 iConference, 68–74.
Nov, O., Arazy, O., & Anderson, D. (2014). Scientists@ home: what drives the quantity and quality of online citizen science participation. PLoS ONE, 9(4), 1–11.
Ogilvie, B. W. (2008). The science of describing: Natural history in Renaissance Europe. Chicago, IL: University of Chicago Press.
Ogunseye, S., Parsons, J., & Lukyanenko, R. (2017). Do crowds go stale? Exploring the Effects of Crowd Reuse on Data Diversity. In WITS 2017. Seoul, South Korea.
Osborn, D. A., Pearse, J. S., Roe, C. A., Magoon, O., Converse, H., Baird, B., … & Miller-Henson, M. (2005). Monitoring rocky intertidal shorelines: a role for the public in resource management (pp. 624–636). Presented at the California and the World Ocean ‘02, conf. proc. American Society of Civil Engineers, Reston, VA.
Parsons, J. (1996). An information model based on classification theory. Management Science, 42(10), 1437–1453.
Parsons, J., Lukyanenko, R., & Wiersma, Y. (2011). Easier citizen science is better. Nature, 471(7336), 37.
Parsons, J., & Wand, Y. (2000). Emancipating instances from the tyranny of classes in information modeling. ACM Transactions on Database Systems, 25(2), 228–268.
Parsons, J., & Wand, Y. (2008a). A question of class. Nature, 455, 1040–1041.
Parsons, J., & Wand, Y. (2008b). Using cognitive principles to guide classification in information systems modeling. MIS Quarterly, 32(4), 839–868.
Perszyk, K. J. (2013). Nonexistent objects: Meinong and contemporary philosophy. Netherlands: Springer.
Ponti, M., Hillman, T., Kullenberg, C., & Kasperowski, D. (2018). Getting it right or being top rank: Games in citizen science. Citizen Science: Theory and Practice, 3(1).
Popper, K. (2014). Conjectures and refutations: The growth of scientific knowledge. Cambridge, Mass: Routledge.
Pultar, E., Raubal, M., Cova, T. J., & Goodchild, M. F. (2009). Dynamic GIS case studies: Wildfire evacuation and volunteered geographic information. Transactions in GIS, 13(s1), 85–104.
Quinn, A. J., & Bederson, B. B. (2011). Human computation: a survey and taxonomy of a growing field (pp. 1403–1412). Presented at the Proceedings of the SIGCHI conference on human factors in computing systems, ACM.
Rosch, E., Mervis, C. B., Gray, W. D., Johnson, D. M., & Boyesbraem, P. (1976). Basic objects in natural categories. Cognitive Psychology, 8(3), 382–439.
Sauermann, H., & Franzoni, C. (2015). Crowd science user contribution patterns and their implications. Proceedings of the National Academy of Sciences, 112(3), 679–684.
Schuurman, D., Baccarne, B., De Marez, L., & Mechant, P. (2012). Smart ideas for smart cities: investigating crowdsourcing for generating and selecting ideas for ICT innovation in a city context. Journal of Theoretical and Applied Electronic commerce Research, 7(3), 49–62.
Searle, J. R. (1995). The construction of social reality. Simon and Schuster.
Seidel, S., Recker, J. C., & Vom Brocke, J. (2013). Sensemaking and sustainable practicing: functional affordances of information systems in green transformations. Management Information Systems Quarterly, 37(4), 1275–1299.
Sein, M., Henfridsson, O., Purao, S., Rossi, M., & Lindgren, R. (2011). Action design research. MIS Quarterly, 35(1), 37.
Smith, E., & Medin, D. (1981). Categories and concepts. Cambridge, Mass: Harvard University Press.
Sonnenberg, C., & vom Brocke, J. (2012). Evaluations in the science of the artificial–reconsidering the build-evaluate pattern in design science research. In Design Science Research in Information Systems (pp. 381–397). Springer.
Stevens, M., Vitos, M., Lewis, J., & Haklay, M. (2013). Participatory monitoring of poaching in the Congo basin. Presented at the 21st GIS Research UK Conference.
Sullivan, B. L., Wood, C. L., Iliff, M. J., Bonney, R. E., Fink, D., & Kelling, S. (2009). eBird: A citizen-based bird observation network in the biological sciences. Biological Conservation, 142(10), 2282–2292.
Surowiecki, J. (2005). The wisdom of crowds. New York, NY: Anchor Books.
Tams, S., Grover, V., & Thatcher, J. (2014). Modern information technology in an old workforce: toward a strategic research agenda. The Journal of Strategic Information Systems, 23(4), 284–304.
Theobald, E. J., Ettinger, A. K., Burgess, H. K., DeBey, L. B., Schmidt, N. R., Froehlich, H. E., … & Harsch, M. A. (2015). Global change and local solutions: Tapping the unrealized potential of citizen science for biodiversity research. Biological Conservation, 181, 236–244.
Venable, J., Pries-Heje, J., & Baskerville, R. (2012). A comprehensive framework for evaluation in design science research. In DESRIST 2012, LNCS 7286 (pp. 423–438). Springer.
Vitos, M., Lewis, J., Stevens, M., & Haklay, M. (2013). Making local knowledge matter: Supporting non-literate people to monitor poaching in Congo (pp. 1–10). Presented at the ACM Symposium on Computing for Development, ACM.
Weinmann, M., Schneider, C., & vom Brocke, J. (2016). Digital nudging. Business & Information Systems Engineering, 58(6), 433–436.
Wiersma, Y. F. (2010). Birding 2.0: Citizen science and effective monitoring in the Web 2.0 world. Avian Conservation and Ecology, 5(2), 13.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Lukyanenko, R., Parsons, J. (2020). Easier Crowdsourcing Is Better: Designing Crowdsourcing Systems to Increase Information Quality and User Participation. In: vom Brocke, J., Hevner, A., Maedche, A. (eds) Design Science Research. Cases. Progress in IS. Springer, Cham. https://doi.org/10.1007/978-3-030-46781-4_3
Download citation
DOI: https://doi.org/10.1007/978-3-030-46781-4_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-46780-7
Online ISBN: 978-3-030-46781-4
eBook Packages: Business and ManagementBusiness and Management (R0)