Skip to main content
Top
Published in: Journal of Science Education and Technology 2/2021

10-02-2021

Practices and Theories: How Can Machine Learning Assist in Innovative Assessment Practices in Science Education

Author: Xiaoming Zhai

Published in: Journal of Science Education and Technology | Issue 2/2021

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

As cutting-edge technologies, such as machine learning (ML), are increasingly involved in science assessments, it is essential to conceptualize how assessment practices are innovated by technologies. To partially meet this need, this article focuses on ML-based science assessments and elaborates on how ML innovates assessment practices in science education. The article starts with an articulation of the “practice” nature of assessment both of learning and for learning, identifying four essential assessment practices: identifying learning goals, eliciting performance, interpreting observations, and decision-making and action-taking. I then extend a three-dimensional framework for innovative assessments, including construct, functionality, and automaticity, and based on which to conceptualize innovative assessments in three levels: substitute, transform, and redefine. Using the framework, I elaborate on how the 10 articles included in this special issue, Applying Machine Learning in Science Assessment: Opportunity and Challenge, advanced our knowledge of the innovations that ML brought to science assessment practices. I contend that the 10 articles exemplify a great deal of effort to transform the four components of assessment practices: ML allows assessments to target complex, diverse, and structural constructs, and thus better approaching the three-dimensional science learning goals of the Next Generation Science Standards (NGSS Lead States, 2013); ML extends the approaches used to eliciting performance and collecting evidence; ML provides a means to better interpreting observations and using evidence; ML supports immediate and complex decision-making and action-taking. I conclude this article by pushing the field to consider the underlying educational theories that are needed for innovative assessment practices and the necessities of establishing a “romance” between assessment practices and the relevant educational theories, which I contend are the prominent challenges to forward innovative and ML-based assessment practices in science education.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Footnotes
1
Author note. In the preparation of the manuscript From Substitution to Redefinition: A Framework of Machine Learning-based Science Assessment (hereby called ML Framework; Zhai et al., 2020a). I benefited from a long conversation with my co-author Mark Urban-Lurain, the co-founder of Automatic Analysis of Constructed Responses (AACR) group at Michigan State University, and Kevin Haudek, the current lead Principal Investigator of AACR. Urban-Lurain has dedicated more than 20 years in studying the use of technology in STEM education, particularly on constructed response assessments, and had valuable insights about multiple aspects of applying technology, such as machine learning (ML), in advancing STEM assessments. The conversation started from a comment Urban-Lurain made on how we should conceptualize “innovative assessments.” More specifically, he urged me to think why we could possibly use the word “redefinition” to portray “innovative assessment.” If assessment is still in forms of multiple-choice, constructed responses, etc., and teachers and students are involved in activities that they used to do, what do we mean by “innovative” to a degree that the assessment may be “redefined” by ML? This section provides information partially serving as a response to Urban-Lurain’ question.
 
Literature
go back to reference Abd-El-Khalick, F., Boujaoude, S., Duschl, R., Lederman, N. G., Mamlok-Naaman, R., & Hofstein, A. (2004). Inquiry in science education: international perspectives. Science Education, 88(3), 397–419.CrossRef Abd-El-Khalick, F., Boujaoude, S., Duschl, R., Lederman, N. G., Mamlok-Naaman, R., & Hofstein, A. (2004). Inquiry in science education: international perspectives. Science Education, 88(3), 397–419.CrossRef
go back to reference Anderson, L. W., Krathwohl, D. R., & Bloom, B. S. (2001). A taxonomy for learning, teaching, and assessing: a revision of Bloom's taxonomy of educational objectives. Longman Anderson, L. W., Krathwohl, D. R., & Bloom, B. S. (2001). A taxonomy for learning, teaching, and assessing: a revision of Bloom's taxonomy of educational objectives. Longman
go back to reference Bennett, R. E. (2018). Educational assessment: what to watch in a rapidly changing world. Educational Measurement: Issues and Practice, 37(4), 7–15CrossRef Bennett, R. E. (2018). Educational assessment: what to watch in a rapidly changing world. Educational Measurement: Issues and Practice, 37(4), 7–15CrossRef
go back to reference Bennett, R. E., Deane, P., & van Rijn, W. P. (2016). From cognitive-domain theory to assessment practice. Educational Psychologist, 51(1), 82–107.CrossRef Bennett, R. E., Deane, P., & van Rijn, W. P. (2016). From cognitive-domain theory to assessment practice. Educational Psychologist, 51(1), 82–107.CrossRef
go back to reference Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education: Principles, Policy & Practice, 5(1), 7–74. Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education: Principles, Policy & Practice, 5(1), 7–74.
go back to reference Chi, M. T., Feltovich, P. J., & Glaser, R. (1981). Categorization and representation of physics problems by experts and novices. Cognitive Science, 5(2), 121–152CrossRef Chi, M. T., Feltovich, P. J., & Glaser, R. (1981). Categorization and representation of physics problems by experts and novices. Cognitive Science, 5(2), 121–152CrossRef
go back to reference Clark, R. E. (1983). Reconsidering research on learning from media. Review of Educational Research, 53(4), 445–459CrossRef Clark, R. E. (1983). Reconsidering research on learning from media. Review of Educational Research, 53(4), 445–459CrossRef
go back to reference Clement, J. (2000). Model based learning as a key research area for science education. International Journal of Science Education, 22(9), 1041–1053CrossRef Clement, J. (2000). Model based learning as a key research area for science education. International Journal of Science Education, 22(9), 1041–1053CrossRef
go back to reference Cronbach, L. J., & Meehl, P. E. (1955). Construct validity in psychological tests. Psychological Bulletin, 52(4), 1–28CrossRef Cronbach, L. J., & Meehl, P. E. (1955). Construct validity in psychological tests. Psychological Bulletin, 52(4), 1–28CrossRef
go back to reference Darling-Hammond, L. (2014). Next generation assessment: Moving beyond the bubble test to support 21st century learning. John Wiley & Sons Darling-Hammond, L. (2014). Next generation assessment: Moving beyond the bubble test to support 21st century learning. John Wiley & Sons
go back to reference DeBoer, G. E., Quellmalz, E. S., Davenport, J. L., Timms, M. J., Herrmann-Abell, C. F., & Buckley, B. C. (2014). Comparing three online testing modalities: using static, active, and interactive online testing modalities to assess middle school students’ understanding of fundamental ideas and use of inquiry skills related to ecosystems. Journal of Research in Science Teaching, 51(4), 523–554.CrossRef DeBoer, G. E., Quellmalz, E. S., Davenport, J. L., Timms, M. J., Herrmann-Abell, C. F., & Buckley, B. C. (2014). Comparing three online testing modalities: using static, active, and interactive online testing modalities to assess middle school students’ understanding of fundamental ideas and use of inquiry skills related to ecosystems. Journal of Research in Science Teaching, 51(4), 523–554.CrossRef
go back to reference Duschl, R. (2008). Science education in three-part harmony: balancing conceptual, epistemic, and social learning goals. Review of Research in Education, 32(1), 268–291CrossRef Duschl, R. (2008). Science education in three-part harmony: balancing conceptual, epistemic, and social learning goals. Review of Research in Education, 32(1), 268–291CrossRef
go back to reference Frezzo, D. C., Behrens, J. T., & Mislevy, R. J. (2010). Design patterns for learning and assessment: facilitating the introduction of a complex simulation-based learning environment into a community of instructors. Journal of Science Education and Technology, 19(2), 105–114. https://doi.org/10.1007/s10956-009-9192-0.CrossRef Frezzo, D. C., Behrens, J. T., & Mislevy, R. J. (2010). Design patterns for learning and assessment: facilitating the introduction of a complex simulation-based learning environment into a community of instructors. Journal of Science Education and Technology, 19(2), 105–114. https://​doi.​org/​10.​1007/​s10956-009-9192-0.CrossRef
go back to reference Gobert, J. D., & Pallant, A. (2004). Fostering students’ epistemologies of models via authentic model-based tasks. Journal of Science Education and Technology, 13(1), 7–22CrossRef Gobert, J. D., & Pallant, A. (2004). Fostering students’ epistemologies of models via authentic model-based tasks. Journal of Science Education and Technology, 13(1), 7–22CrossRef
go back to reference Hickey, D. T., Taasoobshirazi, G., & Cross, D. (2012). Assessment as learning: enhancing discourse, understanding, and achievement in innovative science curricula. Journal of Research in Science Teaching, 49(10), 1240–1270.CrossRef Hickey, D. T., Taasoobshirazi, G., & Cross, D. (2012). Assessment as learning: enhancing discourse, understanding, and achievement in innovative science curricula. Journal of Research in Science Teaching, 49(10), 1240–1270.CrossRef
go back to reference Jescovitch, L. N., Scott, E. E., Cerchiara, J. A., Merrill, J., Urban-Lurain, M., Doherty, J. H., & Haudek, K. C. (2020). Comparison of machine learning performance using analytic and holistic coding approaches across constructed response assessments aligned to a science learning progression. Journal of Science Education and Technology, 1–18 Jescovitch, L. N., Scott, E. E., Cerchiara, J. A., Merrill, J., Urban-Lurain, M., Doherty, J. H., & Haudek, K. C. (2020). Comparison of machine learning performance using analytic and holistic coding approaches across constructed response assessments aligned to a science learning progression. Journal of Science Education and Technology, 1–18
go back to reference Kane, M. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50(1), 1–73CrossRef Kane, M. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50(1), 1–73CrossRef
go back to reference Kelly, G. J., McDonald, S., & Wickman, P. O. (2012). Science learning and epistemology. In Second international handbook of science education (pp. 281–291). Springer Kelly, G. J., McDonald, S., & Wickman, P. O. (2012). Science learning and epistemology. In Second international handbook of science education (pp. 281–291). Springer
go back to reference Krajcik, J. S., & Mun, K. (2014). Promises and challenges of using learning technologies to promote student learning of science. Handbook of Research on Science Education, 2, 337–360 Krajcik, J. S., & Mun, K. (2014). Promises and challenges of using learning technologies to promote student learning of science. Handbook of Research on Science Education, 2, 337–360
go back to reference Lamb, R., Hand, B., & Kavner, A. (2020). Computational modeling of the effects of the science writing heuristic on student critical thinking in science using machine learning. Journal of Science Education and Technology, 1–15 Lamb, R., Hand, B., & Kavner, A. (2020). Computational modeling of the effects of the science writing heuristic on student critical thinking in science using machine learning. Journal of Science Education and Technology, 1–15
go back to reference Lee, H. S., Gweon, G. H., Lord, T., Paessel, N., Pallant, A., & Pryputniewicz, S. (2021). Machine learning-enabled automated feedback: supporting students’ revision of scientific arguments based on data drawn from simulation. Journal of Science Education and Technology. https://doi.org/10.1007/s10956-020-09889-7. Lee, H. S., Gweon, G. H., Lord, T., Paessel, N., Pallant, A., & Pryputniewicz, S. (2021). Machine learning-enabled automated feedback: supporting students’ revision of scientific arguments based on data drawn from simulation. Journal of Science Education and Technology. https://​doi.​org/​10.​1007/​s10956-020-09889-7.
go back to reference Liaw, H., Yu, Y. R., Chou, C. C., & Chiu, M. H. (2020). Relationships between facial expressions, prior knowledge, and multiple representations: a case of conceptual change for kinematics instruction. Journal of Science Education and Technology, 1–12 Liaw, H., Yu, Y. R., Chou, C. C., & Chiu, M. H. (2020). Relationships between facial expressions, prior knowledge, and multiple representations: a case of conceptual change for kinematics instruction. Journal of Science Education and Technology, 1–12
go back to reference Magnusson, S., Krajcik, J., & Borko, H. (1999). Nature, sources, and development of pedagogical content knowledge for science teaching. In Examining pedagogical content knowledge (pp. 95–132). Springer Magnusson, S., Krajcik, J., & Borko, H. (1999). Nature, sources, and development of pedagogical content knowledge for science teaching. In Examining pedagogical content knowledge (pp. 95–132). Springer
go back to reference Messick, S. (1994). The interplay of evidence and consequences in the validation of performance assessments. Educational researcher, 23(2), 13–23CrossRef Messick, S. (1994). The interplay of evidence and consequences in the validation of performance assessments. Educational researcher, 23(2), 13–23CrossRef
go back to reference Mislevy, R., & Haertel, G. (2006). Implications of evidence-centered design for educational testing. Educational measurement: issues and practice, 25(4), 6–20CrossRef Mislevy, R., & Haertel, G. (2006). Implications of evidence-centered design for educational testing. Educational measurement: issues and practice, 25(4), 6–20CrossRef
go back to reference Mislevy, R. J. (2016). How developments in psychology and technology challenge validity argumentation. Journal of Educational Measurement, 53(3), 265–292CrossRef Mislevy, R. J. (2016). How developments in psychology and technology challenge validity argumentation. Journal of Educational Measurement, 53(3), 265–292CrossRef
go back to reference National Research Council. (2012). A framework for K-12 science education: practices, crosscutting concepts, and core ideas. National Academies Press National Research Council. (2012). A framework for K-12 science education: practices, crosscutting concepts, and core ideas. National Academies Press
go back to reference NGSS Lead States. (2013). Next generation science standards: For states, by states. National Academies Press. NGSS Lead States. (2013). Next generation science standards: For states, by states. National Academies Press.
go back to reference Nicolaidou, I., Kyza, E. A., Terzian, F., Hadjichambis, A., & Kafouris, D. (2011). A framework for scaffolding students’ assessment of the credibility of evidence. Journal of Research in Science Teaching, 48(7), 711–744CrossRef Nicolaidou, I., Kyza, E. A., Terzian, F., Hadjichambis, A., & Kafouris, D. (2011). A framework for scaffolding students’ assessment of the credibility of evidence. Journal of Research in Science Teaching, 48(7), 711–744CrossRef
go back to reference Osborne, J. (2010). Arguing to learn in science: the role of collaborative, critical. Science, 1183944(463), 328 Osborne, J. (2010). Arguing to learn in science: the role of collaborative, critical. Science, 1183944(463), 328
go back to reference Osborne, J. F., Henderson, J. B., MacPherson, A., Szu, E., Wild, A., & Yao, S. Y. (2016). The development and validation of a learning progression for argumentation in science. Journal of Research in Science Teaching, 53(6), 821–846CrossRef Osborne, J. F., Henderson, J. B., MacPherson, A., Szu, E., Wild, A., & Yao, S. Y. (2016). The development and validation of a learning progression for argumentation in science. Journal of Research in Science Teaching, 53(6), 821–846CrossRef
go back to reference Pellegrino, J. W. (2018). Sciences of learning and development: some thoughts from the learning sciences. Applied Developmental Science, 1–9 Pellegrino, J. W. (2018). Sciences of learning and development: some thoughts from the learning sciences. Applied Developmental Science, 1–9
go back to reference Pellegrino, J. W., Chudowsky, N., & Glaser, R. (2001). Knowing what students know: the science and design of educational assessment. ERIC Pellegrino, J. W., Chudowsky, N., & Glaser, R. (2001). Knowing what students know: the science and design of educational assessment. ERIC
go back to reference Pellegrino, J. W., Wilson, M. R., Koenig, J. A., & Beatty, A. S. (2014). Developing assessments for the Next Generation Science Standards. ERIC Pellegrino, J. W., Wilson, M. R., Koenig, J. A., & Beatty, A. S. (2014). Developing assessments for the Next Generation Science Standards. ERIC
go back to reference Penfield, R. D., & Lee, O. (2010). Test-based accountability: potential benefits and pitfalls of science assessment with student diversity. Journal of Research in Science Teaching, 47(1), 6–24CrossRef Penfield, R. D., & Lee, O. (2010). Test-based accountability: potential benefits and pitfalls of science assessment with student diversity. Journal of Research in Science Teaching, 47(1), 6–24CrossRef
go back to reference Rosenberg, J. M., & Krist, C. (2020). Combining machine learning and qualitative methods to elaborate students’ ideas about the generality of their model-based explanations. Journal of Science Education and Technology, 1–13 Rosenberg, J. M., & Krist, C. (2020). Combining machine learning and qualitative methods to elaborate students’ ideas about the generality of their model-based explanations. Journal of Science Education and Technology, 1–13
go back to reference Ruiz-Primo, M. A., & Furtak, E. M. (2007). Exploring teachers’ informal formative assessment practices and students’ understanding in the context of scientific inquiry. Journal of Research in Science Teaching, 44(1), 57–84CrossRef Ruiz-Primo, M. A., & Furtak, E. M. (2007). Exploring teachers’ informal formative assessment practices and students’ understanding in the context of scientific inquiry. Journal of Research in Science Teaching, 44(1), 57–84CrossRef
go back to reference Schwarz, C. V., Reiser, B. J., Davis, E. A., Kenyon, L., Achér, A., & Fortus, D. (2009). Developing a learning progression for scientific modeling: making scientific modeling accessible and meaningful for learners. Journal of Research in Science Teaching, 46(6), 632–654 Schwarz, C. V., Reiser, B. J., Davis, E. A., Kenyon, L., Achér, A., & Fortus, D. (2009). Developing a learning progression for scientific modeling: making scientific modeling accessible and meaningful for learners. Journal of Research in Science Teaching, 46(6), 632–654
go back to reference Shavelson, R. J., Young, D. B., Ayala, C. C., Brandon, P. R., Furtak, E. M., Ruiz-Primo, M. A., & Yin, Y. (2008). On the impact of curriculum-embedded formative assessment on learning: a collaboration between curriculum and assessment developers. Applied measurement in education, 21(4), 295–314CrossRef Shavelson, R. J., Young, D. B., Ayala, C. C., Brandon, P. R., Furtak, E. M., Ruiz-Primo, M. A., & Yin, Y. (2008). On the impact of curriculum-embedded formative assessment on learning: a collaboration between curriculum and assessment developers. Applied measurement in education, 21(4), 295–314CrossRef
go back to reference Shepard, L. A., Penuel, W. R., & Pellegrino, J. W. (2018). Using learning and motivation theories to coherently link formative assessment, grading practices, and large-scale assessment. Educational measurement: issues and practice, 37(1), 21–34CrossRef Shepard, L. A., Penuel, W. R., & Pellegrino, J. W. (2018). Using learning and motivation theories to coherently link formative assessment, grading practices, and large-scale assessment. Educational measurement: issues and practice, 37(1), 21–34CrossRef
go back to reference Sung, S. H., Li, C., Chen, G., Huang, X., Xie, C., Massicotte, J., & Shen, J. (2020). How does augmented observation facilitate multimodal representational thinking? Applying deep learning to decode complex student construct. Journal of Science Education and Technology, 1–17 Sung, S. H., Li, C., Chen, G., Huang, X., Xie, C., Massicotte, J., & Shen, J. (2020). How does augmented observation facilitate multimodal representational thinking? Applying deep learning to decode complex student construct. Journal of Science Education and Technology, 1–17
go back to reference Wang, C., Liu, X., Wang, L., Sun, Y., & Zhang, H. (2020). Automated scoring of Chinese grades 7–9 students’ competence in interpreting and arguing from evidence. Journal of Science Education and Technology, 1–14 Wang, C., Liu, X., Wang, L., Sun, Y., & Zhang, H. (2020). Automated scoring of Chinese grades 7–9 students’ competence in interpreting and arguing from evidence. Journal of Science Education and Technology, 1–14
go back to reference Wiliam, D. (2011). What is assessment for learning? Studies in Educational Evaluation, 37(1), 3–14CrossRef Wiliam, D. (2011). What is assessment for learning? Studies in Educational Evaluation, 37(1), 3–14CrossRef
go back to reference Yoo, J., & Kim, J. (2014). Can Online Discussion Participation Predict Group Project Performance? Investigating the Roles of Linguistic Features and Participation Patterns. International Journal of Artificial Intelligence in Education, 24(1), 8–32.CrossRef Yoo, J., & Kim, J. (2014). Can Online Discussion Participation Predict Group Project Performance? Investigating the Roles of Linguistic Features and Participation Patterns. International Journal of Artificial Intelligence in Education, 24(1), 8–32.CrossRef
go back to reference Zhai, X., Haudek, K. C., Shi, L., Nehm, R., & Urban-Lurain, M. (2020a). From substitution to redefinition: a framework of machine learning-based science assessment. Journal of Research in Science Teaching, 57(9), 1430–1459. https://doi.org/10.1002/tea.21658. Zhai, X., Haudek, K. C., Shi, L., Nehm, R., & Urban-Lurain, M. (2020a). From substitution to redefinition: a framework of machine learning-based science assessment. Journal of Research in Science Teaching, 57(9), 1430–1459. https://​doi.​org/​10.​1002/​tea.​21658.
go back to reference Zhai, X., Haudek, K. C., Stuhlsatz, M. A., & Wilson, C. (2020b). Evaluation of construct-irrelevant variance yielded by machine and human scoring of a science teacher PCK constructed response assessment. Studies in Educational Evaluation, 67, 100916 Zhai, X., Haudek, K. C., Stuhlsatz, M. A., & Wilson, C. (2020b). Evaluation of construct-irrelevant variance yielded by machine and human scoring of a science teacher PCK constructed response assessment. Studies in Educational Evaluation, 67, 100916
go back to reference Zhai, X., Li, M., & Guo, Y. (2018). Teachers’ use of learning progression-based formative assessment to inform teachers’ instructional adjustment: a case study of two physics teachers’ instruction. International Journal of Science Education, 40(15), 1832–1856CrossRef Zhai, X., Li, M., & Guo, Y. (2018). Teachers’ use of learning progression-based formative assessment to inform teachers’ instructional adjustment: a case study of two physics teachers’ instruction. International Journal of Science Education, 40(15), 1832–1856CrossRef
go back to reference Zhai, X., Yin, Y., Pellegrino, J. W., Haudek, K. C., & Shi, L. (2020e). Applying machine learning in science assessment: a systematic review. Studies in Science Education, 56(1), 111–151 Zhai, X., Yin, Y., Pellegrino, J. W., Haudek, K. C., & Shi, L. (2020e). Applying machine learning in science assessment: a systematic review. Studies in Science Education, 56(1), 111–151
Metadata
Title
Practices and Theories: How Can Machine Learning Assist in Innovative Assessment Practices in Science Education
Author
Xiaoming Zhai
Publication date
10-02-2021
Publisher
Springer Netherlands
Published in
Journal of Science Education and Technology / Issue 2/2021
Print ISSN: 1059-0145
Electronic ISSN: 1573-1839
DOI
https://doi.org/10.1007/s10956-021-09901-8

Other articles of this Issue 2/2021

Journal of Science Education and Technology 2/2021 Go to the issue

Premium Partners