skip to main content
10.1145/2663204.2663264acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
research-article

The Additive Value of Multimodal Features for Predicting Engagement, Frustration, and Learning during Tutoring

Published:12 November 2014Publication History

ABSTRACT

Detecting learning-centered affective states is difficult, yet crucial for adapting most effectively to users. Within tutoring in particular, the combined context of student task actions and tutorial dialogue shape the student's affective experience. As we move toward detecting affect, we may also supplement the task and dialogue streams with rich sensor data. In a study of introductory computer programming tutoring, human tutors communicated with students through a text-based interface. Automated approaches were leveraged to annotate dialogue, task actions, facial movements, postural positions, and hand-to-face gestures. These dialogue, nonverbal behavior, and task action input streams were then used to predict retrospective student self-reports of engagement and frustration, as well as pretest/posttest learning gains. The results show that the combined set of multimodal features is most predictive, indicating an additive effect. Additionally, the findings demonstrate that the role of nonverbal behavior may depend on the dialogue and task context in which it occurs. This line of research identifies contextual and behavioral cues that may be leveraged in future adaptive multimodal systems.

References

  1. Arroyo, I., Cooper, D.G., Burleson, W., Woolf, B.P., Muldner, K. and Christopherson, R.M. 2009. Emotion Sensors Go To School. 14th International Conference on Artificial Intelligence in Education, 17--24. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Cooper, D.G., Muldner, K., Arroyo, I., Woolf, B.P. and Burleson, W. 2010. Ranking Feature Sets for Emotion Models used in Classroom Based Intelligent Tutoring Systems. Proceedings of the 18th International Conference on User Modeling, Adaptation, and Personalization, 135--146. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. D'Mello, S.K. and Graesser, A.C. 2010. Multimodal Semiautomated Affect Detection From Conversational Cues, Gross Body Language, and Facial Features. User Modeling and UserAdapted Interaction. 20, 2, 147--187. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. D'Mello, S.K. and Kory, J. 2012. Consistent but Modest: A Meta-Analysis on Unimodal and Multimodal Affect Detection Accuracies from 30 Studies. Proceedings of the 14th ACM International Conference on Multimodal Interaction, 31--38. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. D'Mello, S.K., Lehman, B., Pekrun, R. and Graesser, A.C. 2014. Confusion Can Be Beneficial for Learning. Learning & Instruction. 29, 153--170.Google ScholarGoogle ScholarCross RefCross Ref
  6. Ekman, P., Friesen, W. V. and Hager, J.C. 2002. Facial Action Coding System. A Human Face.Google ScholarGoogle Scholar
  7. Grafsgaard, J.F., Fulton, R.M., Boyer, K.E., Wiebe, E.N. and Lester, J.C. 2012. Multimodal Analysis of the Implicit Affective Channel in Computer-Mediated Textual Communication. Proceedings of the 14th ACM International Conference on Multimodal Interaction, 145--152. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Grafsgaard, J.F., Wiggins, J.B., Boyer, K.E., Wiebe, E.N. and Lester, J.C. 2013. Automatically Recognizing Facial Expression: Predicting Engagement and Frustration. Proceedings of the 6th International Conference on Educational Data Mining, 43--50.Google ScholarGoogle Scholar
  9. Grafsgaard, J.F., Wiggins, J.B., Boyer, K.E., Wiebe, E.N. and Lester, J.C. 2013. Automatically Recognizing Facial Indicators of Frustration: A Learning-Centric Analysis. Proceedings of the 5th International Conference on Affective Computing and Intelligent Interaction, 159--165. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Harrigan, J.A. and O'Connell, D.M. 1996. How Do You Look When Feeling Anxious? Facial Displays of Anxiety. Personality and Individual Differences. 21, 2, 205--212.Google ScholarGoogle ScholarCross RefCross Ref
  11. Hart, S.G. and Staveland, L.E. 1988. Development of NASATLX (Task Load Index): Results of Empirical and Theoretical Research. Human Mental Workload. P.A. Hancock and N. Meshkati, eds. Elsevier Science. 139--183.Google ScholarGoogle Scholar
  12. Joshi, J., Goecke, R., Parker, G. and Breakspear, M. 2013. Can Body Expressions Contribute to Automatic Depression Analysis? Proceedings of the 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, 1--7.Google ScholarGoogle Scholar
  13. Kapoor, A. and Picard, R.W. 2005. Multimodal Affect Recognition in Learning Environments. Proceedings of the 13th Annual ACM International Conference on Multimedia, 677--682. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Littlewort, G., Bartlett, M.S., Salamanca, L.P. and Reilly, J. 2011. Automated Measurement of Children's Facial Expressions during Problem Solving Tasks. Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, 30--35.Google ScholarGoogle Scholar
  15. Littlewort, G., Whitehill, J., Wu, T., Fasel, I., Frank, M., Movellan, J.R. and Bartlett, M.S. 2011. The Computer Expression Recognition Toolbox (CERT). Proceedings of the IEEE International Conference on Automatic Face and Gesture Recognition, 298--305.Google ScholarGoogle Scholar
  16. Marx, J.D. and Cummings, K. 2007. Normalized Change. American Journal of Physics. 75, 1, 87--91.Google ScholarGoogle ScholarCross RefCross Ref
  17. O'Brien, H.L. and Toms, E.G. 2010. The Development and Evaluation of a Survey to Measure User Engagement. Journal of the American Society for Information Science and Technology. 61, 1, 50--69. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Scherer, S., Stratou, G., Mahmoud, M., Boberg, J., Gratch, J., Rizzo, A. and Morency, L.-P. 2013. Automatic Behavior Descriptors for Psychological Disorder Analysis. Proceedings of the 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, 1--8.Google ScholarGoogle Scholar
  19. Symonds, M.R.E. and Moussalli, A. 2010. A Brief Guide to Model Selection, Multimodel Inference and Model Averaging in Behavioural Ecology using Akaike's Information Criterion. Behavioral Ecology and Sociobiology. 65, 1, 13--21.Google ScholarGoogle ScholarCross RefCross Ref
  20. Vail, A.K. and Boyer, K.E. 2014. Adapting to Personality Over Time: Examining the Effectiveness of Dialogue Policy Progressions in Task-Oriented Interaction. Proceedings of the 15th Annual SIGDIAL Meeting on Discourse and Dialogue, 41--50.Google ScholarGoogle Scholar
  21. Vail, A.K. and Boyer, K.E. 2014. Identifying Effective Moves in Tutoring: On the Refinement of Dialogue Act Annotation Schemes. Proceedings of the 12th International Conference on Intelligent Tutoring Systems, 199--209.Google ScholarGoogle Scholar
  22. VanLehn, K., Graesser, A.C., Jackson, G.T., Jordan, P., Olney, A. and Rosé, C.P. 2007. When Are Tutorial Dialogues More Effective Than Reading? Cognitive Science. 31, 1, 3--62.Google ScholarGoogle ScholarCross RefCross Ref
  23. Wiebe, E.N., Lamb, A., Hardy, M. and Sharek, D. 2014. Measuring Engagement in Video Game-based Environments: Investigation of the User Engagement Scale. Computers in Human Behavior. 32, 123--132. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Woolf, B.P., Burleson, W., Arroyo, I., Dragon, T., Cooper, D.G. and Picard, R.W. 2009. Affect-Aware Tutors: Recognising and Responding to Student Affect. International Journal of Learning Technology. 4, 3--4, 129--164. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Wu, T., Butko, N.J., Ruvolo, P., Whitehill, J., Bartlett, M.S. and Movellan, J.R. 2012. Multi-Layer Architectures for Facial Action Unit Recognition. IEEE Transactions on Systems, Man and Cybernetics, Part B: Cybernetics. 42, 4, 1027--1038.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. The Additive Value of Multimodal Features for Predicting Engagement, Frustration, and Learning during Tutoring

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ICMI '14: Proceedings of the 16th International Conference on Multimodal Interaction
      November 2014
      558 pages
      ISBN:9781450328852
      DOI:10.1145/2663204

      Copyright © 2014 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 12 November 2014

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      ICMI '14 Paper Acceptance Rate51of127submissions,40%Overall Acceptance Rate453of1,080submissions,42%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader