skip to main content
research-article

Multimodal human attention detection for reading from facial expression, eye gaze, and mouse dynamics

Published:04 November 2016Publication History
Skip Abstract Section

Abstract

Affective computing has recently become an important area in human-computer interaction research. Techniques have been developed to enable computers to understand human affects or emotions, in order to predict human intention more precisely and provide better service to user to enhance user experience. In this paper, we investigate into the detection of human attention level as a useful from of human affect, which could be influential in intelligent e-learning applications. We adopt ubiquitous hardware available in most computer systems, namely, webcam and mouse. Information from multiple input modalities is fused together for effective human attention detection. We invite human subjects to carry out experiments in reading articles when being imposed upon different kinds of distraction to induce them into different levels of attention. Machine-learning techniques are applied to identify useful features to recognize human attention level by building up user-independent models. Our result indicate performance improvement with multimodal inputs from webcam and mouse over that of a single device. We believe that our work has revealed and intersting affective computing direction with potential application in e-learning.

References

  1. Adamczyk, P.D. and Bailey, B.P. If not now, when? Proceedings of Conference on Human fators in Computing Systems - CHI '04, ACM Press (2004), 271--278. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Ambadar, Z., Cohn, J.F., and Reed, L.I. All Smiles are Not Created Equal: Morphology and Timing of Smiles Perceived as Amused, Polite, and Embarrassed/Nervous. Journal of Nonverbal Behavior 33, 1 (2009), 17--34.Google ScholarGoogle ScholarCross RefCross Ref
  3. Anderson, A. R., Christenson, S.L., Sinclair, M.F., and Lehr, C.A. Check & Connect: The importance of relationships for promoting engagement with school. Journal of School Psychology 42, 2 (2004), 95--113.Google ScholarGoogle ScholarCross RefCross Ref
  4. Ashraf, A.B., Lucey, S., Cohn, J.F., et al. The Painful Face - Pain Expression Recognition Using Active Appearance Models. Image and Vision Computing 27, 12(2009), 1788--1796. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Avila, C., Furnham, A., and McClelland, A. The influence of distracting familiar vocal music on cognitive performance of introverts and extraverts. Psychology of Music 40, 1 (2012), 84--93.Google ScholarGoogle ScholarCross RefCross Ref
  6. Bixler, R. and D'Mello, S. Detecting boredom and engagement during writing with keystroke analysis, task appraisals, and stable traits. Proceedings of International Conference on Intelligent User Interfaces - IUI '13, ACM Press (2013), 225--234. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Bolt, R.A. "Put-that-there." Proceedings of Conference on Computer Graphics and Interactive Techniques - SIGGRAPH '80, ACM Press (1980), 262--270. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Calvo, R.A. and D'Mello, S. Affect Detection: An Interdiciplinary Review of Models, Methods, and Their Applications. IEEE Transactions on Appective Computing 1, 1 (2010), 18--37. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Carrino, S., Péclat, A., Mugellini, E., Abou Khaled, O., and Ingold, R. Humans and smart environments. Proceedings of International Conference on Multimodal Interfaces - ICMI '11, ACM Press (2011), 105--112. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Conningham, D.W., Kleiner, M., Bülthoff, H.H., and Wallraven, C. The components of conversational facial expressions. Proceedings of Symposium on Applied perception in Graphics and Visualization - APGV '04, ACM Press (2004), 143--150. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Divjak, M. and Bischof, H. Eye blink based fatigue detection for prevention of computer vision syndrome. Proceedings of IAPR Conference on Machine Vision Applications, (2009), 350--353.Google ScholarGoogle Scholar
  12. Ekman, P. Universals and Cultural Differences in Facial Expression of Emotion. Nebraska Symposium on Motivation, (1972), 207--283.Google ScholarGoogle Scholar
  13. Fox, A.B., Rosen, J., and Crawford, M. Distractions, Distractions: Does Instant Messaging Affect College Students' Performance on a Concurrent Reading Comprehension Task? CyberPsychology & Behavior 12, (2009), 51--53.Google ScholarGoogle ScholarCross RefCross Ref
  14. Fu, Y., Leong, H.V., Ngai, G., Huang, M.X., and Chan, S.C.F. Physiological Mouse: Towards an Emotion-Aware Mouse. Proceedings of International Conference Workshops on Computer Software and Applications, IEEE (2014), 258--263. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Gross, R., Matthews, I., Cohn, J., Kanade, T., and Baker, S. Multi-PIE. Proceedings of International Conference on Automatic Face & Gesture Recognition, IEEE (2008), 1--8.Google ScholarGoogle Scholar
  16. Henderson, J.M. and Hollingworth, A. Eye movements during scene viewing: an overview. In G. Underwood, ed., Eye Guidance in Reading and Scene Perception. Elsevier Science, 1998, 269--293.Google ScholarGoogle Scholar
  17. Huang, M., Ngai, G., Hua, K., Chan, S., and Leong, H.V. Identifying User-Specific Facial Affects from Spontaneous Expressions with Minimal Annotation. IEEE Transactions on Affective Computing PP, 99 (2015), 1--14.Google ScholarGoogle Scholar
  18. Huang, M.X., Kwok, T.C.K., Ngai, G., Leong H.V., and Chan, S.CV.F. Building a Self-Learning Eye Gaze Model from User Interaction Data. Proceedings of International Conference on Multimedia - MM '14, ACM Press (2014), 1017--1020. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Huang, X.M., Li, J., Ngai, G., and Leong, H.V. StressClick: Sensing Stress from Gaze-Click Patterns. To appear in Proceedings of International Conference on Multimedia, (2016). Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Inhoff, A.W. and Rayner, K. Parafoveal word processing during eye fixations in reading: Effects of word frequency. Preception & Psychophysics 40, 6 (1986), 431--439.Google ScholarGoogle Scholar
  21. Ji, Q., Lan, P., and Looney, C.A probabilistic framework for modeling ad real-time monitoring human fatigue. IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans 36, 5 (2006), 862--875. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Liversedge, S.P. and Findlay, J.M. Saccadic eye movements and cognition. Trends in Congnitive Sciences 4, 1 (2000), 6--14.Google ScholarGoogle ScholarCross RefCross Ref
  23. Oviatt, S.L. Multimodal interfaces. In Human-Computer Interaction Handbook: Fundamentals, Evolving Technologies, and Emergine Applications. L. Erlbaum Associates Inc, Hillsdale, NJ, USA, 2007, 286--304. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Picard, R. W. Affective computing. MIT Press, Cambridge, MA, USA, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Radach, R. and Kennedy, A. Theoretical Perspectives on eye movements in reading: Past controversies, current issues, and an agenda for future research. European Journal of Cognitive Psychology 16, 1--2 (2004), 3--26.Google ScholarGoogle ScholarCross RefCross Ref
  26. Rayner, K. Eye Movements in reading and information processing: 20 years of research. Psychological bulletin 124, 3 (1998), 372--422.Google ScholarGoogle Scholar
  27. Reichle, E.D., Reineberg, A.E., and Schooler, J.W. Eye Movements During Mindless Reading. Psychological Science 21, 9 (2010), 1300--1310.Google ScholarGoogle Scholar
  28. Rodrigue, M., Son, J., Giesbrecht, B., Truk, M., and Höllerer, T. Spatio-Temporal Detection of Divided Attention in reading Applications Using EEG and Eye Tracking Proceedings of International Conference on Intelligent User Interfaces - IUI '15, ACM Press (2015), 121--125. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Salvucci, D.D. and Goldberg, J.H. Identifying fixations and saccades in eye-tracking protocols. Proceedings of Symposium on Eye Tracking Research & Applications - ETRA '00, ACM Press (2000), 71--78. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Saragih, J.M., Lucey, S., and Cohn, J.F. Deformable Model Fitting by Regularized Landmark Mean-Shift. International Journal of Computer Vision 91, 2 (2010), 200--215 Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Schleicher, R., Galley, N., Briest, S., and Galley, L. Blinks and saccades as indicators of fatigue in sleepiness warnings: looking tired? Ergonomics 51, 7 (2008), 982--1010.Google ScholarGoogle ScholarCross RefCross Ref
  32. Sen, T. and Megaw, T. The Effects of Task Variables and Prolonged Perfomance on Saccadic Eye Movement Parameters. In A.G. Gale and F. Johnson, eds., Theoretical and Applied Aspects of Eye Movement Research, Elsevier, 1984, 103--111.Google ScholarGoogle Scholar
  33. Talavera, L. An Evaluation of Filter and Wrapper Methods for Feature Selection in Categorical Clustering. In A. Famili, J. Kok, J. Pena, A. Siebes and A. Feelders, eds., Advances in Intelligent Data Analysis VI. Springer Berlin Heidlberg, 2005, 440--451. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Tsalakanidou, F. and Malassiotis, S. Real-time 2D+3D facial action and expression recognition. Pattern Recognition 43, 5 (2010), 1763--1775. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. Tsoulouhas, G., Georgiou, D., and Karkos, A. Detection of Learners' Affective State Based on Mouse Movement. Journal of Computing 3, 11 (2011), 9--18.Google ScholarGoogle Scholar
  36. Vizer, L.M., Zhou, L., and Sears, A. Automated stress detection using keystroke and linguistic features: An exploratory study International Journal of Human-Computer Studies 67, 10 (2009), 870--886. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Weigle, C. and Banks, D.C. Analysis of eye-tracking experiments performed on a Tobii T6. Presented at the Conference on Visualization and Data Analysis, (2008), San José California, USA.Google ScholarGoogle Scholar
  38. Whitehill, J., Serpell, Z., Foster, A., and Movellan, J.R. The Faces of Engagement: Automatic Recognition of Student Engagementform Facial Expressions. IEEE Transactions on Affective Computing 5, 1 (2004), 86--98.Google ScholarGoogle Scholar
  39. Xiong, X. and De la Torre, F. Supervised Descent Method and Its Applications to Face Alignment. Proceedings of Conference on Computer Vision and Pattern Recognition, IEEE (2013), 532--539. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Yamauchi, T. Mouse Trajectories and State Anxiety: Feature Selection with Random Forest. Proceedings of Humaine Association Conference on Affective Computing and Intelligent Interaction, IEEE (2013), 399--404. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Zeng, Z., Hu, Y., Roisman, G.I., Wen, Z., Fu, Y., and Huang, T.S. Audio-Visual Spontaneous Emotion Recognition. Artifical Intelligence for Human Computing 4451, (2007), 72--90. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Zeng, Z., Pantic, M., Roisman, G.I., and Huang, T.S. A Survey of Affect Recognition Methods: Audio, visual, and Spontaneous Expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence 31, 1 (2009), 39--58. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Multimodal human attention detection for reading from facial expression, eye gaze, and mouse dynamics

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader