ABSTRACT
Eye-gaze activity provides rich information about individuals' engagement in social interactions. Gaze is one of the strongest visual cues in face-to-face interaction. Previous studies have examined how eye gaze can be used to coordinate social interactions such as turn taking and identifying the focus of attention. In this study, we investigate the role of gaze during collaborative problem-solving tasks, specifically how individuals perform different gazing activities when holding different team roles in pair programming, and also whether the differences in eye gaze (if any) provide predictive insight into learning outcomes. We analyzed 40 students' eye-gaze activities, which were annotated for each second during a collaborative problem-solving task (~50 min on average). The results show that students' roles in the collaborative task have a significant relationship with eye-gaze activities. Moreover, participants' gaze activities can provide predictive insight into their post-test scores. These findings suggest that simple activity measures such as relative frequency of eye-gaze activities can be very useful in understanding the collaborative process.
- Peter Auer. 2017. Gaze, addressee selection and turn-taking in three-party interaction. Journal of InLiSt - Interaction and Linguistic Structures 60.Google Scholar
- Kent Beck. 2000. Extreme programming eXplained: embrace change. Addison-Wesley. Google ScholarDigital Library
- Sallyann Bryant and Benedict du Boulay. 2008. Pair programming and the mysterious role of the navigator. International Journal of Human-Computer Studies 66, 7: 519--529. Google ScholarDigital Library
- Teresa Busjahn, Roman Bednarik, Andrew Begel, Martha Crosby, James H. Paterson, Carsten Schulte, Bonita Sharif, and Sascha Tamm. 2015. Eye Movements in Code Reading: Relaxing the Linear Order. In 2015 IEEE 23rd International Conference on Program Comprehension, 255--265. Google ScholarDigital Library
- Mehmet Celepkolu and Kristy Elizabeth Boyer. 2018. Thematic Analysis of Students' Reflections on Pair Programming in CS1. In Proceedings of the 49th ACM Technical Symposium on Computer Science Education - SIGCSE '18, 771--776. Google ScholarDigital Library
- Jan Chong and Tom Hurlbutt. 2007. The Social Dynamics of Pair Programming. In 29th International Conference on Software Engineering (ICSE07), 354--363. Google ScholarDigital Library
- Martha E. Crosby and Jan Stelovsky. 1990. How do we read algorithms? A case study. Computer 23, 1: 25--35. Google ScholarDigital Library
- Sarah D'Angelo and Andrew Begel. 2017. Improving Communication Between Pair Programmers Using Shared Gaze Awareness. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems - CHI '17, 6245--6290. Google ScholarDigital Library
- Gwyneth Doherty-Sneddon and Fiona G. Phelps. 2005. Gaze aversion: A response to cognitive or social difficulty? Memory & Cognition 33, 4: 727--733.Google ScholarCross Ref
- Starkey Duncan and Donald W. Fiske. 2015. Face-to-Face Interaction. Routledge.Google Scholar
- Randi A. Engle, Jennifer M. Langer-Osuna, and Maxine McKinney de Royston. 2014. Toward a Model of Influence in Persuasive Discussions: Negotiating Quality, Authority, Privilege, and Access Within a Student-Led Argument. Journal of the Learning Sciences 23, 2: 245--268.Google ScholarCross Ref
- Margaret Foddy. 1978. Patterns of Gaze in Cooperative and Competitive Negotiation. Human Relations 31, 11: 925--938.Google ScholarCross Ref
- Tom Foulsham, Esther Walker, and Alan Kingstone. 2011. The where, what and when of gaze allocation in the lab and the natural environment. Vision Research 51, 17: 1920--1931.Google ScholarCross Ref
- Megan Freeth, Tom Foulsham, and Alan Kingstone. 2013. What Affects Social Attention? Social Presence, Eye Contact and Autistic Traits. PLoS ONE 8(1): e53286.Google ScholarCross Ref
- Andrew C Gallup, Joseph J Hale, David J T Sumpter, Simon Garnier, Alex Kacelnik, John R Krebs, and Iain D Couzin. 2012. Visual attention and the acquisition of information in human crowds. In Proceedings of the National Academy of Sciences, 109, 19: 7245--7250.Google ScholarCross Ref
- Shuchi Grover, Marie Bienkowski, Amir Tamrakar, Behjat Siddiquie, David Salter, and Ajay Divakaran. 2016. Multimodal analytics to study collaborative problem solving in pair programming. In Proceedings of the Sixth International Conference on Learning Analytics & Knowledge - LAK '16, 516--517. Google ScholarDigital Library
- Simon Ho, Tom Foulsham, and Alan Kingstone. 2015. Speaking and Listening with the Eyes: Gaze Signaling during Dyadic Interactions. PLoS ONE 10(8): e0136905.Google ScholarCross Ref
- Danielle L. Jones and Scott D. Fleming. 2013. What use is a backseat driver? A qualitative investigation of pair programming. In 2013 IEEE Symposium on Visual Languages and Human Centric Computing, 103--110.Google Scholar
- Adam Kendon. 1967. Some functions of gaze-direction in social interaction. Acta Psychologica 26: 22--63.Google ScholarCross Ref
- Sonu Chopra Khullar and Norman I. Badler. 2001. Where to Look? Automating Attending Behaviors of Virtual Human Characters. Autonomous Agents and Multi-Agent Systems 4, 1/2: 9--23. Google ScholarDigital Library
- Chris L. Kleinke. 1986. Gaze and Eye Contact: A Research Review. Psychological Bulletin 100, 1: 78--100.Google ScholarCross Ref
- Meng-Lung Lai, Meng-Jung Tsai, Fang-Ying Yang, Chung-Yuan Hsu, Tzu-Chien Liu, Silvia Wen-Yu Lee, Min-Hsien Lee, Guo-Li Chiou, Jyh-Chong Liang, and Chin-Chung Tsai. 2013. A review of using eye-tracking technology in exploring learning from 2000 to 2012. Educational Research Review 10: 90--115.Google ScholarCross Ref
- Kaitlin E. W Laidlaw, Tom Foulsham, Gustav Kuhn, and Alan Kingstone. 2011. Potential social interactions are important to social attention. Proceedings of the National Academy of Sciences 108, 14: 5548--53.Google ScholarCross Ref
- Colleen M. Lewis and Niral Shah. 2015. How Equity and Inequity Can Emerge in Pair Programming. In Proceedings of the Eleventh annual International Conference on International Computing Education Research - ICER '15, 41--50. Google ScholarDigital Library
- Charlie McDowell, Linda Werner, Heather Bullock, and Julian Fernald. 2002. The effects of pair-programming on performance in an introductory programming course. ACM SIGCSE Bulletin 34, 1: 38. Google ScholarDigital Library
- Nachiappan Nagappan, Laurie Williams, Miriam Ferzli, Eric Wiebe, Kai Yang, Carol Miller, and Suzanne Balik. 2003. Improving the CS1 experience with pair programming. ACM SIGCSE Bulletin 35, 1: 359. Google ScholarDigital Library
- Markus Nivala, Florian Hauser, Jurgen Mottok, and Hans Gruber. 2016. Developing visual expertise in software engineering: An eye tracking study. In 2016 IEEE Global Engineering Education Conference (EDUCON), 613--620.Google ScholarCross Ref
- Marc-Antoine Nüssli. 2011. Dual Eye-Tracking Methods for the Study of Remote Collaborative Problem Solving. PhD thesis, Ecole Polytechnique Federale de Lausanne.Google Scholar
- Unaizah Obaidellah, Mohammed Al Haek, and Peter C. H. Cheng. 2018. A Survey on the Usage of Eye-Tracking in Computer Programming. ACM Computing Surveys 51, 1: 1--58. Google ScholarDigital Library
- Laura Plonka, Helen Sharp, and Janet Van Der Linden. 2012. Disengagement in Pair Programming: Does It Matter? In Proceedings of the 34th International Conference on Software Engineering, 496--506. Google ScholarDigital Library
- Evan F. Risko, Kaitlin Laidlaw, Megan Freeth, Tom Foulsham, and Alan Kingstone. 2012. Social attention with real versus reel stimuli: toward an empirical approach to concerns about ecological validity. Frontiers in Human Neuroscience 6: 143.Google ScholarCross Ref
- Kshitij Sharma, Patrick Jermann, Marc-Antoine Nüssli, and Pierre Dillenbourg. 2012. Gaze Evidence for Different Activities in Program Understanding. In Proceedings of 24th Annual conference of Psychology of Programming Interest Group.Google Scholar
- Tim J. Smith and Parag K. Mital. 2013. Attentional synchrony and the influence of viewing task on gaze behavior in static and dynamic scenes. Journal of Vision 13, 8: 16--16.Google ScholarCross Ref
- Randy Stein and Susan E. Brennan. 2004. Another person's eye gaze as a cue in solving programming problems. In Proceedings of the Sixth International Conference on Multimodal interfaces - ICMI '04, 9. Google ScholarDigital Library
- Melissa L. H. Vo, Tim J. Smith, Parag K. Mital, and John M. Henderson. 2012. Do the eyes really have it? Dynamic allocation of attention when viewing moving faces. Journal of Vision 12, 13: 3--3.Google ScholarCross Ref
- Jiahui Wang, Pavlo Antonenko, Mehmet Celepkolu, Yerika Jimenez, Ethan Fieldman, and Ashley Fieldman. 2018. Exploring Relationships Between Eye Tracking and Traditional Usability Testing Data. International Journal of Human--Computer Interaction: 1--12.Google Scholar
- Armin Weinberger, Karsten Stegmann, and Frank Fischer. 2010. Learning to argue online: Scripted groups surpass individuals (unscripted groups do not). Computers in Human Behavior 26, 4: 506--515. Google ScholarDigital Library
- Laurie Williams, Eric Wiebe, Kai Yang, Miriam Ferzli, and Carol Miller. 2002. In Support of Pair Programming in the Introductory Computer Science Course. Computer Science Education 12, 3: 197--212.Google ScholarCross Ref
Index Terms
- Predicting Student Performance Based on Eye Gaze During Collaborative Problem Solving
Recommendations
Eye to Eye: Gaze Patterns Predict Remote Collaborative Problem Solving Behaviors in Triads
Artificial Intelligence in EducationAbstractWe investigated the feasibility of using eye gaze to model collaborative problem solving (CPS) behaviors in 96 triads (N = 288) who used videoconferencing to remotely collaborate on an educational physics game. Trained human raters coded spoken ...
Non-intrusive eye gaze estimation without knowledge of eye pose
FGR' 04: Proceedings of the Sixth IEEE international conference on Automatic face and gesture recognitionEye gaze estimation is to detect the eye gaze of human. By detection of the eye gaze, a machine can read the intention. Eye gaze is very useful to read human intention. In this paper, a new eye gaze estimation method is suggested. The method allows ...
Gaze Speedup: Eye Gaze Assisted Gesture Typing in Virtual Reality
IUI '23: Proceedings of the 28th International Conference on Intelligent User InterfacesMid-air text input in augmented or virtual reality (AR/VR) is an open problem. One proposed solution is gesture typing where the user performs a gesture trace over the keyboard. However, this requires the user to move their hands precisely and ...
Comments