ABSTRACT
The design of remote gesturing technologies is an area of growing interest. Current technologies have taken differing approaches to the representation of remote gesture. It is not clear which approach has the most benefit to task performance. This study therefore compared performance in a collaborative physical (assembly) task using remote gesture systems constructed with combinations of three different gesture formats (unmediated hands only, hands and sketch and digital sketch only) and two different gesture output locations (direct projection into a worker's task space or on an external monitor). Results indicated that gesturing with an unmediated representation of the hands leads to faster performance with no loss of accuracy. Comparison of gesture output locations did not find a significant difference between projecting gestures and presenting them on external monitors. These results are discussed in relation to theories of conversational grounding and the design of technologies from a 'mixed ecologies' perspective.
- Argyle, M. Bodily Communication. (1988). Routledge.Google Scholar
- Bekker, M. M., Olson, J. S., & Olson, G. M. Analysis of gestures in face-to-face design teams provides guidance for how to use groupware in design. In Proceedings of DIS 95. ACM Press. (1995). 157--166. Google ScholarDigital Library
- Chapanis, A., Ochsman, R. B., Parrish, R. N., &Weeks, G. D. Studies in interactive communication: I. The effects of four communication modes on the behavior of teams during cooperative problem-solving. Human Factors, 14(6), (1972). 487--509.Google Scholar
- Clark, H. H. Using language. (1996). Cambridge: Cambridge University Press.Google Scholar
- Clark, H. H., & Krych, M. A. Speaking while monitoring addressees for understanding. Journal of Memory & Language, 50(1), (2004). 62--81.Google Scholar
- Clark, H. H., & Wilkes-Gibbs, D. Referring as a collaborative process. Cognition, 22(1), (1986). 1--39.Google Scholar
- Clark, H. H., & Brennan, S.E. Grounding in Communication. In L.B. Resnick, R.M. Levine, & S.D. Teasley (Eds.). Perspectives on socially shared cognition, (1991). 127--149. Washington, DC: APA.Google Scholar
- Fraser, M. Working with Objects in Collaborative Environments. (2001) Unpublished PhD thesis, University of Nottingham.Google Scholar
- Fussell, S. R., Setlock, L. D., Yang, J., Ou, J., Mauer, E. M., & Kramer, A. Gestures over video streams to support remote collaboration on physical tasks. Human-Computer Interaction, 19, (2004). 273--309. Google ScholarDigital Library
- Fussell, S.R., Kraut, R.E., & Siegel, J. Coordination of communication: Effects of shared visual context on collaborative work. Proceedings of CSCW 2000, ACM Press (2000). 21--30. Google ScholarDigital Library
- Fussell, S.R., Setlock, L.D., & Parker, E.M. Where do helpers look? Gaze targets during collaborative physical tasks. In Proceedings of CHI 2003 (Extended Abstracts), ACM Press (2003).768--769. Google ScholarDigital Library
- Gergle, D., Kraut, R.E., & Fussell, S.R. Action as language in a shared visual space. Proceedings of CSCW 2004. ACM Press (2004). 487--496. Google ScholarDigital Library
- Heath, C. C. and Luff P. K. Convergent activities: collaborative work and multimedia technology in London Underground Line Control Rooms, in D. Middleton and Y. Engestrom (eds.), Cognition and Communication at Work: Distributed Cognition in the Workplace. Cambridge University Press (1996). 96--130.Google Scholar
- Karsenty, L. Cooperative work and shared visual context: An empirical study of comprehension problems and in side-by-side and remote help dialogues. Human-Computer Interaction, 14(3), (1999). 283--315. Google ScholarDigital Library
- Kirk, D. S. & Stanton Fraser, D. The Effects of Remote Gesturing on Distance Instruction. In Proceedings of CSCL 2005. LEA (2005). Google ScholarDigital Library
- Kirk, D., Crabtree, A. & Rodden, T. Ways of the Hands. In Proceedings of ECSCW 2005. Kluwer (2005). Google ScholarDigital Library
- Kraut, R. E., Fussell, S. R., & Siegel, J. Visual information as a conversational resource in collaborative physical tasks. Human Computer Interaction, 18(1), (2003). 13-49. Google ScholarDigital Library
- Kraut, R. E., Gergle, D., & Fussell, S. R. The use of visual information in shared visual spaces: Informing the development of virtual co-presence. In Proceedings of CSCW 2002, ACM Press. (2002). 31--40. Google ScholarDigital Library
- Kraut, R. E., Miller, M. D., & Siegel, J. Collaboration in performance of physical tasks: Effects on outcomes and communication. Proceedings of CSCW 1996 ACM Press (1996). 57--66. Google ScholarDigital Library
- Kuzuoka, H. Spatial workspace collaboration: A Sharedview video support system for remote collaboration capability. Proceedings of CHI'92 ACM Press (1992). 533--540. Google ScholarDigital Library
- Kuzuoka, H., Kosaka, J., Yamazaki, K., Suga, Y., Yamazaki, A., Luff, P., & Heath, C. Mediating Dual Ecologies. Proceedings of CSCW 2004 ACM Press (2004). 477--486). Google ScholarDigital Library
- Kuzuoka, H., Kosuge, T., & Tanaka, K. GestureCam: A video communication system for sympathetic remote collaboration. Proceedings of CSCW 1994 ACM Press (1994) 35--43. Google ScholarDigital Library
- Kuzuoka, H., Oyama, S., Yamazaki, K., Suzuki, K., & Mitsuishi, M. GestureMan: A mobile robot that embodies a remote instructor's actions. Proceedings of CSCW 2000 ACM Press (2000). 155--162. Google ScholarDigital Library
- Luff, P., Heath, C., Kuzuoka, H., Hindmarsh, J., Yamazaki, K., and Oyama, S., Fractured ecologies: creating environments for collaboration, Special Issue of the HCI Journal: 'Talking About Things: Mediated Conversation about Objects', 18, 1 & 2, (2003). 51--84. Google ScholarDigital Library
- McNeill, D. Hand and Mind. What gestures reveal about thought. (1996) Chicago: University of Chicago Press.Google Scholar
- Ou, J., Fussell, S. R., Chen, X., Setlock, L. D., & Yang, J. Gestural communication over video stream: Supporting multimodal interaction for remote collaborative physical tasks. In Proceedings of ICMI 2003, (2003) Google ScholarDigital Library
- Ou, J., Min, L., Yang, J. and Fussell, S. R. Effects of Task Properties, Partner Actions and Message Content on Eye Gaze Patterns in a Collaborative Task. In Proceedings of CHI 2005. ACM Press (2005). 231 -- 240. Google ScholarDigital Library
- Rauscher, F. H., Krauss, R. M. & Chen, Y. (1996) Gesture, Speech and Lexical Access: The Role of Lexical Movements in Speech Production. Psychological Science. 7 4 226--231.Google ScholarCross Ref
- Sakata, N., Kurata, T., Kato, T., Kourogi, M. and Kuzuoka, H. WACL: Supporting Telecommunications Using Wearable Active Camera with Laser Pointer. Proceedings of Int. Symp. on Wearable Computers 2003. IEEE Press (2003). Google ScholarDigital Library
- Tang, A., Boyle, M. and Greenberg, S. Display and Presence Disparity in Mixed Presence Groupware. Proceedings of Australasian user Interface. ACM Press (2004). 73--82. Google ScholarDigital Library
- Tang, J. C. Findings from observational studies of collaborative work. International Journal of Man-Machine Studies, 34, (1991). 143--160. Google ScholarDigital Library
Index Terms
- Comparing remote gesture technologies for supporting collaborative physical tasks
Recommendations
Turn it this way: grounding collaborative action with remote gestures
CHI '07: Proceedings of the SIGCHI Conference on Human Factors in Computing SystemsRemote gesture systems have been shown to provide a significant enhancement to performance in collaborative physical tasks, an effect ascribed to the ability of remote gestures to help ground deictic references. The argument that this effect works by ...
Ownership and control of point of view in remote assistance
CHI '13: Proceedings of the SIGCHI Conference on Human Factors in Computing SystemsIn this paper we investigate user performance and user behavior, related to the issue of who controls the point of view in a remote assistance scenario. We describe an experiment that examined users completing two different tasks with the aid of a ...
Effects of Enhanced Gaze Presentation on Gaze Leading in Remote Collaborative Physical Tasks
CHI '18: Proceedings of the 2018 CHI Conference on Human Factors in Computing SystemsWith respect to collaborative physical tasks, gaze and gestures play significant roles when referring to physical objects. In video-mediated communication, however, such nonverbal cues become "ineffectual" when they are presented via a 2D monitor, ...
Comments