ABSTRACT
This paper investigates integration, coordination, and transition strategies of gaze and hand input for 3D object manipulation in VR. Specifically, this work aims to understand whether incorporating gaze input can benefit VR object manipulation tasks, and how it should be combined with hand input for improved usability and efficiency. We designed four gaze-supported techniques that leverage different combination strategies for object manipulation and evaluated them in two user studies. Overall, we show that gaze did not offer significant performance benefits for transforming objects in the primary working space, where all objects were located in front of the user and within the arm-reach distance, but can be useful for a larger environment with distant targets. We further offer insights regarding combination strategies of gaze and hand input, and derive implications that can help guide the design of future VR systems that incorporate gaze input for 3D object manipulation.
Supplemental Material
Available for Download
- Ferran Argelaguet and Carlos Andujar. 2013. A survey of 3D object selection techniques for virtual environments. Computers & Graphics 37, 3 (2013), 121–136. https://doi.org/10.1016/j.cag.2012.12.003Google ScholarDigital Library
- Benjamin Bach, Ronell Sicat, Johanna Beyer, Maxime Cordeil, and Hanspeter Pfister. 2017. The hologram in my hand: How effective is interactive exploration of 3d visualizations in immersive tangible augmented reality?IEEE transactions on visualization and computer graphics 24, 1(2017), 457–467. https://doi.org/10.1109/TVCG.2017.2745941Google ScholarCross Ref
- Costas Boletsis. 2017. The new era of virtual reality locomotion: A systematic literature review of techniques and a proposed typology. Multimodal Technologies and Interaction 1, 4 (2017), 24. https://doi.org/10.3390/mti1040024Google ScholarCross Ref
- Gunnar AV Borg. 1982. Psychophysical bases of perceived exertion.Medicine & Science in Sports & Exercise(1982).Google Scholar
- Doug A. Bowman and Larry F. Hodges. 1997. An Evaluation of Techniques for Grabbing and Manipulating Remote Objects in Immersive Virtual Environments. In Proceedings of the 1997 Symposium on Interactive 3D Graphics (Providence, Rhode Island, USA) (I3D ’97). Association for Computing Machinery, New York, NY, USA, 35–ff.https://doi.org/10.1145/253284.253301Google ScholarDigital Library
- Doug A. Bowman, Donald B. Johnson, and Larry F. Hodges. 2001. Testbed Evaluation of Virtual Environment Interaction Techniques. Presence: Teleoperators and Virtual Environments 10, 1(2001), 75–95. https://doi.org/10.1162/105474601750182333Google ScholarDigital Library
- Doug A. Bowman, Ryan P. McMahan, and Eric D. Ragan. 2012. Questioning Naturalism in 3D User Interfaces. Commun. ACM 55, 9 (Sept. 2012), 78–88. https://doi.org/10.1145/2330667.2330687Google ScholarDigital Library
- Wolfgang Büschel, Annett Mitschick, Thomas Meyer, and Raimund Dachselt. 2019. Investigating Smartphone-Based Pan and Zoom in 3D Data Spaces in Augmented Reality. In Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services (Taipei, Taiwan) (MobileHCI ’19). Association for Computing Machinery, New York, NY, USA, Article 2, 13 pages. https://doi.org/10.1145/3338286.3340113Google ScholarDigital Library
- Ishan Chatterjee, Robert Xiao, and Chris Harrison. 2015. Gaze+Gesture: Expressive, Precise and Targeted Free-Space Interactions. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (Seattle, Washington, USA) (ICMI ’15). Association for Computing Machinery, New York, NY, USA, 131–138. https://doi.org/10.1145/2818346.2820752Google ScholarDigital Library
- Shujie Deng, Nan Jiang, Jian Chang, Shihui Guo, and Jian J Zhang. 2017. Understanding the impact of multimodal interaction using gaze informed mid-air gesture control in 3D virtual objects manipulation. International Journal of Human-Computer Studies 105 (2017), 68–80. https://doi.org/10.1016/j.ijhcs.2017.04.002Google ScholarCross Ref
- Joachim Diepstraten, Daniel Weiskopf, and Thomas Ertl. 2002. Transparency in interactive technical illustrations. In Computer Graphics Forum, Vol. 21. Wiley Online Library, 317–325. https://doi.org/10.1111/1467-8659.t01-1-00591Google ScholarCross Ref
- Niklas Elmqvist and Philippas Tsigas. 2007. A taxonomy of 3D occlusion management techniques. In 2007 IEEE Virtual Reality Conference. IEEE, 51–58. https://doi.org/10.1109/VR.2007.352463Google ScholarCross Ref
- Barrett M. Ens, Rory Finnegan, and Pourang P. Irani. 2014. The Personal Cockpit: A Spatial Interface for Effective Task Switching on Head-Worn Displays. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Toronto, Ontario, Canada) (CHI ’14). Association for Computing Machinery, New York, NY, USA, 3171–3180. https://doi.org/10.1145/2556288.2557058Google ScholarDigital Library
- Facebook. [n.d.]. Oculus Medium. Retrieved September 6, 2020 from https://www.oculus.com/medium/Google Scholar
- Anna Maria Feit, Shane Williams, Arturo Toledo, Ann Paradiso, Harish Kulkarni, Shaun Kane, and Meredith Ringel Morris. 2017. Toward Everyday Gaze Input: Accuracy and Precision of Eye Tracking and Implications for Design. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). Association for Computing Machinery, New York, NY, USA, 1118–1130. https://doi.org/10.1145/3025453.3025599Google ScholarDigital Library
- Tiare Feuchtner and Jörg Müller. 2018. Ownershift: Facilitating Overhead Interaction in Virtual Reality with an Ownership-Preserving Hand Space Shift. In Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology (Berlin, Germany) (UIST ’18). Association for Computing Machinery, New York, NY, USA, 31–43. https://doi.org/10.1145/3242587.3242594Google ScholarDigital Library
- S. Frees and G. D. Kessler. 2005. Precise and rapid interaction through scaled manipulation in immersive virtual environments. In IEEE Proceedings. VR 2005. Virtual Reality, 2005.99–106. https://doi.org/10.1109/VR.2005.1492759Google ScholarCross Ref
- Epic Games. [n.d.]. Unreal Engine VR Mode. Retrieved September 6, 2020 from https://docs.unrealengine.com/en-US/Engine/Editor/VR/index.html/Google Scholar
- Christoph Gebhardt, Brian Hecox, Bas van Opheusden, Daniel Wigdor, James Hillis, Otmar Hilliges, and Hrvoje Benko. 2019. Learning Cooperative Personalized Policies from Gaze Data. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology (New Orleans, LA, USA) (UIST ’19). Association for Computing Machinery, New York, NY, USA, 197–208. https://doi.org/10.1145/3332165.3347933Google ScholarDigital Library
- Google. [n.d.]. Introducing Blocks. Retrieved September 6, 2020 from https://arvr.google.com/blocks/Google Scholar
- Sandra G Hart. 2006. NASA-task load index (NASA-TLX); 20 years later. In Proceedings of the human factors and ergonomics society annual meeting, Vol. 50. Sage publications Sage CA: Los Angeles, CA, 904–908. https://doi.org/10.1177/154193120605000909Google ScholarCross Ref
- Devamardeep Hayatpur, Seongkook Heo, Haijun Xia, Wolfgang Stuerzlinger, and Daniel Wigdor. 2019. Plane, Ray, and Point: Enabling Precise Spatial Manipulations with Shape Constraints. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology (New Orleans, LA, USA) (UIST ’19). Association for Computing Machinery, New York, NY, USA, 1185–1195. https://doi.org/10.1145/3332165.3347916Google ScholarDigital Library
- Juan David Hincapié-Ramos, Xiang Guo, Paymahn Moghadasian, and Pourang Irani. 2014. Consumed Endurance: A Metric to Quantify Arm Fatigue of Mid-Air Interactions. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Toronto, Ontario, Canada) (CHI ’14). Association for Computing Machinery, New York, NY, USA, 1063–1072. https://doi.org/10.1145/2556288.2557130Google ScholarDigital Library
- Teresa Hirzle, Jan Gugenheimer, Florian Geiselhart, Andreas Bulling, and Enrico Rukzio. 2019. A Design Space for Gaze Interaction on Head-Mounted Displays. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). Association for Computing Machinery, New York, NY, USA, Article 625, 12 pages. https://doi.org/10.1145/3290605.3300855Google ScholarDigital Library
- Mozilla Hubs. [n.d.]. Hubs - Private social VR in your web browser. Retrieved Augest 19, 2020 from https://hubs.mozilla.com/Google Scholar
- Aulikki Hyrskykari, Howell Istance, and Stephen Vickers. 2012. Gaze Gestures or Dwell-Based Interaction?. In Proceedings of the Symposium on Eye Tracking Research and Applications (Santa Barbara, California) (ETRA ’12). Association for Computing Machinery, New York, NY, USA, 229–232. https://doi.org/10.1145/2168556.2168602Google ScholarDigital Library
- VRChat Inc.[n.d.]. VRChat. Retrieved September 6, 2020 from https://www.vrchat.com/Google Scholar
- Robert J. K. Jacob. 1990. What You Look at is What You Get: Eye Movement-Based Interaction Techniques. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Seattle, Washington, USA) (CHI ’90). Association for Computing Machinery, New York, NY, USA, 11–18. https://doi.org/10.1145/97243.97246Google ScholarDigital Library
- Sujin Jang, Wolfgang Stuerzlinger, Satyajit Ambike, and Karthik Ramani. 2017. Modeling Cumulative Arm Fatigue in Mid-Air Interaction Based on Perceived Exertion and Kinetics of Arm Motion. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). Association for Computing Machinery, New York, NY, USA, 3328–3339. https://doi.org/10.1145/3025453.3025523Google ScholarDigital Library
- Mikko Kytö, Barrett Ens, Thammathip Piumsomboon, Gun A. Lee, and Mark Billinghurst. 2018. Pinpointing: Precise Head- and Eye-Based Target Selection for Augmented Reality. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada) (CHI ’18). Association for Computing Machinery, New York, NY, USA, Article 81, 14 pages. https://doi.org/10.1145/3173574.3173655Google ScholarDigital Library
- Joseph J LaViola Jr, Ernst Kruijff, Ryan P McMahan, Doug Bowman, and Ivan P Poupyrev. 2017. 3D user interfaces: theory and practice. Addison-Wesley Professional.Google Scholar
- Johnson Ching-Hong Li. 2016. Effect size measures in a two-independent-samples case with nonnormal and nonhomogeneous data. Behavior research methods 48, 4 (2016), 1560–1574. https://doi.org/10.3758/s13428-015-0667-zGoogle ScholarCross Ref
- Zhenxing Li, Deepak Akkil, and Roope Raisamo. 2019. Gaze Augmented Hand-Based Kinesthetic Interaction: What You See is What You Feel. IEEE transactions on haptics 12, 2 (2019), 114–127. https://doi.org/10.1109/TOH.2019.2896027Google ScholarCross Ref
- Chang Liu, Alexander Plopski, and Jason Orlosky. 2020. OrthoGaze: Gaze-based Three-dimensional Object Manipulation using Orthogonal Planes. Computers & Graphics(2020). https://doi.org/10.1016/j.cag.2020.04.005Google ScholarCross Ref
- Mingyu Liu, Mathieu Nancel, and Daniel Vogel. 2015. Gunslinger: Subtle Arms-down Mid-Air Interaction. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology (Charlotte, NC, USA) (UIST ’15). Association for Computing Machinery, New York, NY, USA, 63–71. https://doi.org/10.1145/2807442.2807489Google ScholarDigital Library
- D. Mendes, F. M. Caputo, A. Giachetti, A. Ferreira, and J. Jorge. 2019. A Survey on 3D Virtual Object Manipulation: From the Desktop to Immersive Virtual Environments. Computer Graphics Forum 38, 1 (2019), 21–45. https://doi.org/10.1111/cgf.13390Google ScholarCross Ref
- Daniel Mendes, Filipe Relvas, Alfredo Ferreira, and Joaquim Jorge. 2016. The Benefits of DOF Separation in Mid-Air 3D Object Manipulation. In Proceedings of the 22nd ACM Conference on Virtual Reality Software and Technology (Munich, Germany) (VRST ’16). Association for Computing Machinery, New York, NY, USA, 261–268. https://doi.org/10.1145/2993369.2993396Google ScholarDigital Library
- Minecraft. [n.d.]. Minecraft Official Site. Retrieved Augest 19, 2020 from https://www.minecraft.net/Google Scholar
- Ken Pfeuffer, Jason Alexander, Ming Ki Chong, and Hans Gellersen. 2014. Gaze-Touch: Combining Gaze with Multi-Touch for Interaction on the Same Surface. In Proceedings of the 27th Annual ACM Symposium on User Interface Software and Technology(Honolulu, Hawaii, USA) (UIST ’14). Association for Computing Machinery, New York, NY, USA, 509–518. https://doi.org/10.1145/2642918.2647397Google ScholarDigital Library
- Ken Pfeuffer, Benedikt Mayer, Diako Mardanbegi, and Hans Gellersen. 2017. Gaze + Pinch Interaction in Virtual Reality. In Proceedings of the 5th Symposium on Spatial User Interaction (Brighton, United Kingdom) (SUI ’17). Association for Computing Machinery, New York, NY, USA, 99–108. https://doi.org/10.1145/3131277.3132180Google ScholarDigital Library
- Jeffrey S. Pierce, Brian C. Stearns, and Randy Pausch. 1999. Voodoo Dolls: Seamless Interaction at Multiple Scales in Virtual Environments. In Proceedings of the 1999 Symposium on Interactive 3D Graphics (Atlanta, Georgia, USA) (I3D ’99). Association for Computing Machinery, New York, NY, USA, 141–145. https://doi.org/10.1145/300523.300540Google ScholarDigital Library
- Matti Pouke, Antti Karhu, Seamus Hickey, and Leena Arhippainen. 2012. Gaze Tracking and Non-Touch Gesture Based Interaction Method for Mobile 3D Virtual Spaces. In Proceedings of the 24th Australian Computer-Human Interaction Conference (Melbourne, Australia) (OzCHI ’12). Association for Computing Machinery, New York, NY, USA, 505–512. https://doi.org/10.1145/2414536.2414614Google ScholarDigital Library
- Ivan Poupyrev, Mark Billinghurst, Suzanne Weghorst, and Tadao Ichikawa. 1996. The Go-Go Interaction Technique: Non-Linear Mapping for Direct Manipulation in VR. In Proceedings of the 9th Annual ACM Symposium on User Interface Software and Technology(Seattle, Washington, USA) (UIST ’96). Association for Computing Machinery, New York, NY, USA, 79–80. https://doi.org/10.1145/237091.237102Google ScholarDigital Library
- Warren Robinett and Richard Holloway. 1992. Implementation of Flying, Scaling and Grabbing in Virtual Worlds. In Proceedings of the 1992 Symposium on Interactive 3D Graphics (Cambridge, Massachusetts, USA) (I3D ’92). Association for Computing Machinery, New York, NY, USA, 189–192. https://doi.org/10.1145/147156.147201Google ScholarDigital Library
- Kunhee Ryu, Joong-Jae Lee, and Jung-Min Park. 2019. GG Interaction: a gaze–grasp pose interaction for 3D virtual object selection. Journal on Multimodal User Interfaces 13, 4 (2019), 383–393. https://doi.org/10.1007/s12193-019-00305-yGoogle ScholarCross Ref
- Jeff Sauro and Joseph S. Dumas. 2009. Comparison of Three One-Question, Post-Task Usability Questionnaires. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Boston, MA, USA) (CHI ’09). Association for Computing Machinery, New York, NY, USA, 1599–1608. https://doi.org/10.1145/1518701.1518946Google ScholarDigital Library
- Martin Schrepp, Andreas Hinderks, and Jörg Thomaschewski. 2017. Design and Evaluation of a Short Version of the User Experience Questionnaire (UEQ-S).IJIMAI 4, 6 (2017), 103–108. https://doi.org/10.9781/ijimai.2017.09.001Google ScholarCross Ref
- Marcos Serrano, Barrett Ens, Xing-Dong Yang, and Pourang Irani. 2015. Gluey: Developing a Head-Worn Display Interface to Unify the Interaction Experience in Distributed Display Environments. In Proceedings of the 17th International Conference on Human-Computer Interaction with Mobile Devices and Services (Copenhagen, Denmark) (MobileHCI ’15). Association for Computing Machinery, New York, NY, USA, 161–171. https://doi.org/10.1145/2785830.2785838Google ScholarDigital Library
- Ludwig Sidenmark and Hans Gellersen. 2019. Eye, Head and Torso Coordination During Gaze Shifts in Virtual Reality. ACM Trans. Comput.-Hum. Interact. 27, 1, Article 4 (Dec. 2019), 40 pages. https://doi.org/10.1145/3361218Google ScholarDigital Library
- Ludwig Sidenmark and Hans Gellersen. 2019. Eye&Head: Synergetic Eye and Head Movement for Gaze Pointing and Selection. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology (New Orleans, LA, USA) (UIST ’19). Association for Computing Machinery, New York, NY, USA, 1161–1174. https://doi.org/10.1145/3332165.3347921Google ScholarDigital Library
- Ludwig Sidenmark, Diako Mardanbegi, Argenis Ramirez Gomez, Christopher Clarke, and Hans Gellersen. 2020. BimodalGaze: Seamlessly Refined Pointing with Gaze and Filtered Gestural Head Movement. In ACM Symposium on Eye Tracking Research and Applications (Stuttgart, Germany) (ETRA ’20 Full Papers). Association for Computing Machinery, New York, NY, USA, Article 8, 9 pages. https://doi.org/10.1145/3379155.3391312Google ScholarDigital Library
- Adalberto L. Simeone, Andreas Bulling, Jason Alexander, and Hans Gellersen. 2016. Three-Point Interaction: Combining Bi-Manual Direct Touch with Gaze. In Proceedings of the International Working Conference on Advanced Visual Interfaces (Bari, Italy) (AVI ’16). Association for Computing Machinery, New York, NY, USA, 168–175. https://doi.org/10.1145/2909132.2909251Google ScholarDigital Library
- Vincent Sitzmann, Ana Serrano, Amy Pavel, Maneesh Agrawala, Diego Gutierrez, Belen Masia, and Gordon Wetzstein. 2018. Saliency in VR: How do people explore virtual environments?IEEE transactions on visualization and computer graphics 24, 4(2018), 1633–1642. https://doi.org/10.1109/TVCG.2018.2793599Google ScholarDigital Library
- Dana Slambekova, Reynold Bailey, and Joe Geigel. 2012. Gaze and Gesture Based Object Manipulation in Virtual Worlds. In Proceedings of the 18th ACM Symposium on Virtual Reality Software and Technology (Toronto, Ontario, Canada) (VRST ’12). Association for Computing Machinery, New York, NY, USA, 203–204. https://doi.org/10.1145/2407336.2407380Google ScholarDigital Library
- Sophie Stellmach and Raimund Dachselt. 2013. Still Looking: Investigating Seamless Gaze-Supported Selection, Positioning, and Manipulation of Distant Targets. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Paris, France) (CHI ’13). Association for Computing Machinery, New York, NY, USA, 285–294. https://doi.org/10.1145/2470654.2470695Google ScholarDigital Library
- Sophie Stellmach, Sebastian Stober, Andreas Nürnberger, and Raimund Dachselt. 2011. Designing Gaze-Supported Multimodal Interactions for the Exploration of Large Image Collections. In Proceedings of the 1st Conference on Novel Gaze-Controlled Applications (Karlskrona, Sweden) (NGCA ’11). Association for Computing Machinery, New York, NY, USA, Article 1, 8 pages. https://doi.org/10.1145/1983302.1983303Google ScholarDigital Library
- Richard Stoakley, Matthew J. Conway, and Randy Pausch. 1995. Virtual Reality on a WIM: Interactive Worlds in Miniature. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’95). ACM Press/Addison-Wesley Publishing Co., USA, 265–272. https://doi.org/10.1145/223904.223938Google ScholarDigital Library
- Amedeo Troiano, Francesco Naddeo, Erik Sosso, Gianfranco Camarota, Roberto Merletti, and Luca Mesin. 2008. Assessment of force and fatigue in isometric contractions of the upper trapezius muscle by surface EMG signal and perceived exertion scale. Gait & Posture 28, 2 (2008), 179 – 186. https://doi.org/10.1016/j.gaitpost.2008.04.002Google ScholarCross Ref
- Jayson Turner, Jason Alexander, Andreas Bulling, and Hans Gellersen. 2015. Gaze+RST: Integrating Gaze and Multitouch for Remote Rotate-Scale-Translate Tasks. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (Seoul, Republic of Korea) (CHI ’15). Association for Computing Machinery, New York, NY, USA, 4179–4188. https://doi.org/10.1145/2702123.2702355Google ScholarDigital Library
- Jayson Turner, Jason Alexander, Andreas Bulling, Dominik Schmidt, and Hans Gellersen. 2013. Eye pull, eye push: Moving objects between large screens and personal devices with gaze and touch. In IFIP Conference on Human-Computer Interaction. Springer, 170–186. https://doi.org/10.1007/978-3-642-40480-1_11Google ScholarCross Ref
- Jayson Turner, Andreas Bulling, Jason Alexander, and Hans Gellersen. 2013. Eye Drop: An Interaction Concept for Gaze-Supported Point-to-Point Content Transfer. In Proceedings of the 12th International Conference on Mobile and Ubiquitous Multimedia(Luleå, Sweden) (MUM ’13). Association for Computing Machinery, New York, NY, USA, Article 37, 4 pages. https://doi.org/10.1145/2541831.2541868Google ScholarDigital Library
- Jayson Turner, Andreas Bulling, Jason Alexander, and Hans Gellersen. 2014. Cross-Device Gaze-Supported Point-to-Point Content Transfer. In Proceedings of the Symposium on Eye Tracking Research and Applications (Safety Harbor, Florida) (ETRA ’14). Association for Computing Machinery, New York, NY, USA, 19–26. https://doi.org/10.1145/2578153.2578155Google ScholarDigital Library
- Jayson Turner, Andreas Bulling, and Hans Gellersen. 2011. Combining Gaze with Manual Interaction to Extend Physical Reach. In Proceedings of the 1st International Workshop on Pervasive Eye Tracking & Mobile Eye-Based Interaction (Beijing, China) (PETMEI ’11). Association for Computing Machinery, New York, NY, USA, 33–36. https://doi.org/10.1145/2029956.2029966Google ScholarDigital Library
- András Vargha and Harold D Delaney. 2000. A critique and improvement of the CL common language effect size statistics of McGraw and Wong. Journal of Educational and Behavioral Statistics 25, 2(2000), 101–132. https://doi.org/10.3102/10769986025002101Google ScholarCross Ref
- Manuel Veit, Antonio Capobianco, and Dominique Bechmann. 2009. Influence of Degrees of Freedom’s Manipulation on Performances during Orientation Tasks in Virtual Reality Environments. In Proceedings of the 16th ACM Symposium on Virtual Reality Software and Technology (Kyoto, Japan) (VRST ’09). Association for Computing Machinery, New York, NY, USA, 51–58. https://doi.org/10.1145/1643928.1643942Google ScholarDigital Library
- Eduardo Velloso, Jayson Turner, Jason Alexander, Andreas Bulling, and Hans Gellersen. 2015. An Empirical Investigation of Gaze Selection in Mid-Air Gestural 3D Manipulation. In Human-Computer Interaction – INTERACT 2015. Springer International Publishing, Cham, 315–330. https://doi.org/10.1007/978-3-319-22668-2_25Google ScholarDigital Library
- Simon Voelker, Sebastian Hueber, Christian Holz, Christian Remy, and Nicolai Marquardt. 2020. GazeConduits: Calibration-Free Cross-Device Collaboration through Gaze and Touch. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI ’20). Association for Computing Machinery, New York, NY, USA, 1–10. https://doi.org/10.1145/3313831.3376578Google ScholarDigital Library
- Tobii VR. [n.d.]. Hand-Eye Coordination. Retrieved Augest 12, 2020 from https://vr.tobii.com/sdk/develop/unity/unity-examples/hand-eye-coordination/Google Scholar
- Johann Wentzel, Greg d’Eon, and Daniel Vogel. 2020. Improving Virtual Reality Ergonomics Through Reach-Bounded Non-Linear Input Amplification. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI ’20). Association for Computing Machinery, New York, NY, USA, 1–12. https://doi.org/10.1145/3313831.3376687Google ScholarDigital Library
- Curtis Wilkes and Doug A. Bowman. 2008. Advantages of Velocity-Based Scaling for Distant 3D Manipulation. In Proceedings of the 2008 ACM Symposium on Virtual Reality Software and Technology (Bordeaux, France) (VRST ’08). Association for Computing Machinery, New York, NY, USA, 23–29. https://doi.org/10.1145/1450579.1450585Google ScholarDigital Library
- Jacob O. Wobbrock, Leah Findlater, Darren Gergle, and James J. Higgins. 2011. The Aligned Rank Transform for Nonparametric Factorial Analyses Using Only Anova Procedures. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Vancouver, BC, Canada) (CHI ’11). Association for Computing Machinery, New York, NY, USA, 143–146. https://doi.org/10.1145/1978942.1978963Google ScholarDigital Library
- Yukang Yan, Chun Yu, Xiaojuan Ma, Shuai Huang, Hasan Iqbal, and Yuanchun Shi. 2018. Eyes-Free Target Acquisition in Interaction Space around the Body for Virtual Reality. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada) (CHI ’18). Association for Computing Machinery, New York, NY, USA, 1–13. https://doi.org/10.1145/3173574.3173616Google ScholarDigital Library
- Difeng Yu, Hai-Ning Liang, Kaixuan Fan, Heng Zhang, Charles Fleming, and Konstantinos Papangelis. 2020. Design and Evaluation of Visualization Techniques of Off-Screen and Occluded Targets in Virtual Reality Environments. IEEE Transactions on Visualization and Computer Graphics 26, 9(2020), 2762–2774. https://doi.org/10.1109/TVCG.2019.2905580Google ScholarCross Ref
- Difeng Yu, Qiushi Zhou, Joshua Newn, Tilman Dingler, Eduardo Velloso, and Jorge Goncalves. 2020. Fully-Occluded Target Selection in Virtual Reality. IEEE Transactions on Visualization and Computer Graphics 26, 12(2020), 3402–3413. https://doi.org/10.1109/TVCG.2020.3023606Google ScholarCross Ref
- Difeng Yu, Qiushi Zhou, Benjamin Tag, Tilman Dingler, Eduardo Velloso, and Jorge Goncalves. 2020. Engaging Participants during Selection Studies in Virtual Reality. In 2020 IEEE Conference on Virtual Reality and 3D User Interfaces (VR). IEEE, 500–509. https://doi.org/10.1109/VR46266.2020.00071Google ScholarCross Ref
- Shumin Zhai, Carlos Morimoto, and Steven Ihde. 1999. Manual and Gaze Input Cascaded (MAGIC) Pointing. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Pittsburgh, Pennsylvania, USA) (CHI ’99). Association for Computing Machinery, New York, NY, USA, 246–253. https://doi.org/10.1145/302979.303053Google ScholarDigital Library
Index Terms
- Gaze-Supported 3D Object Manipulation in Virtual Reality
Recommendations
Gaze + pinch interaction in virtual reality
SUI '17: Proceedings of the 5th Symposium on Spatial User InteractionVirtual reality affords experimentation with human abilities beyond what's possible in the real world, toward novel senses of interaction. In many interactions, the eyes naturally point at objects of interest while the hands skilfully manipulate in 3D ...
Thumble: One-Handed 3D Object Manipulation Using a Thimble-Shaped Wearable Device in Virtual Reality
UIST '22 Adjunct: Adjunct Proceedings of the 35th Annual ACM Symposium on User Interface Software and TechnologyConventional controllers or hand-tracking interactions in VR cause hand fatigue while manipulating 3D objects because repetitive wrist rotation and hand movements are often required. As a solution to this inconvenience, we propose Thumble, a novel ...
Glasses with haptic feedback of gaze gestures
CHI EA '14: CHI '14 Extended Abstracts on Human Factors in Computing SystemsWe introduce eyeglasses that present haptic feedback when using gaze gestures for input. The glasses utilize vibrotactile actuators to provide gentle stimulation to three locations on the user's head. We describe two initial user studies that were ...
Comments