Abstract
Computer graphics hardware supporting real-time interactive 3D animation has the potential to support effective user interfaces by enabling virtual 3D workspaces. However, this potential requires development of viewpoint movement techniques that support rapid and controlled movement through workspaces. Rapid movement through large distances avoids wasted work time; controlled movement near target objects allows the user to examine and interact with objects in the workspace. Current techniques for viewpoint movement typically use high velocities to cover distances rapidly, but high velocities are hard to control near objects. This paper describes a new technique for targeted viewpoint movement that solves this problem. The key idea is to have the user indicate a point of interest (target) on a 3D object and use the distance to this target to move the viewpoint logarithmically, by moving the same relative percentage of distance to the target on every animation cycle. the result is rapid motion over distances that slows as the viewpoint approaches the target object. The technique can be used with 2D and multidimensional input devices. We also extend the technique to move objects in the workspace.
- 1 Badler, Norman I., Kamran H. Manoochehri, and David Baraff. Multi-Dimensional input techniques and articulated figure positioning by multiple constraints. in Proceedings o/ the 1986 Workshop on Interactive 319 Graphics (Chapel Hill, NC, October 1986). ACM, New York~ 1987~ 151-169. Google ScholarDigital Library
- 2 Bier, Eric A. Skitters and jacks: interactive 3D positioning tools. In Proceedings of the 1986 Workshop on Interactive 3D Graphics (Chapel Hill, NC, October 1986). ACM, New York, 1987, 183-196. Google ScholarDigital Library
- 3 Bier, Eric A. Snap-dragging in three dimensions. Proceedings of the 1990 Symposium on Interactive 3D Graphics (Snowbird, Utah, March, 1990). In Computer Graphics 24, 2 (March 1990), 193-204. Google ScholarDigital Library
- 4 Brooks, Frederick P. Jr. A dynamic graphics system for simulating virtual buildings. In Proceedings of the 1986 Workshop on Interactive 31) Graphics (Chapel Hill, NC, October 1986). ACM, New York, 1987, 9-21. Google ScholarDigital Library
- 5 Chen, Michael, S. Joy Mountford, and Abigail Sellen. A study in interactive 3-D rotation using 2-D control devices. Proceedings of SIGGRAPH'88 (Atlanta, Georgla~ August 1988). In Computer Graphics 25, 4 (August 1988), 121-129. Google ScholarDigital Library
- 6 Evans, Kenneth B., Peter P. Tanner, and Marceli Wein. Tablet-based valuators that provide one, two, or three degrees of freedom. Proceedings of SIGGRAPH'81 (Dallas, Texas, August 1981). In Computer Graphics 15, 3 (August 1981), 91-97. Google ScholarDigital Library
- 7 Fairchild, Kim M., Steven E. Poltrock, and George W. Furnas. Semnet: three-dimensional graphic representations of large knowledge bases. In Cognitive science and its applications for human-computer interaction, Guindon, R. (ed), Lawrence Erlbaum, 1988.Google Scholar
- 8 Mackinlay, Jock D., Stuart K. Card, and George G. Robertson. A semantic analysis of the design space of input devices. Human-Computer Interaction, to appear in vol. 5, 1990.Google Scholar
- 9 Nielson, Gregory M., and Dan R. Olsen Jr. Direct manipulation techniques for 3D objects using 2D locator devices. In Proceedings of the 1986 Workshop on Interactive 3D Graphics, (Chapel Hill, NC, October 1986). ACM, New York, 175-182. Google ScholarDigital Library
- 10 Pew, R. W. Human perceptual-motor performance. In Human Information Processing: Tutorials in Performance and Cognition, Kantowitz, B. H. (ed), Lawrence Erlbaum, 1974, 1-40.Google Scholar
- 11 Phillips, Cary B., and Norman I. Badler. Jack: a toolkit for manipulating articulated figures. In Proceedings of the A CM SIGGRAPH Symposium on User Interface Software, (Banff, Canada, October 1988). ACM, New York, 221-229. Google ScholarDigital Library
- 12 Robertson, George G., Stuart K. Card, and Jock D. Mackinlay. The cognitive coprocessor architecture for interactive user interfaces. In Proceedings of the A CM SIGGRAPH Symposium on User Interface Software and Technology, (Williamsburg, Virginia, November 1989). ACM, New York, 10-18. Google ScholarDigital Library
- 13 Sturman, David 3., David Zeltzer, and Steve Pieper. Hands-on interaction with virtual enviornments. In Proceedings of the A CM SIGGRAPH Symposium on User Interface So~ware and Technology, (Williamsburg, Virginia, November 1989). ACM, New York, 19- 24. Google ScholarDigital Library
- 14 Ware, Colin and Steve Osborne. Exploration and virtual camera control in virtual three dimensional environments. Proceedings of the 1990 Symposium on Interactive 3D Graphics (Snowbird, Utah, March 1990). In Computer Graphics 24, 2 (March 1990), 175-183. Google ScholarDigital Library
- 15 Weimer, David, and S. K. Ganapathy. A synthetic visual environment with hand gesturing and voice input. In Proceedings of CHI'89, (Austin, Texas, April 1989). ACM, New York, 235-240. Google ScholarDigital Library
Index Terms
- Rapid controlled movement through a virtual 3D workspace
Recommendations
Rapid controlled movement through a virtual 3D workspace
SIGGRAPH '90: Proceedings of the 17th annual conference on Computer graphics and interactive techniquesComputer graphics hardware supporting real-time interactive 3D animation has the potential to support effective user interfaces by enabling virtual 3D workspaces. However, this potential requires development of viewpoint movement techniques that support ...
Object-wise 3D Gaze Mapping in Physical Workspace
AH '18: Proceedings of the 9th Augmented Human International ConferenceUnderstanding the intention of other people is a fundamental social skill in human communication. Eye behavior is an important, yet implicit communication cue. In this work, we focus on enabling people to see the users' gaze associated with objects in ...
DEEP: 3D Gaze Pointing in Virtual Reality Leveraging Eyelid Movement
UIST '22: Proceedings of the 35th Annual ACM Symposium on User Interface Software and TechnologyGaze-based target suffers from low input precision and target occlusion. In this paper, we explored to leverage the continuous eyelid movement to support high-efficient and occlusion-robust dwell-based gaze pointing in virtual reality. We first ...
Comments