ABSTRACT
We present BeThere, a proof-of-concept system designed to explore 3D input for mobile collaborative interactions. With BeThere, we explore 3D gestures and spatial input which allow remote users to perform a variety of virtual interactions in a local user's physical environment. Our system is completely self-contained and uses depth sensors to track the location of a user's fingers as well as to capture the 3D shape of objects in front of the sensor. We illustrate the unique capabilities of our system through a series of interactions that allow users to control and manipulate 3D virtual content. We also provide qualitative feedback from a preliminary user study which confirmed that users can complete a shared collaborative task using our system.
- Bauer, M., and Kortuem, G. Where Are You Pointing At? A Study of Remote Collaboration in a Wearable Videoconference System. In Wearable Computers. (1999), 151--161. Google ScholarDigital Library
- Benko, H., Jota, R., Wilson, A. D., and Inesc-id, V. MirageTable: Freehand Interaction on a Projected Augmented Reality Tabletop. In Proc. of CHI (2012), 199--208. Google ScholarDigital Library
- Bérard, F., Ip, J., Benovoy, M., El-Shimy, D., Blum, J., and Cooperstock, J. Did minority report get it wrong? superiority of the mouse over 3d input devices in a 3d placement task. In Proc. of INTERACT (2009), 400--414. Google ScholarDigital Library
- Billinghurst, M., Cheok, A., Kato, H., and Prince, S. Real World Teleconferencing. In Proc. of CHI (1999), 194--195. Google ScholarDigital Library
- Butler, A., and Izadi, S. SideSight: multi-touch interaction around small devices. In Proc. of UIST (2008), 3--6. Google ScholarDigital Library
- Ens, B., Ahlström, D., Cockburn, A., and Irani, P. Characterizing user performance with assisted direct off-screen pointing. In Proc. of MobileHCI (2011), 485--494. Google ScholarDigital Library
- Fitzmaurice, G. W., Zhai, S., and Chignell, M. H. Virtual reality for palmtop computers. In Trans. on Info. Sys., vol. 11 (1993), 197--218. Google ScholarDigital Library
- Friedrich, D. W. ARVIKA: Augmented Reality for Development, Production and Service. In Proc. of ISMAR (2002), 1--2. Google ScholarDigital Library
- Fussell, S., Setlock, L., Yang, J., Ou, J., and Mauer, E. Gestures Over Video Streams to Support Remote Collaboration on Physical Tasks. In Hum.-Comput. Interact. 19 (Sept. 2004), 273--309. Google ScholarDigital Library
- Gauglitz, S., Lee, C., Turk, M., and Hollerer, T. Integrating the Physical Environment into Mobile Remote Collaboration. In Proc. of Mobile HCI (2012), 241--250. Google ScholarDigital Library
- Gergle, D., and Clark, A. T. See What I m Saying? Using Dyadic Mobile Eye Tracking to Study Collaborative Reference. In Proc. CSCW (2011), 435--444. Google ScholarDigital Library
- Gurevich, P., Lanir, J., Cohen, B., and Stone, R. TeleAdvisor: a versatile augmented reality tool for remote assistance. In Proc. of CHI (2012), 619--622. Google ScholarDigital Library
- Gustafson, S., and Bierwirth, D. Imaginary interfaces: spatial interaction with empty hands and without visual feedback. In Proc. of UIST (2010), 3--12. Google ScholarDigital Library
- Gutwin, C., and Penner, R. Improving interpretation of remote gestures with telepointer traces. Proc. of CSCW (2002), 49. Google ScholarDigital Library
- Hackenberg, G. Lightweight Palm and Finger Tracking for Real-Time 3D Gesture Control. Building, March 2010 (2011), 19--26. Google ScholarDigital Library
- Hancock, M., ten Cate, T., and Carpendale, S. Sticky tools: full 6dof force-based interaction for multi-touch tables. In Proc. of ITS, ITS '09 (2009), 133--140. Google ScholarDigital Library
- Harrison, C., Benko, H., Wilson, A. D., and Way, O. M. OmniTouch: Wearable multitouch interaction everywhere. In Proc. UIST (2011), 441--450. Google ScholarDigital Library
- Harrison, C., and Hudson, S. Abracadabra: wireless, high-precision, and unpowered finger input for very small mobile devices. In Proc. UIST (2009), 121--124. Google ScholarDigital Library
- Hilliges, O., Kim, D., Izadi, S., Weiss, M., and Wilson, A. Holodesk: direct 3d interactions with a situated see-through display. In Proc. of CHI (2012), 2421--2430. Google ScholarDigital Library
- Izadi, S. KinectFusion: Real-time 3D Reconstruction and Interaction Using a Moving Depth Camera. In Proc. of UIST (2011), 559--568. Google ScholarDigital Library
- Jones, B., Sodhi, R., Bailey, B., Maciocci, G., and Forsyth, D. Depth-based Around Device Interaction for Multiscale Navigation. In Proc. MobileHCI (2012), 83--92. Google ScholarDigital Library
- Kratz, S., and Rohs, M. HoverFlow: Expanding the Design Space of Around-Device Interaction. In MobileHCI (2009), 1--8. Google ScholarDigital Library
- Kratz, S., and Rohs, M. PalmSpace: Continuous Around-Device Gestures vs. Multitouch for 3D Rotation Tasks on Mobile Devices Categories and Subject Descriptors Interaction on Mobile Devices. In Proc. of AVI (2012), 181--188. Google ScholarDigital Library
- Li, W., Agrawala, M., Curless, B., and Salesin, D. Automated Generation of Interactive 3D Exploded View Diagrams. In Proc. of SIGGRAPH (2008), 101--107. Google ScholarDigital Library
- Lincoln, P., Welch, G., Nashel, A., Ilie, A., State, A., and Fuchs, H. Animatronic Shader Lamps Avatars. In Proc. of ISMAR, Ieee (Oct. 2009), 27--33. Google ScholarDigital Library
- Maimone, A. A First Look at a Telepresence System with Room-Sized Real-Time 3D Capture. In Proc. ICAT (2011), 4--9.Google Scholar
- Martinet, A., Casiez, G., and Grisoni, L. Integrality and separability of multitouch interaction techniques in 3d manipulation tasks. Trans. on Vis. and CG 18, 3 (Mar. 2012), 369--380. Google ScholarDigital Library
- Nancel, M., Wagner, J., Pietriga, E., Chapuis, O., Mackay, W., Orsay, F., and Orsay, F. Mid-air Pan-and-Zoom on Wall-sized Displays. In Proc of CHI, no. May (2011), 177--186. Google ScholarDigital Library
- Neill, J. O., Castellani, S., Roulland, F., and Hairon, N. From Ethnographic Study to Mixed Reality: A Remote Collaborative Troubleshooting System. Proc. of CSCW (2011), 225--234. Google ScholarDigital Library
- Oikonomidis, I., Kyriazis, N., and Argyros, A. Efficient model-based 3D tracking of hand articulations using Kinect. Proc of the BMVC (2011), 101.1--101.11.Google ScholarCross Ref
- Ranjan, A., Birnholtz, J. P., and Balakrishnan, R. Dynamic Shared Visual Spaces: Experimenting with Automatic Camera Control in a Remote Repair Task. In Proc. of CHI (2007), 1177--1186. Google ScholarDigital Library
- Raskar, R., Welch, G., Cutts, M., Lake, A., Stesin, L., and Fuchs, H. The office of the future. In Proc. SIGGRAPH (1998), 179--188. Google ScholarDigital Library
- Reisman, J. L., Davidson, P. L., and Han, J. Y. A screen-space formulation for 2d and 3d direct manipulation. In Proc. of UIST, UIST'09 (2009), 69--78. Google ScholarDigital Library
- Ring, M.-t. F., Ashbrook, D., Baudisch, P., White, S., Str, P.-d.-h., and Potsdam, D. Nenya: Subtle and Eyes-Free Mobile Input with a. In Proc. CHI (2011), 2043--2046. Google ScholarDigital Library
- Sakong, K. Supporting telepresence by visual and physical cues in distributed 3D collaborative design environments. In Proc. CHI (2006), 1283--1288. Google ScholarDigital Library
- Stafford, A. Piekarski, W., and Thomas, B. Implementation of god-like interaction techniques. In Proc. of ISMAR (2006) (2006), 165--172. Google ScholarDigital Library
- Tang, J., and Minneman, S. VideoDraw: A Video Interface for Collaborative Drawing. In ACM Trans. Inf. Syst. 9 (1991), 170--184. Google ScholarDigital Library
- Wigdor, D., Forlines, C., Baudisch, P., Barnwell, J., and Shen, C. Lucid touch: a see-through mobile device. In Proc. of UIST, ACM (2007), 269--278. Google ScholarDigital Library
- Wilson, A., and Robbins, D. Playtogether: Playing games across multiple interactive tabletops. In Proc. of IUI (2006).Google Scholar
- Wilson, A. D. Robust Computer Vision-Based Detection of Pinching for One and Two-Handed Gesture Input. In Proc of UIST (2006), 255--258. Google ScholarDigital Library
Index Terms
- BeThere: 3D mobile collaboration with spatial input
Recommendations
Air+touch: interweaving touch & in-air gestures
UIST '14: Proceedings of the 27th annual ACM symposium on User interface software and technologyWe present Air+Touch, a new class of interactions that interweave touch events with in-air gestures, offering a unified input modality with expressiveness greater than each input modality alone. We demonstrate how air and touch are highly complementary: ...
The Eye in Extended Reality: A Survey on Gaze Interaction and Eye Tracking in Head-worn Extended Reality
With innovations in the field of gaze and eye tracking, a new concentration of research in the area of gaze-tracked systems and user interfaces has formed in the field of Extended Reality (XR). Eye trackers are being used to explore novel forms of spatial ...
Extending interaction for smart watches: enabling bimanual around device control
CHI EA '14: CHI '14 Extended Abstracts on Human Factors in Computing SystemsThe size of a smart watch limits the available interactive surface for the user. Most current smart watches use a combination of a touch screen and physical buttons. Unfortunately, a small touch screen's usability is limited when it can be easily ...
Comments