ABSTRACT
Connected vehicles collect and share data by communicating with road infrastructure, with each other, the web, IoT systems, and with their occupants’ personal devices. Part of this data is presented to drivers via a multitude of interactive devices and systems. Thus, one challenge that arises in such a complex environment is effective and safe operation of the various interactive systems, e.g., the in-vehicle infotainment (IVI). In this paper, we present a synopsis of input modalities from the literature of automotive user interfaces (AutoUIs) for media consumption inside connected vehicles.
- B. I. Ahmad, C. Hare, H. Singh, A. Shabani, B. Lindsay, L. Skrypchuk, P. Langdon, and S. Godsill. 2018. Selection Facilitation Schemes for Predictive Touch with Mid-Air Pointing Gestures in Automotive Displays. In AutomotiveUI ’18. ACM.Google Scholar
- B. I. Ahmad, P. M. Langdon, S. J. Godsill, R. Hardy, E. Dias, and L. Skrypchuk. 2014. Interactive Displays in Vehicles: Improving Usability with a Pointing Gesture Tracker and Bayesian Intent Predictors. In AutomotiveUI ’14. ACM, 1–8.Google Scholar
- B. I. Ahmad, J. Murphy, P. M. Langdon, and S. J. Godsill. 2014. Filtering perturbed in-vehicle pointing gesture trajectories: Improving the reliability of intent inference. In WMLSP ’14. 1–6. https://doi.org/10.1109/MLSP.2014.6958860Google Scholar
- B. I. Ahmad, J. K. Murphy, P. M. Langdon, S. J. Godsill, R. Hardy, and L. Skrypchuk. 2016. Intent Inference for Hand Pointing Gesture-Based Interactions in Vehicles. IEEE T. Cybernetics 46, 4 (2016). https://doi.org/10.1109/TCYB.2015.2417053Google ScholarCross Ref
- A. Aiordăchioae, R.-D. Vatavu, and D.-M. Popovici. 2019. A Design Space for Vehicular Lifelogging to Support Creation of Digital Content in Connected Cars. In EICS ’19. ACM. https://doi.org/10.1145/3319499.3328234Google ScholarDigital Library
- M. Alpern and K. Minardo. 2003. Developing a Car Gesture Interface for Use as a Secondary Task. In CHI EA ’03. ACM, New York, NY, USA, 932–933. https://doi.org/10.1145/765891.766078Google ScholarDigital Library
- I. Alvarez, A. Martin Hammond, J. Dunbar, J. Taiber, D.-M. Wilson, and J. E. Gilbert. 2010. Voice Interfaced Vehicle User Help. In AutomotiveUI ’10. ACM, 42–49. https://doi.org/10.1145/1969773.1969782Google ScholarDigital Library
- L. Angelini, J. Baumgartner, F. Carrino, S. Carrino, M. Caon, O. A. Khaled, J. Sauer, D. Lalanne, E. Mugellini, and A. Sonderegger. 2016. A Comparison of Three Interaction Modalities in the Car: Gestures, Voice and Touch. In IHM ’16.Google Scholar
- L. Angelini, F. Carrino, S. Carrino, M. Caon, O.A. Khaled, J. Baumgartner, A. Sonderegger, D. Lalanne, and E. Mugellini. 2014. Gesturing on the Steering Wheel: A User-Elicited Taxonomy.Google Scholar
- L. Angelini, F. Carrino, S. Carrino, M. Caon, D. Lalanne, O. A. Khaled, and E. Mugellini. 2013. Opportunistic Synergy: A Classifier Fusion Engine for Micro-Gesture Recognition. In AutomotiveUI ’13. 8. https://doi.org/10.1145/2516540.2516563Google ScholarDigital Library
- L. Arce-Plevnik, A. Cohen, and T. Shor. 2016. IoT in Automotive Industry: Connecting Cars. Nanotechnology and Nanosensors (04 2016).Google Scholar
- L.-B. Bilius. 2020. A Smartwatch-based User Interface for In-Vehicle Interactions. In DAS ’20.Google Scholar
- I. Bourmeyster, J. Chetoni, S. Cucchi, N. Griggio, A. Guido, G. Moroni, A. Riccio, M. Stanzani, and F. Valente. 1996. Voice controlled mobile phone for car environment. In ESP ’96. 1–4. https://ieeexplore.ieee.org/document/7082870Google Scholar
- D. Brand, K. Büchele, and A. Meschtscherjakov. 2016. Pointing at the HUD: Gesture Interaction Using a Leap Motion. In AutomotiveUI ’16 Adjunct. ACM, New York, NY, USA, 167–172. https://doi.org/10.1145/3004323.3004343Google ScholarDigital Library
- G. Burnett, E. Crundall, D. Large, G. Lawson, and L. Skrypchuk. 2013. A Study of Unidirectional Swipe Gestures on In-Vehicle Touch Screens. In AutomotiveUI ’13. ACM, 22–29. https://doi.org/10.1145/2516540.2516545Google ScholarDigital Library
- C.-J. Chung and W.-Y. Chen. 2001. Low complexity voice recognition in car environment. In SPAWC ’01. 380–383. https://doi.org/10.1109/SPAWC.2001.923931Google Scholar
- N. Deo, A. Rangesh, and M. Trivedi. 2016. In-vehicle Hand Gesture Recognition using Hidden Markov models. In ITSC ’16. 2179–2184. https://doi.org/10.1109/ITSC.2016.7795908Google ScholarDigital Library
- H. Detjen, S. Faltaous, S. Geisler, and S. Schneegass. 2019. User-Defined Voice and Mid-Air Gesture Commands for Maneuver-Based Interventions in Automated Vehicles. In MuC ’19. 8. https://doi.org/10.1145/3340764.3340798Google ScholarDigital Library
- R. Ecker, V. Broy, A. Butz, and A. De Luca. 2009. PieTouch: A Direct Touch Gesture Interface for Interacting with in-Vehicle Information Systems. In MobileHCI ’09. ACM. https://doi.org/10.1145/1613858.1613887Google ScholarDigital Library
- R. Ecker, V. Broy, K. Hertzschuch, and A. Butz. 2010. Visual Cues Supporting Direct Touch Gesture Interaction with In-Vehicle Information Systems. In AutomotiveUI ’10. ACM, 80–87. https://doi.org/10.1145/1969773.1969788Google ScholarDigital Library
- M. FakhrHosseini, J. Sterkenburg, S. Landry, J. Ryan, and M. Jeon. 2017. Design Process of Sonically-Enhanced Air Gesture Controls in Vehicles. In AutomotiveUI ’17. https://doi.org/10.1145/3131726.3131874Google ScholarDigital Library
- H. J. Fariman, H. J. Alyamani, M. Kavakli, and L. Hamey. 2016. Designing a User-Defined Gesture Vocabulary for an in-Vehicle Climate Control System. In OzCHI ’16. ACM, 391–395. https://doi.org/10.1145/3010915.3010955Google ScholarDigital Library
- F. Faubel, M. Georges, K. Kumatani, A. Bruhn, and D. Klakow. 2011. Improving hands-free speech recognition in a car through audio-visual voice activity detection. In 2011 Joint Work. on Hands-free Speech Comm. and Microphone Arrays.Google Scholar
- L. Fissore, P. Laface, and P. Ruscitti. 1992. HMM modeling for speaker independent voice dialing in car environment. In ASSP ’92. 249–252 vol.1. https://doi.org/10.1109/ICASSP.1992.225925Google Scholar
- T. M. Gable, K. R. May, and B. N. Walker. 2014. Applying Popular Usability Heuristics to Gesture Interaction in the Vehicle. In Adjunct AutomotiveUI ’14. ACM, New York, NY, USA, 1–7. https://doi.org/10.1145/2667239.2667298Google ScholarDigital Library
- B.-F. Gheran and R.-D. Vatavu. 2020. From Controls on the Steering Wheel to Controls on the Finger: Using Smart Rings for In-Vehicle Interactions. In DIS ’20 Companion. https://doi.org/10.1145/3393914.3395851Google ScholarDigital Library
- M. H. Hassan, W. S. Beardslee, and T. K. Arnold. 1993. Voice recognition digital cipher lock for smart vehicles. In MSCS ’93. 352–355 vol.1. https://doi.org/10.1109/MWSCAS.1993.343037Google ScholarCross Ref
- J. F. Hessam, M. Zancanaro, M. Kavakli, and M. Billinghurst. 2017. Towards Optimization of Mid-Air Gestures for in-Vehicle Interactions. In OZCHI ’17. ACM, NY, USA, 126–134. https://doi.org/10.1145/3152771.3152785Google ScholarDigital Library
- J. Hook, P. Stenton, M. F. Ursu, G. Schofield, and R.-D. Vatavu. 2019. Welcome From the ACM TVX 2019 Chairs. In TVX ’19. ACM, New York, NY, USA. https://dl.acm.org/doi/proceedings/10.1145/3317697Google Scholar
- R. Häuslschmid, B. Menrad, and A. Butz. 2015. Freehand vs. micro gestures in the car: Driving performance and user experience. In 3DUI ’15. 159–160. https://doi.org/10.1109/3DUI.2015.7131749Google Scholar
- S. Khan, H. Akmal, I. Ali, and N. Naeem. 2017. Efficient and unique learning of in-car voice control for engineering education. In INMIC ’17. 1–6. https://doi.org/10.1109/INMIC.2017.8289467Google Scholar
- J. Kim, M. Jeong, and S. C. Lee. 2019. Why Did This Voice Agent Not Understand Me?”: Error Recovery Strategy for in-Vehicle Voice User Interface. In AutomotiveUI ’19. ACM, 146–150. https://doi.org/10.1145/3349263.3351513Google ScholarDigital Library
- K. Kim and J. Choi. 2019. Passengers’ Gesture Recognition Model in Self-driving Vehicles : Gesture Recognition Model of the Passengers’ Obstruction of the Vision of the Driver. In ICCCS ’19. 239–242. https://doi.org/10.1109/CCOMS.2019.8821649Google Scholar
- T. Kujala and H. Grahn. 2017. Visual Distraction Effects of In-Car Text Entry Methods: Comparing Keyboard, Handwriting and Voice Recognition. In AutomotiveUI ’17. ACM, 1–10. https://doi.org/10.1145/3122986.3122987Google ScholarDigital Library
- S. H. Lee, S.-O. Yoon, and J. H. Shin. 2015. On-Wheel Finger Gesture Control for in-Vehicle Systems on Central Consoles. In AutomotiveUI ’15 Adjunct. ACM, New York, NY, USA, 94–99. https://doi.org/10.1145/2809730.2809739Google ScholarDigital Library
- P. Lei, M. Chen, and J. Wang. 2019. Speech enhancement for in-vehicle voice control systems using wavelet analysis and blind source separation. IET Intelligent Transport Systems 13, 4 (2019), 693–702. https://doi.org/10.1049/iet-its.2018.5094Google ScholarCross Ref
- H. Lin. 2019. Using Passenger Elicitation for Developing Gesture Design Guidelines for Adjusting Highly Automated Vehicle Dynamics. In DIS 2019 Companion. ACM, New York, NY, USA, 97–100. https://doi.org/10.1145/3301019.3324878Google ScholarDigital Library
- J. Ma and Y. Du. 2017. Study on the evaluation method of in-vehicle gesture control. In ICCSSE ’17. 145–148. https://doi.org/10.1109/CCSSE.2017.8087912Google Scholar
- J. Mahetalia, R. Mehta, S. Varudandi, and K. Samdani. 2018. A Comparative Study on Gesture Recognition Systems for Car Infotainment Systems. In SCAN ’18. 1–7. https://doi.org/10.1109/ICSCAN.2018.8541236Google Scholar
- A. Mahr, C. Endres, C. Müller, and T. Schneeberger. 2011. Determining Human-Centered Parameters of Ergonomic Micro-Gesture Interaction for Drivers Using the Theater Approach. In AutomotiveUI ’11.Google Scholar
- Y. Manabe and T. Imamura. 2018. Construction of a Voice Direction Database for the Destination Input Interface of Autonomous Vehicles. In SMC ’18. 813–818. https://doi.org/10.1109/SMC.2018.00146Google ScholarDigital Library
- U. E. Manawadu, M. Kamezaki, M. Ishikawa, T. Kawano, and S. Sugano. 2016. A hand gesture based driver-vehicle interface to control lateral and longitudinal motions of an autonomous vehicle. In SMC ’16. https://doi.org/10.1109/SMC.2016.7844497Google ScholarDigital Library
- K. May, T. M. Gable, X. Wu, R. R. Sardesai, and B. N. Walker. 2016. Choosing the Right Air Gesture: Impacts of Menu Length and Air Gesture Type on Driver Workload. In AutomotiveUI ’16 Adj. 6. https://doi.org/10.1145/3004323.3004330Google ScholarDigital Library
- K. R. May, T. M. Gable, and B. N. Walker. 2014. A Multimodal Air Gesture Interface for In Vehicle Menu Navigation. In Adjunct AutomotiveUI ’14. ACM, New York, NY, USA, 1–6. https://doi.org/10.1145/2667239.2667280Google ScholarDigital Library
- K. R. May, T. M. Gable, and B. N. Walker. 2017. Designing an In-Vehicle Air Gesture Set Using Elicitation Methods. In AutomotiveUI ’17. ACM, New York, NY, USA, 74–83. https://doi.org/10.1145/3122986.3123015Google ScholarDigital Library
- R. Neßelrath, M. M. Moniri, and M. Feld. 2016. Combining Speech, Gaze, and Micro-gestures for the Multimodal Control of In-Car Functions. In IE ’16. 190–193. https://doi.org/10.1109/IE.2016.42Google Scholar
- E. Ohn-Bar, C. Tran, and M. Trivedi. 2012. Hand Gesture-Based Visual User Interface for Infotainment. In AutomotiveUI ’12. ACM, New York, NY, USA, 111–115. https://doi.org/10.1145/2390256.2390274Google ScholarDigital Library
- P. Paolini, P. Cremonesi, G. Lekakos, and M. Obrist. 2013. Chairs’ Welcome. In EuroITV ’13. ACM, New York, NY, USA.Google Scholar
- F. Parada-Loira, E. González-Agulla, and J. L. Alba-Castro. 2014. Hand gestures to control infotainment equipment in cars. In IVS ’14. 1–6. https://doi.org/10.1109/IVS.2014.6856614Google Scholar
- B. Pfleging, S. Schneegass, and A. Schmidt. 2012. Multimodal Interaction in the Car: Combining Speech and Gestures on the Steering Wheel. In AutomotiveUI ’12. ACM, 155–162. https://doi.org/10.1145/2390256.2390282Google ScholarDigital Library
- B. Pogorelc, A. Lugmayr, B. Stockleben, R.-D. Vatavu, N. Tahmasebi, E. Serral, E. Stojmenova, B. Imperl, T. Risse, G. Zenz, and M. Gams. 2013. Ambient Bloom: New Business, Content, Design and Models to Increase the Semantic Ambient Media Experience. Multimedia Tools and Applications 66, 1 (2013), 7–32. http://dx.doi.org/10.1007/s11042-012-1228-4Google ScholarDigital Library
- B. Pogorelc, R.-D. Vatavu, A. Lugmayr, B. Stockleben, T. Risse, J. Kaario, E.C. Lomonaco, and M. Gams. 2012. Semantic Ambient Media: From Ambient Advertising to Ambient-Assisted Living. Multimedia Tools App. 58, 2 (2012), 399–425.Google ScholarDigital Library
- A.S.M.M. Rahman, J. Saboune, and A. El Saddik. 2011. Motion-Path Based in Car Gesture Control of the Multimedia Devices. In DIVANet ’11. ACM, New York, NY, USA, 69–76. https://doi.org/10.1145/2069000.2069013Google ScholarDigital Library
- M. Raja, V. Ghaderi, and S. Sigg. 2018. WiBot! In-Vehicle Behaviour and Gesture Recognition Using Wireless Network Edge. In ICDCS ’18. 376–387. https://doi.org/10.1109/ICDCS.2018.00045Google Scholar
- B. Reimer, B. Mehler, J. Dobres, H. McAnulty, A. Mehler, D. Munger, and A. Rumpold. 2014. Effects of an “Expert Mode” Voice Command System on Task Performance, Glance Behavior & Driver Physiology. In AutomotiveUI ’14. ACM.Google Scholar
- A. Riener, A. Ferscha, F. Bachmair, P. Hagmüller, A. Lemme, D. Muttenthaler, D. Pühringer, H. Rogner, A. Tappe, and F. Weger. 2013. Standardization of the In-Car Gesture Interaction Space. In AutomotiveUI ’13. ACM, 14–21.Google Scholar
- W. A. Rogers, A. D. Fisk, A. Collins McLaughlin, and R. Pak. 2005. Touch a Screen or Turn a Knob: Choosing the Best Device for the Job. Human Factors 47, 2 (2005), 271–288. https://doi.org/10.1518/0018720054679452Google ScholarCross Ref
- F. Roider and T. Gross. 2018. I See Your Point: Integrating Gaze to Enhance Pointing Gesture Accuracy While Driving. In AutomotiveUI ’18. ACM, New York, NY, USA, 351–358. https://doi.org/10.1145/3239060.3239084Google ScholarDigital Library
- F. Roider and K. Raab. 2018. Implementation and Evaluation of Peripheral Light Feedback for Mid-Air Gesture Interaction in the Car. In IE ’18. 87–90.Google Scholar
- F. Roider, L. Reisig, and T. Gross. 2018. Just Look: The Benefits of Gaze-Activated Voice Input in the Car. In Adjunct AutomotiveUI ’18. ACM, New York, NY, USA, 210–214. https://doi.org/10.1145/3239092.3265968Google ScholarDigital Library
- F. Roider, S. Rümelin, B. Pfleging, and T. Gross. 2017. The Effects of Situational Demands on Gaze, Speech and Gesture Input in the Vehicle. In AutomotiveUI ’17.Google Scholar
- F. Sachara, T. Kopinski, A. Gepperth, and U. Handmann. 2017. Free-hand gesture recognition with 3D-CNNs for in-car infotainment control in real-time. In ITS ’17. 959–964. https://doi.org/10.1109/ITSC.2017.8317684Google ScholarDigital Library
- P. Sauras-Perez, J. Taiber, and J. Smith. 2014. Variability analysis of in-car gesture interaction. In ICCVE ’14. 777–780. https://doi.org/10.1109/ICCVE.2014.7297657Google Scholar
- O.A. Schipor, R.-D. Vatavu, and J. Vanderdonckt. 2019. Euphoria: A Scalable, Event-Driven Architecture for Designing Interactions Across Heterogeneous Devices in Smart Environments. Inf. and Software Technology 109 (2019), 43–59.Google ScholarDigital Library
- O. A. Schipor and R.-D. Vatavu. 2019. Towards Interactions with Augmented Reality Systems in Hyper-Connected Cars. In EICS Workshops 2019. 76–82. http://ceur-ws.org/Vol-2503/paper1_12.pdfGoogle Scholar
- G. Shakeri, J. H. Williamson, and S. Brewster. 2017. Novel Multimodal Feedback Techniques for In-Car Mid-Air Gesture Interaction. In AutomotiveUI ’17. ACM, New York, NY, USA, 84–93. https://doi.org/10.1145/3122986.3123011Google ScholarDigital Library
- G. Shakeri, J. H. Williamson, and S. Brewster. 2018. May the Force Be with You: Ultrasound Haptic Feedback for Mid-Air Gesture Interaction in Cars. In AutomotiveUI ’18. ACM, 1–10. https://doi.org/10.1145/3239060.3239081Google ScholarDigital Library
- G. Shakeri, J. H. Williamson, and S. A. Brewster. 2017. Bimodal Feedback for In-Car Mid-Air Gesture Interaction. In ICMI ’17. ACM, New York, NY, USA, 518–519. https://doi.org/10.1145/3136755.3143033Google ScholarDigital Library
- M. Sheik-Nainar, J. Huber, R. Bose, and N. Matic. 2016. Force-Enabled TouchPad in Cars: Improving Target Selection Using Absolute Input. In CHI EA ’16. ACM, 2697–2704. https://doi.org/10.1145/2851581.2892390Google ScholarDigital Library
- L. Shen, S. Huo, M. Chen, F. Li, F. Yao, and H. Shen. 2017. Multi-touch gesture recognition algorithm of vehicle electronic devices-based on Bezier curve optimization strategy. In YAC ’17. 720–723. https://doi.org/10.1109/YAC.2017.7967503Google Scholar
- K. A. Smith, C. Csech, D. Murdoch, and G. Shaker. 2018. Gesture Recognition Using mm-Wave Sensor for Human-Car Interface. IEEE Sensors Letters 2, 2 (2018), 1–4. https://doi.org/10.1109/LSENS.2018.2810093Google ScholarCross Ref
- J. Sterkenburg, J. Johnson, S. Landry, and M. Jeon. 2016. Development Tool for Rapid Evaluation of Eyes-Free In-Vehicle Gesture Controls. In AutomotiveUI ’16 Adjunct. 15–19. https://doi.org/10.1145/3004323.3004357Google ScholarDigital Library
- J. Sterkenburg, S. Landry, and M. Jeon. 2017. Eyes-Free In-Vehicle Gesture Controls: Auditory-Only Displays Reduced Visual Distraction and Workload. In AutomotiveUI ’17. 195–200. https://doi.org/10.1145/3131726.3131747Google ScholarDigital Library
- A. Tamoto and K. Itou. 2019. Voice Authentication by Text Dependent Single Utterance for In-Car Environment. In SoICT ’19. ACM, New York, NY, USA, 336–341. https://doi.org/10.1145/3368926.3369669Google ScholarDigital Library
- M. P. Tarvekar. 2018. Hand Gesture Recognition System for Touch-Less Car Interface Using Multiclass Support Vector Machine. In ICICCS ’18. 1929–1932. https://doi.org/10.1109/ICCONS.2018.8663003Google Scholar
- S. Tateno, Y. Zhu, and F. Meng. 2019. Hand Gesture Recognition System for In-car Device Control Based on Infrared Array Sensor. In SICE ’19. 701–706. https://doi.org/10.23919/SICE.2019.8859832Google Scholar
- A. Tewari, B. Taetz, F. Grandidier, and D. Stricker. [n.d.]. A Probabilistic Combination of CNN and RNN Estimates for Hand Gesture Based Interaction in Car, year=2017. In ISMAR ’17 Adjunct. 1–6.Google Scholar
- M. T. Tombeng, R. Najoan, and N. Karel. 2018. Smart Car: Digital Controlling System Using Android Smartwatch Voice Recognition. In CITSM ’18. 1–5. https://doi.org/10.1109/CITSM.2018.8674359Google Scholar
- S. Tsurufuji, H. Ohnishi, M. Iida, R. Suzuki, and Y. Sumi. 1991. A Voice Activated Car Audio System. In CE ’91. 340–341. https://doi.org/10.1109/ICCE.1991.733216Google Scholar
- S. Tsurufuji, H. Ohnishi, M. Iida, R. Suzuki, and Y. Sumi. 1991. A voice activated car audio system. IEEE Transactions on Consumer Electronics 37, 3 (1991), 592–597. https://doi.org/10.1109/30.85572Google ScholarDigital Library
- J. Vanderdonckt, M. Zen, and R.-D. Vatavu. 2019. AB4Web: An On-Line A/B Tester for Comparing User Interface Design Alternatives. Proc. ACM Hum.-Comput. Interact. 3, EICS, Article 18 (June 2019), 28 pages. https://doi.org/10.1145/3331160Google ScholarDigital Library
- R.-D. Vatavu and J. O. Wobbrock. 2015. Formalizing Agreement Analysis for Elicitation Studies: New Measures, Significance Test, and Toolkit. In CHI ’15. ACM, New York, NY, USA, 1325–1334. https://doi.org/10.1145/2702123.2702223Google ScholarDigital Library
- R.-D. Vatavu and J. O. Wobbrock. 2016. Between-Subjects Elicitation Studies: Formalization and Tool Support. In CHI ’16. ACM, New York, NY, USA, 3390–3402. https://doi.org/10.1145/2858036.2858228Google ScholarDigital Library
- B. Verma and A. Choudhary. 2018. Framework for dynamic hand gesture recognition using Grassmann manifold for intelligent vehicles. IET Intelligent Transport Systems 12, 7 (2018), 721–729. https://doi.org/10.1049/iet-its.2017.0331Google ScholarCross Ref
- F. Weidner and W. Broll. 2019. Interact with Your Car: A User-Elicited Gesture Set to Inform Future in-Car User Interfaces. In MUM ’19. ACM, New York, NY, USA, Article 11, 12 pages. https://doi.org/10.1145/3365610.3365625Google ScholarDigital Library
- S. Werner. 2014. The Steering Wheel as a Touch Interface: Using Thumb-Based Gesture Interfaces as Control Inputs While Driving. In AutomotiveUI ’14 Adjunct. ACM, NY, USA, 1–4. https://doi.org/10.1145/2667239.2667299Google ScholarDigital Library
- J. O. Wobbrock, M. R. Morris, and A. D. Wilson. 2009. User-Defined Gestures for Surface Computing. In CHI ’09. ACM, New York, NY, USA, 1083–1092. https://doi.org/10.1145/1518701.1518866Google ScholarDigital Library
- P.N.Y. Wong, D.P. Brumby, H.V.R Babu, and K. Kobayashi. 2019. Voices in Self-Driving Cars Should Be Assertive to More Quickly Grab a Distracted Driver’s Attention. In AutomotiveUI ’19. 165–176. https://doi.org/10.1145/3342197.3344535Google ScholarDigital Library
- J. Wu and M. M. Trivedi. 2006. Visual Modules for Head Gesture Analysis in Intelligent Vehicle Systems. In IVS ’06. https://doi.org/10.1109/IVS.2006.1689598Google Scholar
- Q. Ye, L. Yang, and G. Xue. 2018. Hand-free Gesture Recognition for Vehicle Infotainment System Control. In VNC ’18. 1–2.Google Scholar
- P. Zheng, M. McDonald, and C. Pickering. 2008. Effects of Intuitive Voice Interfaces on Driving and In-vehicle Task Performance. In ITS ’08. 610–615. https://doi.org/10.1109/ITSC.2008.4732605Google Scholar
- M. Zobl, M. Geiger, B. Schuller, M. Lang, and G. Rigoll. 2003. A real-time system for hand gesture controlled operation of in-car devices. In ME ’03, Vol. 3. III–541. https://doi.org/10.1109/ICME.2003.1221368Google Scholar
Recommendations
An AR-Enabled Interactive Car Door to Extend In-Car Infotainment Systems for Rear Seat Passengers
CHI EA '21: Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing SystemsModern cars create a high-tech interactive space by providing entertainment and information functionalities to the driver and partly to the passengers. By introducing rear-seat infotainment systems especially in luxury cars, manufacturers started to ...
InShift: A Shifting Infotainment System to Enhance Co-Driver Experience and Collaboration
AutomotiveUI '21 Adjunct: 13th International Conference on Automotive User Interfaces and Interactive Vehicular ApplicationsCar manufacturers introduced a variety of non-driving-related features to enhance the trip experience for drivers and passengers. However, using an in-vehicle infotainment system (IVIS) can be mentally demanding and cause driver distraction. In addition,...
Interactive documentaries: A Golden Age
SPECIAL ISSUE: TV and Video Entertainment EnvironmentsThis article is motivated by the opportunity presented by recent advances in information and communication technology—particularly by faster broadband connections and faster digital media processing capabilities—for interactive television to extend and ...
Comments