Skip to main content
Erschienen in: Journal on Multimodal User Interfaces 1/2024

19.10.2023 | Original Paper

Comparing alternative modalities in the context of multimodal human–robot interaction

verfasst von: Suprakas Saren, Abhishek Mukhopadhyay, Debasish Ghose, Pradipta Biswas

Erschienen in: Journal on Multimodal User Interfaces | Ausgabe 1/2024

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

With the advancement of interactive technology, alternative input modalities are often used, instead of conventional ones, to create intuitive, efficient, and user-friendly avenues of controlling and collaborating with robots. Researchers have examined the efficacy of natural interaction modalities such as gesture or voice in single and dual-task scenarios. These investigations have aimed to address the potential of the modalities on diverse applications encompassing activities like online shopping, precision agriculture, and mechanical component assembly, which involve tasks like object pointing and selection. This article aims to address the impact on user performance in a practical human–robot interaction application where a fixed-base robot is controlled through the utilization of natural alternative modalities. We explored this by investigating the impact of single-task and dual-task conditions on user performance for object picking and dropping. We undertook two user studies—one focusing on single-task scenarios, employing a fixed-base robot for object picking and dropping and the other encompassing dual-task conditions, utilizing a mobile robot for a driving scenario. We measured task completion times and estimated cognitive workload through the NASA Task Load Index (TLX), which offers a subjective, multidimensional scale measuring the perceived cognitive workload of a user. The studies revealed that the ranking of completion times for the alternative modalities remained consistent across both single-task and dual-task scenarios. However, the ranking based on perceived cognitive load was different. In the single-task study, the gesture-based modality resulted the highest TLX score, contrasting with the dual-task study, where the highest TLX score was associated with the eye gaze-based modality. Likewise, the speech-based modality achieved a lower TLX score compared to eye gaze and gesture in the single-task study, but its TLX score in the dual-task study was between gesture and eye gaze. These outcomes suggest that the efficacy of alternative modalities is contingent not only on user preferences but also on the specific situational context.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Biswas P, Langdon P (2014) Eye-gaze tracking based interaction in India. Procedia Comput Sci 39:59–66CrossRef Biswas P, Langdon P (2014) Eye-gaze tracking based interaction in India. Procedia Comput Sci 39:59–66CrossRef
2.
Zurück zum Zitat Guo J et al (2019) A novel robotic guidance system with eye-gaze tracking control for needle-based interventions. IEEE Trans Cognit Dev Syst 13(1):179–188CrossRef Guo J et al (2019) A novel robotic guidance system with eye-gaze tracking control for needle-based interventions. IEEE Trans Cognit Dev Syst 13(1):179–188CrossRef
3.
Zurück zum Zitat Palinko O et al (2015) Eye gaze tracking for a humanoid robot. In: 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids), IEEE Palinko O et al (2015) Eye gaze tracking for a humanoid robot. In: 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids), IEEE
4.
Zurück zum Zitat Sharma S et al (2016) Gesture-based interaction for individuals with developmental disabilities in India. In: Proceedings of the 18th International ACM SIGACCESS Conference on Computers and Accessibility Sharma S et al (2016) Gesture-based interaction for individuals with developmental disabilities in India. In: Proceedings of the 18th International ACM SIGACCESS Conference on Computers and Accessibility
5.
Zurück zum Zitat He J et al (2014) Texting while driving: Is speech-based text entry less risky than handheld text entry? Accid Anal Prev 72:287–295ADSCrossRefPubMed He J et al (2014) Texting while driving: Is speech-based text entry less risky than handheld text entry? Accid Anal Prev 72:287–295ADSCrossRefPubMed
6.
Zurück zum Zitat Klamka K et al (2015) Look & pedal: Hands-free navigation in zoomable information spaces through gaze-supported foot input. In: Proceedings of the 2015 ACM on international conference on multi-modal interaction Klamka K et al (2015) Look & pedal: Hands-free navigation in zoomable information spaces through gaze-supported foot input. In: Proceedings of the 2015 ACM on international conference on multi-modal interaction
7.
Zurück zum Zitat Manawadu UE et al (2017) A multimodal human-machine interface enabling situation-Adaptive control inputs for highly automated vehicles. In: 2017 IEEE Intelligent Vehicles Symposium (IV), IEEE Manawadu UE et al (2017) A multimodal human-machine interface enabling situation-Adaptive control inputs for highly automated vehicles. In: 2017 IEEE Intelligent Vehicles Symposium (IV), IEEE
8.
Zurück zum Zitat Bolt R (1980) Put-that-there" Voice and gesture at the graphics interface,". In: Proceedings of the 7th annual conference on Computer graphics and interactive techniques Bolt R (1980) Put-that-there" Voice and gesture at the graphics interface,". In: Proceedings of the 7th annual conference on Computer graphics and interactive techniques
9.
Zurück zum Zitat Hornof AJ and Cavender A (2005) EyeDraw: enabling children with severe motor impairments to draw with their eyes. In: Proceedings of the SIGCHI conference on Human factors in computing systems Hornof AJ and Cavender A (2005) EyeDraw: enabling children with severe motor impairments to draw with their eyes. In: Proceedings of the SIGCHI conference on Human factors in computing systems
10.
Zurück zum Zitat Nancel M et al (2011) Mid-air pan-and-zoom on wall-sized displays. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems Nancel M et al (2011) Mid-air pan-and-zoom on wall-sized displays. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
11.
Zurück zum Zitat Serpiva V et al (2021) Dronepaint: swarm light painting with DNN-based gesture recognition. ACM SIGGRAPH 2021 Emerging Technologies, pp 1-4 Serpiva V et al (2021) Dronepaint: swarm light painting with DNN-based gesture recognition. ACM SIGGRAPH 2021 Emerging Technologies, pp 1-4
12.
Zurück zum Zitat Yam-Viramontes B et al (2022) Commanding a drone through body poses, improving the user experience. J Multimod User Interfaces 16(4):357–369CrossRef Yam-Viramontes B et al (2022) Commanding a drone through body poses, improving the user experience. J Multimod User Interfaces 16(4):357–369CrossRef
13.
Zurück zum Zitat Majaranta P and Räihä K-J (2002) Twenty years of eye typing: systems and design issues. In: Proceedings of the 2002 symposium on Eye tracking research & applications Majaranta P and Räihä K-J (2002) Twenty years of eye typing: systems and design issues. In: Proceedings of the 2002 symposium on Eye tracking research & applications
14.
Zurück zum Zitat Kumar M et al (2007) Eyepoint: practical pointing and selection using gaze and keyboard. In: Proceedings of the SIGCHI conference on Human factors in computing systems Kumar M et al (2007) Eyepoint: practical pointing and selection using gaze and keyboard. In: Proceedings of the SIGCHI conference on Human factors in computing systems
15.
Zurück zum Zitat Sharma VK et al (2020) Eye gaze controlled robotic arm for persons with severe speech and motor impairment. In: ACM Symposium on Eye Tracking Research and Applications Sharma VK et al (2020) Eye gaze controlled robotic arm for persons with severe speech and motor impairment. In: ACM Symposium on Eye Tracking Research and Applications
16.
Zurück zum Zitat Oviatt S (1999) Ten myths of multimodal interaction. Commun ACM 42(11):74–81CrossRef Oviatt S (1999) Ten myths of multimodal interaction. Commun ACM 42(11):74–81CrossRef
17.
Zurück zum Zitat Lee M et al (2013) A usability study of multimodal input in an augmented reality environment. Virt Real 17(4):293–305CrossRef Lee M et al (2013) A usability study of multimodal input in an augmented reality environment. Virt Real 17(4):293–305CrossRef
18.
Zurück zum Zitat Hürst W, Van Wezel C (2013) Gesture-based interaction via finger tracking for mobile augmented reality. Multimed Tools Appl 62(1):233–258CrossRef Hürst W, Van Wezel C (2013) Gesture-based interaction via finger tracking for mobile augmented reality. Multimed Tools Appl 62(1):233–258CrossRef
19.
Zurück zum Zitat M. Van den Bergh, et al., "Real-time 3D hand gesture interaction with a robot for understanding directions from humans," 2011 Ro-Man. IEEE, 2011. M. Van den Bergh, et al., "Real-time 3D hand gesture interaction with a robot for understanding directions from humans," 2011 Ro-Man. IEEE, 2011.
20.
Zurück zum Zitat Alvarez-Santos V et al (2014) Gesture-based interaction with voice feedback for a tour-guide robot. J Vis Commun Image Represent 25(2):499–509CrossRef Alvarez-Santos V et al (2014) Gesture-based interaction with voice feedback for a tour-guide robot. J Vis Commun Image Represent 25(2):499–509CrossRef
21.
Zurück zum Zitat Haddadi A et al (2013) Analysis of task-based gestures in human-robot interaction. In: 2013 IEEE International Conference on Robotics and Automation, IEEE Haddadi A et al (2013) Analysis of task-based gestures in human-robot interaction. In: 2013 IEEE International Conference on Robotics and Automation, IEEE
22.
Zurück zum Zitat Al Mahmud J et al (2022) 3D gesture recognition and adaptation for human-robot interaction. IEEE Access 10:116485–116513CrossRef Al Mahmud J et al (2022) 3D gesture recognition and adaptation for human-robot interaction. IEEE Access 10:116485–116513CrossRef
23.
Zurück zum Zitat Coronado E et al (2017) Gesture-based robot control: Design challenges and evaluation with humans. In: 2017 IEEE international conference on robotics and automation (ICRA), IEEE Coronado E et al (2017) Gesture-based robot control: Design challenges and evaluation with humans. In: 2017 IEEE international conference on robotics and automation (ICRA), IEEE
24.
Zurück zum Zitat Hettig J et al (2017) Comparison of gesture and conventional interaction techniques for interventional neuroradiology. Int J Comput Assist Radiol Surg 12(9):1643–1653CrossRefPubMed Hettig J et al (2017) Comparison of gesture and conventional interaction techniques for interventional neuroradiology. Int J Comput Assist Radiol Surg 12(9):1643–1653CrossRefPubMed
25.
Zurück zum Zitat Gips J and Olivieri P (1996) EagleEyes: an eye control system for persons with disabilities. In: The eleventh international conference on technology and persons with disabilities Gips J and Olivieri P (1996) EagleEyes: an eye control system for persons with disabilities. In: The eleventh international conference on technology and persons with disabilities
26.
Zurück zum Zitat Wobbrock JO et al (2008) Longitudinal evaluation of discrete consecutive gaze gestures for text entry. In: Proceedings of the 2008 symposium on Eye tracking research & applications Wobbrock JO et al (2008) Longitudinal evaluation of discrete consecutive gaze gestures for text entry. In: Proceedings of the 2008 symposium on Eye tracking research & applications
27.
Zurück zum Zitat Biswas P, Langdon P (2015) Multimodal intelligent eye-gaze tracking system. Int Jf Human-Comput Interact 31(4):277–294CrossRef Biswas P, Langdon P (2015) Multimodal intelligent eye-gaze tracking system. Int Jf Human-Comput Interact 31(4):277–294CrossRef
28.
Zurück zum Zitat Murthy LRD et al (2021) Eye-gaze-controlled HMDS and MFD for military aircraft. J Aviat Technol Eng 10(2):34CrossRef Murthy LRD et al (2021) Eye-gaze-controlled HMDS and MFD for military aircraft. J Aviat Technol Eng 10(2):34CrossRef
29.
Zurück zum Zitat Lim Y et al (2018) Eye-tracking sensors for adaptive aerospace human-machine interfaces and interactions. In: 2018 5th IEEE international workshop on metrology for aerospace (MetroAeroSpace), IEEE Lim Y et al (2018) Eye-tracking sensors for adaptive aerospace human-machine interfaces and interactions. In: 2018 5th IEEE international workshop on metrology for aerospace (MetroAeroSpace), IEEE
30.
Zurück zum Zitat Jannette M, Vollrath M (2009) Comparison of manual vs. speech-based interaction with in-vehicle information systems. Accid Anal Prev 41(5):924–930CrossRef Jannette M, Vollrath M (2009) Comparison of manual vs. speech-based interaction with in-vehicle information systems. Accid Anal Prev 41(5):924–930CrossRef
31.
Zurück zum Zitat Lee JD et al (2001) Speech-based interaction with in-vehicle computers: the effect of speech-based email on drivers’ attention to the roadway. Hum Factors 43(4):631–640CrossRefPubMed Lee JD et al (2001) Speech-based interaction with in-vehicle computers: the effect of speech-based email on drivers’ attention to the roadway. Hum Factors 43(4):631–640CrossRefPubMed
32.
Zurück zum Zitat Doyle J and Bertolotto M (2006) Combining speech and pen input for effective interaction in mobile geospatial environments. In: Proceedings of the 2006 ACM symposium on Applied computing Doyle J and Bertolotto M (2006) Combining speech and pen input for effective interaction in mobile geospatial environments. In: Proceedings of the 2006 ACM symposium on Applied computing
33.
Zurück zum Zitat Fröhlich J and Wachsmuth I (2013) The visual, the auditory and the haptic–a user study on combining modalities in virtual worlds. In: International Conference on Virtual, Augmented and Mixed Reality, Springer, Berlin, Heidelberg Fröhlich J and Wachsmuth I (2013) The visual, the auditory and the haptic–a user study on combining modalities in virtual worlds. In: International Conference on Virtual, Augmented and Mixed Reality, Springer, Berlin, Heidelberg
34.
Zurück zum Zitat Frisch M et al (2009) Investigating multi-touch and pen gestures for diagram editing on interactive surfaces. In: Proceedings of the ACM international conference on interactive tabletops and surfaces Frisch M et al (2009) Investigating multi-touch and pen gestures for diagram editing on interactive surfaces. In: Proceedings of the ACM international conference on interactive tabletops and surfaces
35.
Zurück zum Zitat Pfeuffer K et al (2014) Gaze-touch: combining gaze with multi-touch for interaction on the same surface. In: Proceedings of the 27th annual ACM symposium on User interface software and technology Pfeuffer K et al (2014) Gaze-touch: combining gaze with multi-touch for interaction on the same surface. In: Proceedings of the 27th annual ACM symposium on User interface software and technology
36.
Zurück zum Zitat Hatscher B and Hansen C (2018) Hand, foot or voice: alternative input modalities for touchless interaction in the medical domain. In: Proceedings of the 20th ACM international conference on multi-modal interaction Hatscher B and Hansen C (2018) Hand, foot or voice: alternative input modalities for touchless interaction in the medical domain. In: Proceedings of the 20th ACM international conference on multi-modal interaction
37.
Zurück zum Zitat Chen Z et al (2017) Multi-modal interaction in augmented reality. In: 2017 IEEE international conference on systems, man, and cybernetics (SMC), IEEE Chen Z et al (2017) Multi-modal interaction in augmented reality. In: 2017 IEEE international conference on systems, man, and cybernetics (SMC), IEEE
38.
Zurück zum Zitat Prabhakar G et al (2020) Interactive gaze and finger controlled HUD for cars. J Multi-Modal User Interfaces 14(1):101–121CrossRef Prabhakar G et al (2020) Interactive gaze and finger controlled HUD for cars. J Multi-Modal User Interfaces 14(1):101–121CrossRef
39.
Zurück zum Zitat Palinko O et al (2016) Robot reading human gaze: Why eye tracking is better than head tracking for human-robot collaboration. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE Palinko O et al (2016) Robot reading human gaze: Why eye tracking is better than head tracking for human-robot collaboration. In: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE
40.
Zurück zum Zitat Craig TL et al (2016) Human gaze commands classification: a shape based approach to interfacing with robots. In: 2016 12th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA), IEEE Craig TL et al (2016) Human gaze commands classification: a shape based approach to interfacing with robots. In: 2016 12th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA), IEEE
41.
Zurück zum Zitat Rudzicz F et al (2015) Speech interaction with personal assistive robots supporting aging at home for individuals with Alzheimer’s disease. ACM Trans Access Comput (TACCESS) 7(2):1–22CrossRef Rudzicz F et al (2015) Speech interaction with personal assistive robots supporting aging at home for individuals with Alzheimer’s disease. ACM Trans Access Comput (TACCESS) 7(2):1–22CrossRef
42.
Zurück zum Zitat Prodanov PJ et al (2002) Voice enabled interface for interactive tour-guide robots. In: IEEE/RSJ International Conference on Intelligent Robots and Systems. Vol 2, IEEE Prodanov PJ et al (2002) Voice enabled interface for interactive tour-guide robots. In: IEEE/RSJ International Conference on Intelligent Robots and Systems. Vol 2, IEEE
43.
Zurück zum Zitat Zinchenko K et al (2016) A study on speech recognition control for a surgical robot. IEEE Trans Industr Inf 13(2):607–615CrossRef Zinchenko K et al (2016) A study on speech recognition control for a surgical robot. IEEE Trans Industr Inf 13(2):607–615CrossRef
44.
Zurück zum Zitat Bingol MC, Aydogmus O (2020) Performing predefined tasks using the human–robot interaction on speech recognition for an industrial robot. Eng Appl Artif Intell 95:103903CrossRef Bingol MC, Aydogmus O (2020) Performing predefined tasks using the human–robot interaction on speech recognition for an industrial robot. Eng Appl Artif Intell 95:103903CrossRef
45.
Zurück zum Zitat Kurnia R et al (2004) Object recognition through human-robot interaction by speech. RO-MAN 2004. In: 13th IEEE International Workshop on Robot and Human Interactive Communication (IEEE Catalog No. 04TH8759), IEEE Kurnia R et al (2004) Object recognition through human-robot interaction by speech. RO-MAN 2004. In: 13th IEEE International Workshop on Robot and Human Interactive Communication (IEEE Catalog No. 04TH8759), IEEE
46.
Zurück zum Zitat Bannat A et al (2009) A multimodal human-robot-interaction scenario: working together with an industrial robot. In: International conference on human-computer interaction, Springer, Berlin, Heidelberg Bannat A et al (2009) A multimodal human-robot-interaction scenario: working together with an industrial robot. In: International conference on human-computer interaction, Springer, Berlin, Heidelberg
47.
Zurück zum Zitat Randelli G et al (2013) Knowledge acquisition through human–robot multimodal interaction. Intel Serv Robot 6(1):19–31CrossRef Randelli G et al (2013) Knowledge acquisition through human–robot multimodal interaction. Intel Serv Robot 6(1):19–31CrossRef
56.
Zurück zum Zitat Hart SG, Staveland LE (1988) Development of NASA-TLX (Task Load Index): results of empirical and theoretical research. Adv Psychol 52:139–183CrossRef Hart SG, Staveland LE (1988) Development of NASA-TLX (Task Load Index): results of empirical and theoretical research. Adv Psychol 52:139–183CrossRef
57.
Zurück zum Zitat Biswas P and Dv J (2018) Eye gaze controlled MFD for military aviation. In: 23rd International Conference on Intelligent User Interfaces Biswas P and Dv J (2018) Eye gaze controlled MFD for military aviation. In: 23rd International Conference on Intelligent User Interfaces
58.
Zurück zum Zitat Karpov A and Ronzhin A (2014) A universal assistive technology with multimodal input and multimedia output interfaces. Universal Access in Human-Computer Interaction. Design and Development Methods for Universal Access: 8th International Conference, UAHCI 2014, Held as Part of HCI International 2014, Heraklion, Crete, Greece, Jun 22–27, 2014, Proceedings, Part I 8, Springer International Publishing Karpov A and Ronzhin A (2014) A universal assistive technology with multimodal input and multimedia output interfaces. Universal Access in Human-Computer Interaction. Design and Development Methods for Universal Access: 8th International Conference, UAHCI 2014, Held as Part of HCI International 2014, Heraklion, Crete, Greece, Jun 22–27, 2014, Proceedings, Part I 8, Springer International Publishing
59.
Zurück zum Zitat Mukhopadhyay A et al (2019) Comparing CNNs for non-conventional traffic participants. In: Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications: Adjunct Proceedings Mukhopadhyay A et al (2019) Comparing CNNs for non-conventional traffic participants. In: Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications: Adjunct Proceedings
Metadaten
Titel
Comparing alternative modalities in the context of multimodal human–robot interaction
verfasst von
Suprakas Saren
Abhishek Mukhopadhyay
Debasish Ghose
Pradipta Biswas
Publikationsdatum
19.10.2023
Verlag
Springer International Publishing
Erschienen in
Journal on Multimodal User Interfaces / Ausgabe 1/2024
Print ISSN: 1783-7677
Elektronische ISSN: 1783-8738
DOI
https://doi.org/10.1007/s12193-023-00421-w

Weitere Artikel der Ausgabe 1/2024

Journal on Multimodal User Interfaces 1/2024 Zur Ausgabe

Premium Partner