Skip to main content
Erschienen in: Autonomous Robots 5/2018

06.11.2017

Early prediction for physical human robot collaboration in the operating room

verfasst von: Tian Zhou, Juan Pablo Wachs

Erschienen in: Autonomous Robots | Ausgabe 5/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

To enable a natural and fluent human robot collaboration flow, it is critical for a robot to comprehend their human peers’ on-going actions, predict their behaviors in the near future, and plan its actions correspondingly. Specifically, the capability of making early predictions is important, so that the robot can foresee the precise timing of a turn-taking event and start motion planning and execution early enough to smooth the turn-taking transition. Such proactive behavior would reduce human’s waiting time, increase efficiency and enhance naturalness in collaborative task. To that end, this paper presents the design and implementation of an early turn-taking prediction algorithm, catered for physical human robot collaboration scenarios. Specifically, a robotic scrub nurse system which can comprehend surgeon’s multimodal communication cues and perform turn-taking prediction is presented. The developed algorithm was tested on a collected data set of simulated surgical procedures in a surgeon–nurse tandem. The proposed turn-taking prediction algorithm is found to be significantly superior to its algorithmic counterparts, and is more accurate than human baseline when little partial input is given (less than 30% of full action). After observing more information, the algorithm can achieve comparable performances as humans with a F1 score of 0.90.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G. S., Davis, A., Dean, J., Devin, M., et al. (2015). TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from tensorflow. org (Vol. 1). Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G. S., Davis, A., Dean, J., Devin, M., et al. (2015). TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from tensorflow. org (Vol. 1).
Zurück zum Zitat Andersen, D., Popescu, V., Cabrera, M. E., Shanghavi, A., Gomez, G., Marley, S., et al. (2016). Virtual annotations of the surgical field through an augmented reality transparent display. The Visual Computer, 32(11), 1481–1498.CrossRef Andersen, D., Popescu, V., Cabrera, M. E., Shanghavi, A., Gomez, G., Marley, S., et al. (2016). Virtual annotations of the surgical field through an augmented reality transparent display. The Visual Computer, 32(11), 1481–1498.CrossRef
Zurück zum Zitat Arsikere, H., Shriberg, E., Ozertem, U. (2015). Enhanced end-of-turn detection for speech to a personal assistant. In 2015 AAAI Spring symposium series. Arsikere, H., Shriberg, E., Ozertem, U. (2015). Enhanced end-of-turn detection for speech to a personal assistant. In 2015 AAAI Spring symposium series.
Zurück zum Zitat Bartlett, M. S., Littlewort, G., Fasel, I., Movellan, J. R. (2003). Real time face detection and facial expression recognition: Development and applications to human computer interaction. In Conference on computer vision and pattern recognition workshop, 2003. CVPRW’03 (Vol. 5, pp. 53–53). IEEE. Bartlett, M. S., Littlewort, G., Fasel, I., Movellan, J. R. (2003). Real time face detection and facial expression recognition: Development and applications to human computer interaction. In Conference on computer vision and pattern recognition workshop, 2003. CVPRW’03 (Vol. 5, pp. 53–53). IEEE.
Zurück zum Zitat Bergstra, J., & Bengio, Y. (2012). Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13, 281–305.MathSciNetMATH Bergstra, J., & Bengio, Y. (2012). Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13, 281–305.MathSciNetMATH
Zurück zum Zitat Buerhaus, P. I., Auerbach, D. I., & Staiger, D. O. (2009). The recent surge in nurse employment: Causes and implications. Health Affairs, 28(4), w657–w668.CrossRef Buerhaus, P. I., Auerbach, D. I., & Staiger, D. O. (2009). The recent surge in nurse employment: Causes and implications. Health Affairs, 28(4), w657–w668.CrossRef
Zurück zum Zitat Cakmak, M., Srinivasa, S. S., Lee, M. K., Kiesler, S., Forlizzi, J. (2011). Using spatial and temporal contrast for fluent robot–human hand-overs. In Proceedings of the 6th international conference on human–robot interaction (pp. 489–496). ACM. Cakmak, M., Srinivasa, S. S., Lee, M. K., Kiesler, S., Forlizzi, J. (2011). Using spatial and temporal contrast for fluent robot–human hand-overs. In Proceedings of the 6th international conference on human–robot interaction (pp. 489–496). ACM.
Zurück zum Zitat Calisgan, E., Haddadi, A., Van der Loos, H. M., Alcazar, J. A., Croft, E. A. (2012). Identifying nonverbal cues for automated human-robot turn-taking. In Robot and human interactive communication (RO-MAN 2012). 21st IEEE International Symposium (pp. 418–423). IEEE. Calisgan, E., Haddadi, A., Van der Loos, H. M., Alcazar, J. A., Croft, E. A. (2012). Identifying nonverbal cues for automated human-robot turn-taking. In Robot and human interactive communication (RO-MAN 2012). 21st IEEE International Symposium (pp. 418–423). IEEE.
Zurück zum Zitat Chan, W. P., Kakiuchi, Y., Okada, K., Inaba, M. (2014). Determining proper grasp configurations for handovers through observation of object movement patterns and inter-object interactions during usage. In 2014 IEEE/RSJ international conference on intelligent robots and systems (IROS 2014) (pp. 1355–1360). IEEE. Chan, W. P., Kakiuchi, Y., Okada, K., Inaba, M. (2014). Determining proper grasp configurations for handovers through observation of object movement patterns and inter-object interactions during usage. In 2014 IEEE/RSJ international conference on intelligent robots and systems (IROS 2014) (pp. 1355–1360). IEEE.
Zurück zum Zitat Ehrlich, S., Wykowska, A., Ramirez-Amaro, K., Cheng, G. (2014). When to engage in interaction #x2014; And how? EEG-based enhancement of robot’s ability to sense social signals in HRI. In 2014 14th IEEE-RAS international conference on humanoid robots (humanoids), November 2014, pp. 1104–1109. Ehrlich, S., Wykowska, A., Ramirez-Amaro, K., Cheng, G. (2014). When to engage in interaction #x2014; And how? EEG-based enhancement of robot’s ability to sense social signals in HRI. In 2014 14th IEEE-RAS international conference on humanoid robots (humanoids), November 2014, pp. 1104–1109.
Zurück zum Zitat Esterman, M., Tamber-Rosenau, B. J., Chiu, Y.-C., & Yantis, S. (2010). Avoiding non-independence in fmri data analysis: Leave one subject out. Neuroimage, 50(2), 572–576.CrossRef Esterman, M., Tamber-Rosenau, B. J., Chiu, Y.-C., & Yantis, S. (2010). Avoiding non-independence in fmri data analysis: Leave one subject out. Neuroimage, 50(2), 572–576.CrossRef
Zurück zum Zitat Harmanec, D., & Klir, G. J. (1994). Measuring total uncertainty in Dempster–Shafer theory: A novel approach. International Journal of General System, 22(4), 405–419.CrossRefMATH Harmanec, D., & Klir, G. J. (1994). Measuring total uncertainty in Dempster–Shafer theory: A novel approach. International Journal of General System, 22(4), 405–419.CrossRefMATH
Zurück zum Zitat Hinton, G. (2010). A practical guide to training restricted Boltzmann machines. Momentum, 9(1), 926. Hinton, G. (2010). A practical guide to training restricted Boltzmann machines. Momentum, 9(1), 926.
Zurück zum Zitat Hughes, K. F., Murphy, R. R. (1992). Ultrasonic robot localization using Dempster–Shafer theory. In San Diego’92 (pp. 2–11). International Society for Optics and Photonics. Hughes, K. F., Murphy, R. R. (1992). Ultrasonic robot localization using Dempster–Shafer theory. In San Diego’92 (pp. 2–11). International Society for Optics and Photonics.
Zurück zum Zitat Jeni, L. A., Cohn, J. F., De La Torre, F. (2013). Facing imbalanced data-recommendations for the use of performance metrics. In 2013 Humaine association conference on affective computing and intelligent interaction (ACII) (pp. 245–251). IEEE. Jeni, L. A., Cohn, J. F., De La Torre, F. (2013). Facing imbalanced data-recommendations for the use of performance metrics. In 2013 Humaine association conference on affective computing and intelligent interaction (ACII) (pp. 245–251). IEEE.
Zurück zum Zitat Kirk, R. E. (1982). Experimental design. New York: Wiley. Kirk, R. E. (1982). Experimental design. New York: Wiley.
Zurück zum Zitat Klir, G., & Yuan, B. (1995). Fuzzy sets and fuzzy logic (Vol. 4). New Jersey: Prentice Hall.MATH Klir, G., & Yuan, B. (1995). Fuzzy sets and fuzzy logic (Vol. 4). New Jersey: Prentice Hall.MATH
Zurück zum Zitat Kose-Bagci, H., Dautenhahn, K., Nehaniv, C. L. (2008). Emergent dynamics of turn-taking interaction in drumming games with a humanoid robot. In RO-MAN 2008—The 17th IEEE international symposium on robot and human interactive communication, August 2008, pp. 346–353. Kose-Bagci, H., Dautenhahn, K., Nehaniv, C. L. (2008). Emergent dynamics of turn-taking interaction in drumming games with a humanoid robot. In RO-MAN 2008—The 17th IEEE international symposium on robot and human interactive communication, August 2008, pp. 346–353.
Zurück zum Zitat Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I. (2010). The extended Cohn–Kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In 2010 IEEE computer society conference on computer vision and pattern recognition workshops (CVPRW) (pp. 94–101). IEEE. Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I. (2010). The extended Cohn–Kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In 2010 IEEE computer society conference on computer vision and pattern recognition workshops (CVPRW) (pp. 94–101). IEEE.
Zurück zum Zitat Mei, J., Liu, M., Wang, Y. F., & Gao, H. (2016). Learning a mahalanobis distance-based dynamic time warping measure for multivariate time series classification. IEEE Transactions on Cybernetics, 46(6), 1363–1374.CrossRef Mei, J., Liu, M., Wang, Y. F., & Gao, H. (2016). Learning a mahalanobis distance-based dynamic time warping measure for multivariate time series classification. IEEE Transactions on Cybernetics, 46(6), 1363–1374.CrossRef
Zurück zum Zitat Morency, L.-P., de Kok, I., Gratch, J. (2008). Context-based recognition during human interactions: Automatic feature selection and encoding dictionary. In Proceedings of the 10th international conference on multimodal interfaces (pp. 181–188). ACM. Available: https://dl.acm.org/citation.cfm?id=1452426. Morency, L.-P., de Kok, I., Gratch, J. (2008). Context-based recognition during human interactions: Automatic feature selection and encoding dictionary. In Proceedings of the 10th international conference on multimodal interfaces (pp. 181–188). ACM. Available: https://​dl.​acm.​org/​citation.​cfm?​id=​1452426.
Zurück zum Zitat Mutlu, B., Shiwa, T., Kanda, T., Ishiguro, H., Hagita, N. (2009). Footing in human–robot conversations: How robots might shape participant roles using gaze cues. In Proceedings of the 4th ACM/IEEE international conference on human robot interaction (pp. 61–68). ACM. Available: https://dl.acm.org/citation.cfm?id=1514109. Mutlu, B., Shiwa, T., Kanda, T., Ishiguro, H., Hagita, N. (2009). Footing in human–robot conversations: How robots might shape participant roles using gaze cues. In Proceedings of the 4th ACM/IEEE international conference on human robot interaction (pp. 61–68). ACM. Available: https://​dl.​acm.​org/​citation.​cfm?​id=​1514109.
Zurück zum Zitat Needleman, J., Buerhaus, P., Pankratz, V. S., Leibson, C. L., Stevens, S. R., & Harris, M. (2011). Nurse staffing and inpatient hospital mortality. New England Journal of Medicine, 364(11), 1037–1045.CrossRef Needleman, J., Buerhaus, P., Pankratz, V. S., Leibson, C. L., Stevens, S. R., & Harris, M. (2011). Nurse staffing and inpatient hospital mortality. New England Journal of Medicine, 364(11), 1037–1045.CrossRef
Zurück zum Zitat Ohn-Bar, E., & Trivedi, M. M. (2014). Hand gesture recognition in real time for automotive interfaces: A multimodal vision-based approach and evaluations. IEEE Transactions on Intelligent Transportation Systems, 15(6), 2368–2377.CrossRef Ohn-Bar, E., & Trivedi, M. M. (2014). Hand gesture recognition in real time for automotive interfaces: A multimodal vision-based approach and evaluations. IEEE Transactions on Intelligent Transportation Systems, 15(6), 2368–2377.CrossRef
Zurück zum Zitat Rabl, L. I., Andersen, M. L., stergaard, D., Bjrn, B., Lilja, B., Mogensen, T. (2011). Republished error management: Descriptions of verbal communication errors between staff. An analysis of 84 root cause analysis-reports from Danish hospitals. Postgraduate Medical Journal, 87(1033), 783–789. Available: https://pmj.bmj.com/content/87/1033/783. Rabl, L. I., Andersen, M. L., stergaard, D., Bjrn, B., Lilja, B., Mogensen, T. (2011). Republished error management: Descriptions of verbal communication errors between staff. An analysis of 84 root cause analysis-reports from Danish hospitals. Postgraduate Medical Journal, 87(1033), 783–789. Available: https://​pmj.​bmj.​com/​content/​87/​1033/​783.
Zurück zum Zitat Raux, A., Eskenazi, M. (2008). Optimizing endpointing thresholds using dialogue features in a spoken dialogue system. In Proceedings of the 9th SIGdial workshop on discourse and dialogue (pp. 1–10). Association for Computational Linguistics. Available: https://dl.acm.org/citation.cfm?id=1622066. Raux, A., Eskenazi, M. (2008). Optimizing endpointing thresholds using dialogue features in a spoken dialogue system. In Proceedings of the 9th SIGdial workshop on discourse and dialogue (pp. 1–10). Association for Computational Linguistics. Available: https://​dl.​acm.​org/​citation.​cfm?​id=​1622066.
Zurück zum Zitat Unhelkar, V. V., Shah, J. A., Siu, H. C. (2014). Comparative performance of human and mobile robotic assistants in collaborative fetch-and-deliver tasks. In Proceedings of the 2014 ACM/IEEE international conference on Human–robot interaction (pp. 82–89). ACM. Unhelkar, V. V., Shah, J. A., Siu, H. C. (2014). Comparative performance of human and mobile robotic assistants in collaborative fetch-and-deliver tasks. In Proceedings of the 2014 ACM/IEEE international conference on Human–robot interaction (pp. 82–89). ACM.
Zurück zum Zitat Wu, H., Siegel, M., Stiefelhagen, R., Yang, J. (2002). Sensor fusion using Dempster–Shafer theory [for context-aware hci]. In Instrumentation and Measurement Technology Conference, 2002. IMTC/2002. Proceedings of the 19th IEEE (Vol. 1, pp. 7–12). IEEE. Wu, H., Siegel, M., Stiefelhagen, R., Yang, J. (2002). Sensor fusion using Dempster–Shafer theory [for context-aware hci]. In Instrumentation and Measurement Technology Conference, 2002. IMTC/2002. Proceedings of the 19th IEEE (Vol. 1, pp. 7–12). IEEE.
Zurück zum Zitat Yamazaki, A., Yamazaki, K., Kuno, Y., Burdelski, M., Kawashima, M., Kuzuoka, H. (2008). Precision timing in human–robot interaction: Coordination of head movement and utterance. In Proceedings of the SIGCHI conference on human factors in computing systems (pp. 131–140). ACM. Available: https://dl.acm.org/citation.cfm?id=1357077. Yamazaki, A., Yamazaki, K., Kuno, Y., Burdelski, M., Kawashima, M., Kuzuoka, H. (2008). Precision timing in human–robot interaction: Coordination of head movement and utterance. In Proceedings of the SIGCHI conference on human factors in computing systems (pp. 131–140). ACM. Available: https://​dl.​acm.​org/​citation.​cfm?​id=​1357077.
Zurück zum Zitat Zheng, W., Zhou, X., Zou, C., & Zhao, L. (2006). Facial expression recognition using kernel canonical correlation analysis (KCCA). IEEE Transactions on Neural Networks, 17(1), 233–238.CrossRef Zheng, W., Zhou, X., Zou, C., & Zhao, L. (2006). Facial expression recognition using kernel canonical correlation analysis (KCCA). IEEE Transactions on Neural Networks, 17(1), 233–238.CrossRef
Metadaten
Titel
Early prediction for physical human robot collaboration in the operating room
verfasst von
Tian Zhou
Juan Pablo Wachs
Publikationsdatum
06.11.2017
Verlag
Springer US
Erschienen in
Autonomous Robots / Ausgabe 5/2018
Print ISSN: 0929-5593
Elektronische ISSN: 1573-7527
DOI
https://doi.org/10.1007/s10514-017-9670-9

Weitere Artikel der Ausgabe 5/2018

Autonomous Robots 5/2018 Zur Ausgabe

Neuer Inhalt