Skip to main content
Top
Published in: Autonomous Robots 3/2014

01-03-2014

A saliency-driven robotic head with bio-inspired saccadic behaviors for social robotics

Authors: Hongsheng He, Shuzhi Sam Ge, Zhengchen Zhang

Published in: Autonomous Robots | Issue 3/2014

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

This paper presents a robotic head for social robots to attend to scene saliency with bio-inspired saccadic behaviors. Scene saliency is determined by measuring low-level static scene information, motion, and object prior knowledge. Towards the extracted saliency spots, the designed robotic head is able to turn gazes in a saccadic manner while obeying eye–head coordination laws with the proposed control scheme. The results of the simulation study and actual applications show the effectiveness of the proposed method in discovering of scene saliency and human-like head motion. The proposed techniques could possibly be applied to social robots to improve social sense and user experience in human–robot interaction.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Appendix
Available only for authorised users
Literature
go back to reference Asfour, T., Welke, K., Azad, P., Ude, A., & Dillmann, R. (2008). The Karlsruhe humanoid head. In Proceedings of IEEE-RAS international conference on humanoid robots (pp. 447–453). Asfour, T., Welke, K., Azad, P., Ude, A., & Dillmann, R. (2008). The Karlsruhe humanoid head. In Proceedings of IEEE-RAS international conference on humanoid robots (pp. 447–453).
go back to reference Breazeal, C. (2000). Sociable machines: Expressive social exchange between humans and robots. Ph.D. thesis, Massachusetts Institute of Technology. Breazeal, C. (2000). Sociable machines: Expressive social exchange between humans and robots. Ph.D. thesis, Massachusetts Institute of Technology.
go back to reference Butko, N., Zhang, L., Cottrell, G., & Movellan J. (2008). Visual saliency model for robot cameras. In IEEE international conference on robotics and automation (pp. 2398–2403). Butko, N., Zhang, L., Cottrell, G., & Movellan J. (2008). Visual saliency model for robot cameras. In IEEE international conference on robotics and automation (pp. 2398–2403).
go back to reference Choi, S.-B., Ban, S.-W., & Lee, M. (2004). Biologically motivated visual attention system using bottom-up saliency map and top-down inhibition. Neural Information Processing—Letters and Review, 2(1), 19–25. Choi, S.-B., Ban, S.-W., & Lee, M. (2004). Biologically motivated visual attention system using bottom-up saliency map and top-down inhibition. Neural Information Processing—Letters and Review, 2(1), 19–25.
go back to reference Corbetta, M., & Shulman, G. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews Neuroscience, 3, 201–215.CrossRef Corbetta, M., & Shulman, G. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews Neuroscience, 3, 201–215.CrossRef
go back to reference Crawford, J., Martinez-Trujillo, J., & Klier, E. (2003). Neural control of three-dimensional eye and head movements. Current Opinion in Neurobiology, 13(6), 655–662.CrossRef Crawford, J., Martinez-Trujillo, J., & Klier, E. (2003). Neural control of three-dimensional eye and head movements. Current Opinion in Neurobiology, 13(6), 655–662.CrossRef
go back to reference Crawford, J., & Vilis, T. (1991). Axes of eye rotation and listing’s law during rotations of the head. Journal of Neurophysiology, 65(3), 407–423. Crawford, J., & Vilis, T. (1991). Axes of eye rotation and listing’s law during rotations of the head. Journal of Neurophysiology, 65(3), 407–423.
go back to reference Dalal, N., & Triggs, B. (2005). Histograms of oriented gradients for human detection. In IEEE Computer Society conference on computer vision and pattern recognition. Dalal, N., & Triggs, B. (2005). Histograms of oriented gradients for human detection. In IEEE Computer Society conference on computer vision and pattern recognition.
go back to reference Donders, F. (1848). Beitrag zur lehre von den bewegungen des menschlichen auges. Holland Beitr Anat Physiol Wiss, 1(104), 384. Donders, F. (1848). Beitrag zur lehre von den bewegungen des menschlichen auges. Holland Beitr Anat Physiol Wiss, 1(104), 384.
go back to reference Doretto, G., Chiuso, A., Wu, Y., & Soatto, S. (2003). Dynamic textures. International Journal of Computer Vision, 51(2), 91–109.CrossRefMATH Doretto, G., Chiuso, A., Wu, Y., & Soatto, S. (2003). Dynamic textures. International Journal of Computer Vision, 51(2), 91–109.CrossRefMATH
go back to reference Gao, D., & Vasconcelos, N. (2007). Bottom-up saliency is a discriminant process. In IEEE 11th international conference on computer vision, ICCV 2007 (pp. 1–6). Gao, D., & Vasconcelos, N. (2007). Bottom-up saliency is a discriminant process. In IEEE 11th international conference on computer vision, ICCV 2007 (pp. 1–6).
go back to reference Ge, S., He, H., & Zhang, Z. (2011). Bottom-up saliency detection for attention determination. Machine Vision and Applications, 24, 1–14. Ge, S., He, H., & Zhang, Z. (2011). Bottom-up saliency detection for attention determination. Machine Vision and Applications, 24, 1–14.
go back to reference Glenn, B., & Vilis, T. (1992). Violations of listing’s law after large eye and head gaze shifts. Journal of Neurophysiology, 68(1), 309–318. Glenn, B., & Vilis, T. (1992). Violations of listing’s law after large eye and head gaze shifts. Journal of Neurophysiology, 68(1), 309–318.
go back to reference Goossens, H., & Opstal, A. (1997). Human eye–head coordination in two dimensions under different sensorimotor conditions. Experimental Brain Research, 114(3), 542–560.CrossRef Goossens, H., & Opstal, A. (1997). Human eye–head coordination in two dimensions under different sensorimotor conditions. Experimental Brain Research, 114(3), 542–560.CrossRef
go back to reference Guitton, D., & Volle, M. (1987). Gaze control in humans: Eye–head coordination during orienting movements to targets within and beyond the oculomotor range. Journal of Neurophysiology, 58(3), 427–459. Guitton, D., & Volle, M. (1987). Gaze control in humans: Eye–head coordination during orienting movements to targets within and beyond the oculomotor range. Journal of Neurophysiology, 58(3), 427–459.
go back to reference Hartley, R., & Zisserman, A. (2000). Multiple view geometry in computer vision (Vol. 2). New York: Cambridge University Press.MATH Hartley, R., & Zisserman, A. (2000). Multiple view geometry in computer vision (Vol. 2). New York: Cambridge University Press.MATH
go back to reference He, H., Ge, S., & Zhang, Z. (2011). Visual attention prediction using saliency determination of scene understanding for social robots. Special issue on towards an effective design of social robots. International Journal of Social Robotics, 3, 457–468.CrossRef He, H., Ge, S., & Zhang, Z. (2011). Visual attention prediction using saliency determination of scene understanding for social robots. Special issue on towards an effective design of social robots. International Journal of Social Robotics, 3, 457–468.CrossRef
go back to reference He, H., Zhang, Z., & Ge, S. (2010). Attention determination for social robots using salient region detection. In International conference on social robotics (pp. 295–304). Heidelberg: Springer. He, H., Zhang, Z., & Ge, S. (2010). Attention determination for social robots using salient region detection. In International conference on social robotics (pp. 295–304). Heidelberg: Springer.
go back to reference Heuring, J., & Murray, D. (1999). Modeling and copying human head movements. IEEE Transactions on Robotics and Automation, 15(6), 1095–1108.CrossRef Heuring, J., & Murray, D. (1999). Modeling and copying human head movements. IEEE Transactions on Robotics and Automation, 15(6), 1095–1108.CrossRef
go back to reference Hwang, A. D., Higgins, E. C., & Pomplun, M. (2009). A model of top-down attentional control during visual search in complex scenes. Journal of Vision, 9(5), 25.1–25.18. Hwang, A. D., Higgins, E. C., & Pomplun, M. (2009). A model of top-down attentional control during visual search in complex scenes. Journal of Vision, 9(5), 25.1–25.18.
go back to reference Itti, L. (2003). Realistic avatar eye and head animation using a neurobiological model of visual attention. Tech. Rep. Defense Technical Information Center Document. Itti, L. (2003). Realistic avatar eye and head animation using a neurobiological model of visual attention. Tech. Rep. Defense Technical Information Center Document.
go back to reference Itti, L. (2005). Models of bottom-up attention and saliency. In L. Itti, G. Rees, & J. K. Tsotsos (Eds.), Neurobiology of attention (pp. 576–582). San Diego, CA: Elsevier. Itti, L. (2005). Models of bottom-up attention and saliency. In L. Itti, G. Rees, & J. K. Tsotsos (Eds.), Neurobiology of attention (pp. 576–582). San Diego, CA: Elsevier.
go back to reference Judd, T., Ehinger, K., Durand, F., & Torralba, A. (2010). Learning to predict where humans look. In International conference on computer vision. Judd, T., Ehinger, K., Durand, F., & Torralba, A. (2010). Learning to predict where humans look. In International conference on computer vision.
go back to reference Kanan, C., Tong, M. H., Zhang, L., & Cottrell, G. W. (2009). Sun: Top-down saliency using natural statistics. Visual Cognition, 17(6–7), 979–1003.CrossRef Kanan, C., Tong, M. H., Zhang, L., & Cottrell, G. W. (2009). Sun: Top-down saliency using natural statistics. Visual Cognition, 17(6–7), 979–1003.CrossRef
go back to reference Laschi, C., Asuni, G., Guglielmelli, E., Teti, G., Johansson, R., Konosu, H., et al. (2008). A bio-inspired predictive sensory-motor coordination scheme for robot reaching and preshaping. Autonomous Robots, 25(1), 85–101.CrossRef Laschi, C., Asuni, G., Guglielmelli, E., Teti, G., Johansson, R., Konosu, H., et al. (2008). A bio-inspired predictive sensory-motor coordination scheme for robot reaching and preshaping. Autonomous Robots, 25(1), 85–101.CrossRef
go back to reference Le Meur, O., Le Callet, P., Barba, D., & Thoreau, D. (2006). A coherent computational approach to model bottom-up visual attention. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(5), 802–817. Le Meur, O., Le Callet, P., Barba, D., & Thoreau, D. (2006). A coherent computational approach to model bottom-up visual attention. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(5), 802–817.
go back to reference Lopes, M., Bernardino, A., Santos-Victor, J., Rosander, K., & von Hofsten, C. (2009). Biomimetic eye-neck coordination. In Proceedings of IEEE international conference on development and learning (pp. 1–8). Lopes, M., Bernardino, A., Santos-Victor, J., Rosander, K., & von Hofsten, C. (2009). Biomimetic eye-neck coordination. In Proceedings of IEEE international conference on development and learning (pp. 1–8).
go back to reference Maini, E., Manfredi, L., Laschi, C., & Dario, P. (2008). Bioinspired velocity control of fast gaze shifts on a robotic anthropomorphic head. Autonomous Robots, 25(1), 37–58.CrossRef Maini, E., Manfredi, L., Laschi, C., & Dario, P. (2008). Bioinspired velocity control of fast gaze shifts on a robotic anthropomorphic head. Autonomous Robots, 25(1), 37–58.CrossRef
go back to reference Medendorp, W., Van Gisbergen, J., Horstink, M., & Gielen, C. (1999). Donders’ law in torticollis. Journal of Neurophysiology, 82(5), 2833. Medendorp, W., Van Gisbergen, J., Horstink, M., & Gielen, C. (1999). Donders’ law in torticollis. Journal of Neurophysiology, 82(5), 2833.
go back to reference Milanese, R., Wechsler, H., Gill, S., Bost, J.-M., & Pun, T. (1994). Integration of bottom-up and top-down cues for visual attention using non-linear relaxation. In IEEE Computer Society conference on computer vision and pattern recognition, Proceedings CVPR’94 (pp. 781–785). Milanese, R., Wechsler, H., Gill, S., Bost, J.-M., & Pun, T. (1994). Integration of bottom-up and top-down cues for visual attention using non-linear relaxation. In IEEE Computer Society conference on computer vision and pattern recognition, Proceedings CVPR’94 (pp. 781–785).
go back to reference Morel, J., & Yu, G. (2009). Asift: A new framework for fully affine invariant image comparison. SIAM Journal on Imaging Sciences, 2(2), 438–469.CrossRefMATHMathSciNet Morel, J., & Yu, G. (2009). Asift: A new framework for fully affine invariant image comparison. SIAM Journal on Imaging Sciences, 2(2), 438–469.CrossRefMATHMathSciNet
go back to reference Nagai, Y., Hosoda, K., Morita, A., & Asada, M. (2003). A constructive model for the development of joint attention. Connection Science, 15(4), 211–229.CrossRef Nagai, Y., Hosoda, K., Morita, A., & Asada, M. (2003). A constructive model for the development of joint attention. Connection Science, 15(4), 211–229.CrossRef
go back to reference Navalpakkam, V., & Itti, L. (2006). An integrated model of top-down and bottom-up attention for optimizing detection speed. In 2006 IEEE Computer Society conference on computer vision and pattern recognition (Vol. 2, pp. 2049–2056). Navalpakkam, V., & Itti, L. (2006). An integrated model of top-down and bottom-up attention for optimizing detection speed. In 2006 IEEE Computer Society conference on computer vision and pattern recognition (Vol. 2, pp. 2049–2056).
go back to reference Oliva, A., Torralba, A., Castelhano, M. S., & Henderson, J. M. (2003). Top-down control of visual attention in object detection. In Proceedings of 2003 IEEE international conference on image processing, ICIP 2003 (Vol. 1, pp. 1–253). Oliva, A., Torralba, A., Castelhano, M. S., & Henderson, J. M. (2003). Top-down control of visual attention in object detection. In Proceedings of 2003 IEEE international conference on image processing, ICIP 2003 (Vol. 1, pp. 1–253).
go back to reference Pagel, M., Maël, E., & Von Der Malsburg, C. (1998). Self calibration of the fixation movement of a stereo camera head. Autonomous Robots, 5(3), 355–367.CrossRef Pagel, M., Maël, E., & Von Der Malsburg, C. (1998). Self calibration of the fixation movement of a stereo camera head. Autonomous Robots, 5(3), 355–367.CrossRef
go back to reference Raphan, T. (1998). Modeling control of eye orientation in three dimensions. I. Role of muscle pulleys in determining saccadic trajectory. Journal of Neurophysiology, 79(5), 2653. Raphan, T. (1998). Modeling control of eye orientation in three dimensions. I. Role of muscle pulleys in determining saccadic trajectory. Journal of Neurophysiology, 79(5), 2653.
go back to reference Seo, H. J., & Milanfar, P. (2009). Nonparametric bottom-up saliency detection by self-resemblance. In IEEE computer society conference on computer vision and pattern recognition workshops. CVPR Workshops 2009 (pp. 45–52). Seo, H. J., & Milanfar, P. (2009). Nonparametric bottom-up saliency detection by self-resemblance. In IEEE computer society conference on computer vision and pattern recognition workshops. CVPR Workshops 2009 (pp. 45–52).
go back to reference Smith, R. (2007). An overview of the tesseract ocr engine. In Proceedings of the ninth international conference on document analysis and recognition. Smith, R. (2007). An overview of the tesseract ocr engine. In Proceedings of the ninth international conference on document analysis and recognition.
go back to reference Tsagarakis, N., Metta, G., Sandini, G., Vernon, D., Beira, R., Becchi, F., et al. (2007). Icub: the design and realization of an open humanoid platform for cognitive and neuroscience research. Advanced Robotics, 21(10), 1151–1175.CrossRef Tsagarakis, N., Metta, G., Sandini, G., Vernon, D., Beira, R., Becchi, F., et al. (2007). Icub: the design and realization of an open humanoid platform for cognitive and neuroscience research. Advanced Robotics, 21(10), 1151–1175.CrossRef
go back to reference Tweed, D. (1997). Three-dimensional model of the human eye–head saccadic system. Journal of Neurophysiology, 77(2), 654. Tweed, D. (1997). Three-dimensional model of the human eye–head saccadic system. Journal of Neurophysiology, 77(2), 654.
go back to reference Viola, P., & Jones, M. (2004). Robust real-time face detection. International Journal of Computer Vision, 57(2), 137–154.CrossRef Viola, P., & Jones, M. (2004). Robust real-time face detection. International Journal of Computer Vision, 57(2), 137–154.CrossRef
go back to reference Westheimer, G. (1957). Kinematics of the eye. Journal of the Optical Society of America, 47, 967–974.CrossRef Westheimer, G. (1957). Kinematics of the eye. Journal of the Optical Society of America, 47, 967–974.CrossRef
Metadata
Title
A saliency-driven robotic head with bio-inspired saccadic behaviors for social robotics
Authors
Hongsheng He
Shuzhi Sam Ge
Zhengchen Zhang
Publication date
01-03-2014
Publisher
Springer US
Published in
Autonomous Robots / Issue 3/2014
Print ISSN: 0929-5593
Electronic ISSN: 1573-7527
DOI
https://doi.org/10.1007/s10514-013-9346-z

Other articles of this Issue 3/2014

Autonomous Robots 3/2014 Go to the issue