Skip to main content
Erschienen in: Autonomous Robots 6/2018

27.12.2017

Skeleton-based bio-inspired human activity prediction for real-time human–robot interaction

verfasst von: Brian Reily, Fei Han, Lynne E. Parker, Hao Zhang

Erschienen in: Autonomous Robots | Ausgabe 6/2018

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Activity prediction is an essential task in practical human-centered robotics applications, such as security, assisted living, etc., which is targeted at inferring ongoing human activities based on incomplete observations. To address this challenging problem, we introduce a novel bio-inspired predictive orientation decomposition (BIPOD) approach to construct representations of people from 3D skeleton trajectories. BIPOD is invariant to scales and viewpoints, runs in real-time on basic computer systems, and is able to recognize and predict activities in an online fashion. Our approach is inspired by biological research in human anatomy. To capture spatio-temporal information of human motions, we spatially decompose 3D human skeleton trajectories and project them onto three anatomical planes (i.e., coronal, transverse and sagittal planes); then, we describe short-term time information of joint motions and encode high-order temporal dependencies. By using Extended Kalman Filters to estimate future skeleton trajectories, we endow our BIPOD representation with the critical capabilities to reduce noisy skeleton observation data and predict the ongoing activities. Experiments on benchmark datasets have shown that our BIPOD representation significantly outperforms previous methods for real-time human activity classification and prediction from 3D skeleton trajectories. Empirical studies using TurtleBot2 and Baxter humanoid robots have also validated that our BIPOD method obtains promising performance, in terms of both accuracy and efficiency, making BIPOD a fast, simple, yet powerful representation for low-latency online activity prediction in human–robot interaction applications.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Aggarwal, J., & Xia, L. (2014). Human activity recognition from 3D data: A review. Pattern Recognition Letters, 48, 70–80.CrossRef Aggarwal, J., & Xia, L. (2014). Human activity recognition from 3D data: A review. Pattern Recognition Letters, 48, 70–80.CrossRef
Zurück zum Zitat Akgun, B., Cakmak, M., Jiang, K., & Thomaz, A. (2012). Keyframe-based learning from demonstration. Internation Journal of Social Robotics, 4(4), 343–355.CrossRef Akgun, B., Cakmak, M., Jiang, K., & Thomaz, A. (2012). Keyframe-based learning from demonstration. Internation Journal of Social Robotics, 4(4), 343–355.CrossRef
Zurück zum Zitat Berndt, H., Emmert, J., & Dietmayer, K. (2008). Continuous driver intention recognition with hidden Markov models. In Intelligent Transportation Systems (pp. 1189–1194). Berndt, H., Emmert, J., & Dietmayer, K. (2008). Continuous driver intention recognition with hidden Markov models. In Intelligent Transportation Systems (pp. 1189–1194).
Zurück zum Zitat Bi, L., Yang, X., & Wang, C. (2013). Inferring driver intentions using a driver model based on queuing network. In Intelligent Vehicles Symposium (pp. 1387–1391). Bi, L., Yang, X., & Wang, C. (2013). Inferring driver intentions using a driver model based on queuing network. In Intelligent Vehicles Symposium (pp. 1387–1391).
Zurück zum Zitat Bosurgi, G., D’Andrea, A., & Pellegrino, O. (2014). Prediction of drivers’ visual strategy using an analytical model. Journal of Transportation Safety & Security, 7, 153–173.CrossRef Bosurgi, G., D’Andrea, A., & Pellegrino, O. (2014). Prediction of drivers’ visual strategy using an analytical model. Journal of Transportation Safety & Security, 7, 153–173.CrossRef
Zurück zum Zitat Boubou, S., & Suzuki, E. (2015). Classifying actions based on histogram of oriented velocity vectors. Journal of Intelligent Information Systems, 44(1), 49–65.CrossRef Boubou, S., & Suzuki, E. (2015). Classifying actions based on histogram of oriented velocity vectors. Journal of Intelligent Information Systems, 44(1), 49–65.CrossRef
Zurück zum Zitat Boussemart, Y., & Cummings, M. L. (2011). Predictive models of human supervisory control behavioral patterns using hidden semi-Markov models. Engineering Applications of Artifical Intelligence, 24, 1252–1262.CrossRef Boussemart, Y., & Cummings, M. L. (2011). Predictive models of human supervisory control behavioral patterns using hidden semi-Markov models. Engineering Applications of Artifical Intelligence, 24, 1252–1262.CrossRef
Zurück zum Zitat Chang, C. C., & Lin, C. J. (2011). LIBSVM: A library for support vector machines. ACM Transaction on Intelligent Systems and Technology, 2, 27:1–27:27. Chang, C. C., & Lin, C. J. (2011). LIBSVM: A library for support vector machines. ACM Transaction on Intelligent Systems and Technology, 2, 27:1–27:27.
Zurück zum Zitat Charles, J., Everingham, M. (2011). Learning shape models for monocular human pose estimation from the Microsoft Xbox Kinect. In IEEE international conference on computer vision. Charles, J., Everingham, M. (2011). Learning shape models for monocular human pose estimation from the Microsoft Xbox Kinect. In IEEE international conference on computer vision.
Zurück zum Zitat Chaudhry, R., Ofli, F., Kurillo, G., Bajcsy, R., & Vidal, R. (2013). Bio-inspired dynamic 3D discriminative skeletal features for human action recognition. In IEEE conference on computer vision and pattern recognition workshop. Chaudhry, R., Ofli, F., Kurillo, G., Bajcsy, R., & Vidal, R. (2013). Bio-inspired dynamic 3D discriminative skeletal features for human action recognition. In IEEE conference on computer vision and pattern recognition workshop.
Zurück zum Zitat Chen, G., Giuliani, M., Clarke, D., Gaschler, A., & Knoll, A. (2014). Action recognition using ensemble weighted multi-instance learning. In IEEE international conference on robotics and automation. Chen, G., Giuliani, M., Clarke, D., Gaschler, A., & Knoll, A. (2014). Action recognition using ensemble weighted multi-instance learning. In IEEE international conference on robotics and automation.
Zurück zum Zitat Dai, F., Zhang, J., & Lu, T. (2011). The study of driver’s starting intentions. In Mechanic Automation and Control Engineering (pp. 2758–2761). Dai, F., Zhang, J., & Lu, T. (2011). The study of driver’s starting intentions. In Mechanic Automation and Control Engineering (pp. 2758–2761).
Zurück zum Zitat Du, Y., Wang, W., & Wang, L. (2015). Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1110–1118). Du, Y., Wang, W., & Wang, L. (2015). Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1110–1118).
Zurück zum Zitat Einicke, G., & White, L. (1999). Robust extended Kalman filtering. IEEE Transactions on Signal Processing, 47(9), 2596–2599.CrossRefMATH Einicke, G., & White, L. (1999). Robust extended Kalman filtering. IEEE Transactions on Signal Processing, 47(9), 2596–2599.CrossRefMATH
Zurück zum Zitat Ellis, C., Masood, S. Z., Tappen, M. F., Laviola, J. J, Jr., & Sukthankar, R. (2013). Exploring the trade-off between accuracy and observational latency in action recognition. International Journal of Computer Vision, 101(3), 420–436.CrossRef Ellis, C., Masood, S. Z., Tappen, M. F., Laviola, J. J, Jr., & Sukthankar, R. (2013). Exploring the trade-off between accuracy and observational latency in action recognition. International Journal of Computer Vision, 101(3), 420–436.CrossRef
Zurück zum Zitat Ganapathi, V., Plagemann, C., Koller, D., & Thrun, S. (2010). Real time motion capture using a single time-of-flight camera. In IEEE conference on computer vision and pattern recognition. Ganapathi, V., Plagemann, C., Koller, D., & Thrun, S. (2010). Real time motion capture using a single time-of-flight camera. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Georgiou, T., & Demiris, Y. (2015). Predicting car states through learned models of vehicle dynamics and user behaviours. In Intelligent vehicles symposium (pp. 1240–1245). Georgiou, T., & Demiris, Y. (2015). Predicting car states through learned models of vehicle dynamics and user behaviours. In Intelligent vehicles symposium (pp. 1240–1245).
Zurück zum Zitat Girshick, R., Shotton, J., Kohli, P., Criminisi, A., & Fitzgibbon, A. (2011). Efficient regression of general-activity human poses from depth images. In IEEE international conference on computer vision. Girshick, R., Shotton, J., Kohli, P., Criminisi, A., & Fitzgibbon, A. (2011). Efficient regression of general-activity human poses from depth images. In IEEE international conference on computer vision.
Zurück zum Zitat Gowayyed, M. A., Torki, M., Hussein, M. E., & El-Saban, M. (2013). Histogram of oriented displacements (HOD): Describing trajectories of human joints for action recognition. In International joint conference on artificial intelligence. Gowayyed, M. A., Torki, M., Hussein, M. E., & El-Saban, M. (2013). Histogram of oriented displacements (HOD): Describing trajectories of human joints for action recognition. In International joint conference on artificial intelligence.
Zurück zum Zitat Gray, H. (1973). Anatomy of the human body. Philadelphia: Lea & Febiger. Gray, H. (1973). Anatomy of the human body. Philadelphia: Lea & Febiger.
Zurück zum Zitat Han, F., Reily, B., Hoff, W., & Zhang, H. (2016). Space-time representation of people based on 3D skeletal data: A review. ArXiv e-prints 1601.01006. Han, F., Reily, B., Hoff, W., & Zhang, H. (2016). Space-time representation of people based on 3D skeletal data: A review. ArXiv e-prints 1601.01006.
Zurück zum Zitat Han, F., Reily, B., Hoff, W., & Zhang, H. (2017). Space-time representation of people based on 3d skeletal data: A review. Computer Vision and Image Understanding, 158, 85–105.CrossRef Han, F., Reily, B., Hoff, W., & Zhang, H. (2017). Space-time representation of people based on 3d skeletal data: A review. Computer Vision and Image Understanding, 158, 85–105.CrossRef
Zurück zum Zitat Harandi, M., Sanderson, C., Hartley, R., & Lovell, B. (2012). Sparse coding and dictionary learning for symmetric positive definite matrices: A kernel approach. Computer Vision-ECCV, 2012, 216–229. Harandi, M., Sanderson, C., Hartley, R., & Lovell, B. (2012). Sparse coding and dictionary learning for symmetric positive definite matrices: A kernel approach. Computer Vision-ECCV, 2012, 216–229.
Zurück zum Zitat He, L., Cf, Zong, & Wang, C. (2012). Driving intention recognition and behaviour prediction based on a double-layer hidden Markov model. Journal of Zhejiang University, 13, 208–217.CrossRef He, L., Cf, Zong, & Wang, C. (2012). Driving intention recognition and behaviour prediction based on a double-layer hidden Markov model. Journal of Zhejiang University, 13, 208–217.CrossRef
Zurück zum Zitat Hoai, M., & De la Torre, F. (2014). Max-margin early event detectors. International Journal of Computer Vision, 107(2), 191–202.MathSciNetCrossRef Hoai, M., & De la Torre, F. (2014). Max-margin early event detectors. International Journal of Computer Vision, 107(2), 191–202.MathSciNetCrossRef
Zurück zum Zitat Hoare, J., & Parker, L. (2010). Using on-line conditional random fields to determine human intent for peer-to-peer human robot teaming. In IEEE/RSJ international conference on intelligent robots and systems. Hoare, J., & Parker, L. (2010). Using on-line conditional random fields to determine human intent for peer-to-peer human robot teaming. In IEEE/RSJ international conference on intelligent robots and systems.
Zurück zum Zitat Hussein, M. E., Torki, M., Gowayyed, M. A., & El-Saban, M. (2013). Human action recognition using a temporal hierarchy of covariance descriptors on 3D joint locations. In International joint conference on artificial intelligence. Hussein, M. E., Torki, M., Gowayyed, M. A., & El-Saban, M. (2013). Human action recognition using a temporal hierarchy of covariance descriptors on 3D joint locations. In International joint conference on artificial intelligence.
Zurück zum Zitat Jin, L., Hou, H., & Jiang, Y. (2011). Driver intention recognition based on continuous hidden Markov model. In Transportation, Mechanical, and Electrical Engineering (pp. 739–742). Jin, L., Hou, H., & Jiang, Y. (2011). Driver intention recognition based on continuous hidden Markov model. In Transportation, Mechanical, and Electrical Engineering (pp. 739–742).
Zurück zum Zitat Jung, H. Y., Lee, S., Heo, Y. S., & Yun, I. D. (2015). Random tree walk toward instantaneous 3D human pose estimation. In IEEE conference on computer vision and pattern recognition. Jung, H. Y., Lee, S., Heo, Y. S., & Yun, I. D. (2015). Random tree walk toward instantaneous 3D human pose estimation. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Kim, Y., Chen, J., Chang, M. C., Wang, X., Provost, E. M., & Lyu, S. (2015). Modeling transition patterns between events for temporal human action segmentation and classification. In IEEE international conference and workshops on automatic face and gesture recognition (FG), Ljubljana (pp. 1–8). Kim, Y., Chen, J., Chang, M. C., Wang, X., Provost, E. M., & Lyu, S. (2015). Modeling transition patterns between events for temporal human action segmentation and classification. In IEEE international conference and workshops on automatic face and gesture recognition (FG), Ljubljana (pp. 1–8).
Zurück zum Zitat Koppula, H. S., Rudhir, G., & Saxena, A. (2013). Learning human activities and object affordances from RGB-D videos. The International Journal of Robotics Research, 32, 951–970.CrossRef Koppula, H. S., Rudhir, G., & Saxena, A. (2013). Learning human activities and object affordances from RGB-D videos. The International Journal of Robotics Research, 32, 951–970.CrossRef
Zurück zum Zitat Li, K., & Fu, Y. (2014). Prediction of human activity by discovering temporal sequence patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36, 1644–1657.CrossRef Li, K., & Fu, Y. (2014). Prediction of human activity by discovering temporal sequence patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36, 1644–1657.CrossRef
Zurück zum Zitat Li, K., Hu, J., & Fu, Y. (2012). Modeling complex temporal composition of actionlets for activity prediction. In European conference on computer vision. Li, K., Hu, J., & Fu, Y. (2012). Modeling complex temporal composition of actionlets for activity prediction. In European conference on computer vision.
Zurück zum Zitat Liu, Q., & Cao, X. (2012). Action recognition using subtensor constraint. In European conference on computer vision. Liu, Q., & Cao, X. (2012). Action recognition using subtensor constraint. In European conference on computer vision.
Zurück zum Zitat López-Mendez, A., Gall, J., Casas, J. R., & Gool, L. J. V. (2012). Metric learning from poses for temporal clustering of human motion. In British machine vision conference. López-Mendez, A., Gall, J., Casas, J. R., & Gool, L. J. V. (2012). Metric learning from poses for temporal clustering of human motion. In British machine vision conference.
Zurück zum Zitat Luo, J., Wang, W., & Qi, H. (2013). Group sparsity and geometry constrained dictionary learning for action recognition from depth maps. In IEEE international conference on computer vision. Luo, J., Wang, W., & Qi, H. (2013). Group sparsity and geometry constrained dictionary learning for action recognition from depth maps. In IEEE international conference on computer vision.
Zurück zum Zitat Mandel, J. (1982). Use of the singular value decomposition in regression analysis. The American Statistician, 36(1), 15–24. Mandel, J. (1982). Use of the singular value decomposition in regression analysis. The American Statistician, 36(1), 15–24.
Zurück zum Zitat McGinnis, M. (1999). Bioregionalism: The tug and pull of place. London: Routledge. McGinnis, M. (1999). Bioregionalism: The tug and pull of place. London: Routledge.
Zurück zum Zitat Meiring, G. A. M., & Myburgh, H. C. (2015). A review of intelligent driving style analysis systems and related artificial intelligence algorithms. Sensors, 15, 30653–30682.CrossRef Meiring, G. A. M., & Myburgh, H. C. (2015). A review of intelligent driving style analysis systems and related artificial intelligence algorithms. Sensors, 15, 30653–30682.CrossRef
Zurück zum Zitat Mori, A., Uchida, S., Kurazume, R., Taniguchi, R. I., Hasegawa, T., & Sakoe, H. (2006). Early recognition and prediction of gestures. In International conference on pattern recognition. Mori, A., Uchida, S., Kurazume, R., Taniguchi, R. I., Hasegawa, T., & Sakoe, H. (2006). Early recognition and prediction of gestures. In International conference on pattern recognition.
Zurück zum Zitat Müller, M., Röder, T., Clausen, M., Eberhardt, B., Krüger, B., & Weber, A. (2007). Documentation mocap database HDM05. Technical report, Universität Bonn. Müller, M., Röder, T., Clausen, M., Eberhardt, B., Krüger, B., & Weber, A. (2007). Documentation mocap database HDM05. Technical report, Universität Bonn.
Zurück zum Zitat Niebles, J. C., & Fei-Fei, L. (2007). A hierarchical model of shape and appearance for human action classification. In IEEE conference on computer vision and pattern recognition. Niebles, J. C., & Fei-Fei, L. (2007). A hierarchical model of shape and appearance for human action classification. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Nikolaidis, S., Hsu, D., & Srinivasa, S. (2017). Human-robot mutual adaptation in collaborative tasks: Models and experiments. The International Journal of Robotics Research, 36(5–7), 618–634.CrossRef Nikolaidis, S., Hsu, D., & Srinivasa, S. (2017). Human-robot mutual adaptation in collaborative tasks: Models and experiments. The International Journal of Robotics Research, 36(5–7), 618–634.CrossRef
Zurück zum Zitat Ofli, F., Chaudhry, R., Kurillo, G., Vidal, R., & Bajcsy, R. (2014). Sequence of the most informative joints (SMIJ): A new representation for human skeletal action recognition. Journal of Visual Communication and Image Representation, 25(1), 24–38.CrossRef Ofli, F., Chaudhry, R., Kurillo, G., Vidal, R., & Bajcsy, R. (2014). Sequence of the most informative joints (SMIJ): A new representation for human skeletal action recognition. Journal of Visual Communication and Image Representation, 25(1), 24–38.CrossRef
Zurück zum Zitat Pentland, A., & Liu, A. (1999). Modeling and prediction of human behavior. Neural Computation, 11(1), 229–242.CrossRef Pentland, A., & Liu, A. (1999). Modeling and prediction of human behavior. Neural Computation, 11(1), 229–242.CrossRef
Zurück zum Zitat Perez-D’Arpino, C., & Shah, J. A. (2015). Fast target prediction of human reaching motion for cooperative human-robot manipulation tasks using time series classification. In 2015 IEEE international conference on robotics and automation (ICRA) (pp. 6175–6182). IEEE. Perez-D’Arpino, C., & Shah, J. A. (2015). Fast target prediction of human reaching motion for cooperative human-robot manipulation tasks using time series classification. In 2015 IEEE international conference on robotics and automation (ICRA) (pp. 6175–6182). IEEE.
Zurück zum Zitat Pieropan, A., Salvi, G., Pauwels, K., & Kjellstrom, H. (2014). Audio-visual classification and detection of human manipulation actions. In IEEE/RSJ international conference on intelligent robots and systems. Pieropan, A., Salvi, G., Pauwels, K., & Kjellstrom, H. (2014). Audio-visual classification and detection of human manipulation actions. In IEEE/RSJ international conference on intelligent robots and systems.
Zurück zum Zitat Plagemann, C., Ganapathi, V., Koller, D., & Thrun, S. (2010). Real-time identification and localization of body parts from depth images. In IEEE international conference on robotics and automation. Plagemann, C., Ganapathi, V., Koller, D., & Thrun, S. (2010). Real-time identification and localization of body parts from depth images. In IEEE international conference on robotics and automation.
Zurück zum Zitat Rahmani, H., Mahmood, A., Mian, A., & Huynh, D. (2014). Real time action recognition using histograms of depth gradients and random decision forests. In IEEE winter conference on applications of computer vision. Rahmani, H., Mahmood, A., Mian, A., & Huynh, D. (2014). Real time action recognition using histograms of depth gradients and random decision forests. In IEEE winter conference on applications of computer vision.
Zurück zum Zitat Ryoo, M. S. (2011). Human activity prediction: Early recognition of ongoing activities from streaming videos. In International conference on computer vision. Ryoo, M. S. (2011). Human activity prediction: Early recognition of ongoing activities from streaming videos. In International conference on computer vision.
Zurück zum Zitat Ryoo, M., Fuchs, T. J., Xia, L., Aggarwal, J. K., & Matthies, L. (2015). Robot-centric activity prediction from first-person videos: What will they do to me? In Proceedings of the tenth annual ACM/IEEE international conference on human–robot interaction (pp. 295–302). ACM. Ryoo, M., Fuchs, T. J., Xia, L., Aggarwal, J. K., & Matthies, L. (2015). Robot-centric activity prediction from first-person videos: What will they do to me? In Proceedings of the tenth annual ACM/IEEE international conference on human–robot interaction (pp. 295–302). ACM.
Zurück zum Zitat Ryoo, M. S., Grauman, K., & Aggarwal, J. K. (2010). A task-driven intelligent workspace system to provide guidance feedback. Computer Vision and Image Understanding, 114(5), 520–534.CrossRef Ryoo, M. S., Grauman, K., & Aggarwal, J. K. (2010). A task-driven intelligent workspace system to provide guidance feedback. Computer Vision and Image Understanding, 114(5), 520–534.CrossRef
Zurück zum Zitat Schwarz, L. A., Mkhitaryan, A., Mateus, D., & Navab, N. (2012). Human skeleton tracking from depth data using geodesic distances and optical flow. Image and Vision Computing, 30(3), 217–226.CrossRef Schwarz, L. A., Mkhitaryan, A., Mateus, D., & Navab, N. (2012). Human skeleton tracking from depth data using geodesic distances and optical flow. Image and Vision Computing, 30(3), 217–226.CrossRef
Zurück zum Zitat Seidenari, L., Varano, V., Berretti, S., Del Bimbo, A., & Pala, P. (2013). Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses. In IEEE conference on computer vision and pattern recognition workshops. Seidenari, L., Varano, V., Berretti, S., Del Bimbo, A., & Pala, P. (2013). Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses. In IEEE conference on computer vision and pattern recognition workshops.
Zurück zum Zitat Shotton, J., Fitzgibbon, A., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., & Blake, A. (2011). Real-time human pose recognition in parts from single depth images. In IEEE conference on computer vision and pattern recognition. Shotton, J., Fitzgibbon, A., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., & Blake, A. (2011). Real-time human pose recognition in parts from single depth images. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Sung, J., Ponce, C., Selman, B., & Saxena, A. (2012). Unstructured human activity detection from RGBD images. In IEEE international conference on robotics and automation. Sung, J., Ponce, C., Selman, B., & Saxena, A. (2012). Unstructured human activity detection from RGBD images. In IEEE international conference on robotics and automation.
Zurück zum Zitat Vantigodi, S., & Babu, R. V. (2013). Real-time human action recognition from motion capture data. In National conference on computer vision, pattern recognition, image processing and graphics. Vantigodi, S., & Babu, R. V. (2013). Real-time human action recognition from motion capture data. In National conference on computer vision, pattern recognition, image processing and graphics.
Zurück zum Zitat Vedaldi, A., & Zisserman, A. (2012). Efficient additive kernels via explicit feature maps. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(3), 480–492.CrossRef Vedaldi, A., & Zisserman, A. (2012). Efficient additive kernels via explicit feature maps. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(3), 480–492.CrossRef
Zurück zum Zitat Viola, P., & Jones, M. (2001). Rapid object detection using a boosted cascade of simple features. In IEEE conference on computer vision and pattern recognition. Viola, P., & Jones, M. (2001). Rapid object detection using a boosted cascade of simple features. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2012). Mining actionlet ensemble for action recognition with depth cameras. In IEEE conference on computer vision and pattern recognition. Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2012). Mining actionlet ensemble for action recognition with depth cameras. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2014a). Learning actionlet ensemble for 3D human action recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(5), 914–927.CrossRef Wang, J., Liu, Z., Wu, Y., & Yuan, J. (2014a). Learning actionlet ensemble for 3D human action recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(5), 914–927.CrossRef
Zurück zum Zitat Wang, Z., Boularias, A., Mulling, K., Scholkopf, B., & Peters, J. (2014c). Anticipatory action selection for human–robot table tennis. Artificial Intelligence, 247, 399–414.MathSciNetCrossRefMATH Wang, Z., Boularias, A., Mulling, K., Scholkopf, B., & Peters, J. (2014c). Anticipatory action selection for human–robot table tennis. Artificial Intelligence, 247, 399–414.MathSciNetCrossRefMATH
Zurück zum Zitat Wu, D., & Shao, L. (2014). Leveraging hierarchical parametric networks for skeletal joints action segmentation and recognition. In IEEE conference on computer vision and pattern recognition. Wu, D., & Shao, L. (2014). Leveraging hierarchical parametric networks for skeletal joints action segmentation and recognition. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Xia, L., & Aggarwal, J. K. (2013). Spatio-temporal depth cuboid similarity feature for activity recognition using depth camera. In IEEE conference on computer vision and pattern recognition. Xia, L., & Aggarwal, J. K. (2013). Spatio-temporal depth cuboid similarity feature for activity recognition using depth camera. In IEEE conference on computer vision and pattern recognition.
Zurück zum Zitat Yang, X., Tian, Y. (2012). EigenJoints-based action recognition using Naï–Bayes-Nearest-Neighbor. In IEEE conference on computer vision and pattern recognition workshop. Yang, X., Tian, Y. (2012). EigenJoints-based action recognition using Naï–Bayes-Nearest-Neighbor. In IEEE conference on computer vision and pattern recognition workshop.
Zurück zum Zitat Yang, X., & Tian, Y. (2014). Effective 3D action recognition using EigenJoints. Journal of Visual Communication and Image Representation, 25(1), 2–11.MathSciNetCrossRef Yang, X., & Tian, Y. (2014). Effective 3D action recognition using EigenJoints. Journal of Visual Communication and Image Representation, 25(1), 2–11.MathSciNetCrossRef
Zurück zum Zitat Yokochi, C., & Rohen, J. W. (2006). Color atlas of anatomy: A photographic study of the human body. Philadelphia: Lippincott Williams & Wilkins. Yokochi, C., & Rohen, J. W. (2006). Color atlas of anatomy: A photographic study of the human body. Philadelphia: Lippincott Williams & Wilkins.
Zurück zum Zitat Yu, G., Yuan, J., & Liu, Z. (2012). Predicting human activities using spatio-temporal structure of interest points. In ACM international conference on multimedia. Yu, G., Yuan, J., & Liu, Z. (2012). Predicting human activities using spatio-temporal structure of interest points. In ACM international conference on multimedia.
Zurück zum Zitat Yu, M., Liu, L., & Shao, L. (2016). Structure-preserving binary representations for RGB-D action recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(8), 1651–1664.CrossRef Yu, M., Liu, L., & Shao, L. (2016). Structure-preserving binary representations for RGB-D action recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(8), 1651–1664.CrossRef
Zurück zum Zitat Zanfir, M., Leordeanu, M., & Sminchisescu, C. (2013). The moving pose: An efficient 3D kinematics descriptor for low-latency action recognition and detection. In IEEE international conference on computer vision. Zanfir, M., Leordeanu, M., & Sminchisescu, C. (2013). The moving pose: An efficient 3D kinematics descriptor for low-latency action recognition and detection. In IEEE international conference on computer vision.
Zurück zum Zitat Zhang, H., & Parker, L. (2011). 4-dimensional local spatio-temporal features for human activity recognition. In IEEE/RSJ international conference on intelligent robots and systems. Zhang, H., & Parker, L. (2011). 4-dimensional local spatio-temporal features for human activity recognition. In IEEE/RSJ international conference on intelligent robots and systems.
Zurück zum Zitat Zhang, H., Reardon, C. M., & Parker, L. E. (2013). Real-time multiple human perception with color-depth cameras on a mobile robot. IEEE Transactions on Cybernetics, 43(5), 1429–1441.CrossRef Zhang, H., Reardon, C. M., & Parker, L. E. (2013). Real-time multiple human perception with color-depth cameras on a mobile robot. IEEE Transactions on Cybernetics, 43(5), 1429–1441.CrossRef
Zurück zum Zitat Zhao, X., Li, X., Pang, C., Zhu, X., & Sheng, Q. Z. (2013). Online human gesture recognition from motion data streams. In ACM international conference on multimedia. Zhao, X., Li, X., Pang, C., Zhu, X., & Sheng, Q. Z. (2013). Online human gesture recognition from motion data streams. In ACM international conference on multimedia.
Zurück zum Zitat Zhu, W., Lan, C., Xing, J., Zeng, W., Li, Y., Shen, L., & Xie, X. (2016). Co-occurrence feature learning for skeleton based action recognition using regularized deep LSTM networks. arXiv preprint arXiv:160307772. Zhu, W., Lan, C., Xing, J., Zeng, W., Li, Y., Shen, L., & Xie, X. (2016). Co-occurrence feature learning for skeleton based action recognition using regularized deep LSTM networks. arXiv preprint arXiv:​160307772.
Metadaten
Titel
Skeleton-based bio-inspired human activity prediction for real-time human–robot interaction
verfasst von
Brian Reily
Fei Han
Lynne E. Parker
Hao Zhang
Publikationsdatum
27.12.2017
Verlag
Springer US
Erschienen in
Autonomous Robots / Ausgabe 6/2018
Print ISSN: 0929-5593
Elektronische ISSN: 1573-7527
DOI
https://doi.org/10.1007/s10514-017-9692-3

Weitere Artikel der Ausgabe 6/2018

Autonomous Robots 6/2018 Zur Ausgabe

Neuer Inhalt