Skip to main content

2018 | OriginalPaper | Buchkapitel

Frame-Level Covariance Descriptor for Action Recognition

verfasst von : Wilson Moreno, Gustavo Garzón, Fabio Martínez

Erschienen in: Advances in Computing

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Activity recognition is a fundamental task in areas such as video-surveillance, gesture recognition, robotics, multimedia applications among much others. Such task remains as an open problem because the variability of many factors such as the appearance of actors, illumination changes in real scenarios and the dynamic developed for each action. Despite favorable results in recent works for several academic datasets, the proposed methodologies require a huge number of training samples and the output descriptor result in a high dimensional array that difficult the implementation in real conditions. This work proposes a spatio-temporal descriptor that model human activities by using a fast regional covariance representation for each frame. At each frame, a set of motion and geometrical map measures are quantified into a pyramidal regional structure to describe the instantaneous action. Such low-level primitive maps are codified into a integral covariance that allows a fast and compact description of local correlation among features. The set of pyramidal-frame-covariances along the video sequence represent a manifold that coexist in a positive Riemannian space. Then, a set of means are approximated in Riemannian space for each regional covariance sequence to represent a very compact action descriptor. The proposed action descriptor is mapped to a Euclidean space to perform an automatic classification using a Support vector Machine. The proposed approach was evaluated in two different public datasets: (1) in UT-Interaction with a k-fold cross-validation scheme was achieved a 70.8% of accuracy with a descriptor size of just 10 features per video sequence and (2) in UCF Sports achieve an accuracy of 71.7% using 13 features.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Brox, T., Malik, J.: Large displacement optical flow: descriptor matching in variational motion estimation. IEEE Trans. Pattern Anal. Mach. Intell. 33(3), 500–513 (2011)CrossRef Brox, T., Malik, J.: Large displacement optical flow: descriptor matching in variational motion estimation. IEEE Trans. Pattern Anal. Mach. Intell. 33(3), 500–513 (2011)CrossRef
2.
Zurück zum Zitat Cao, X., Zhang, H., Deng, C., Liu, Q., Liu, H.: Action recognition using 3D DAISY descriptor. Mach. Vis. Appl. 25(1), 159–171 (2014)CrossRef Cao, X., Zhang, H., Deng, C., Liu, Q., Liu, H.: Action recognition using 3D DAISY descriptor. Mach. Vis. Appl. 25(1), 159–171 (2014)CrossRef
3.
Zurück zum Zitat Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2(3), 27 (2011) Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2(3), 27 (2011)
4.
Zurück zum Zitat Cherla, S., Kulkarni, K., Kale, A., Ramasubramanian, V.: Towards fast, view-invariant human action recognition. In: 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2008, pp. 1–8. IEEE (2008) Cherla, S., Kulkarni, K., Kale, A., Ramasubramanian, V.: Towards fast, view-invariant human action recognition. In: 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2008, pp. 1–8. IEEE (2008)
5.
Zurück zum Zitat Fletcher, P.T., Joshi, S.: Riemannian geometry for the statistical analysis of diffusion tensor data. Sig. Process. 87(2), 250–262 (2007)CrossRef Fletcher, P.T., Joshi, S.: Riemannian geometry for the statistical analysis of diffusion tensor data. Sig. Process. 87(2), 250–262 (2007)CrossRef
6.
Zurück zum Zitat Geronimo, D., Lopez, A.M., Sappa, A.D., Graf, T.: Survey of pedestrian detection for advanced driver assistance systems. IEEE Trans. Pattern Anal. Mach. Intell. 32(7), 1239–1258 (2010)CrossRef Geronimo, D., Lopez, A.M., Sappa, A.D., Graf, T.: Survey of pedestrian detection for advanced driver assistance systems. IEEE Trans. Pattern Anal. Mach. Intell. 32(7), 1239–1258 (2010)CrossRef
7.
Zurück zum Zitat Gowayyed, M.A., Torki, M., Hussein, M.E., El-Saban, M.: Histogram of oriented displacements (HOD): describing trajectories of human joints for action recognition. In: IJCAI (2013) Gowayyed, M.A., Torki, M., Hussein, M.E., El-Saban, M.: Histogram of oriented displacements (HOD): describing trajectories of human joints for action recognition. In: IJCAI (2013)
8.
Zurück zum Zitat Ji, X., Wang, C., Zuo, X., Wang, Y.: Multiple feature voting based human interaction recognition. Int. J. Sig. Process. Image Process. Pattern Recognit. 9(1), 323–334 (2016) Ji, X., Wang, C., Zuo, X., Wang, Y.: Multiple feature voting based human interaction recognition. Int. J. Sig. Process. Image Process. Pattern Recognit. 9(1), 323–334 (2016)
9.
Zurück zum Zitat Laptev, I.: On space-time interest points. Int. J. Comput. Vis. 64(2–3), 107–123 (2005)CrossRef Laptev, I.: On space-time interest points. Int. J. Comput. Vis. 64(2–3), 107–123 (2005)CrossRef
10.
Zurück zum Zitat Laptev, I., Caputo, B., Schüldt, C., Lindeberg, T.: Local velocity-adapted motion events for spatio-temporal recognition. Comput. Vis. Image Underst. 108(3), 207–229 (2007)CrossRef Laptev, I., Caputo, B., Schüldt, C., Lindeberg, T.: Local velocity-adapted motion events for spatio-temporal recognition. Comput. Vis. Image Underst. 108(3), 207–229 (2007)CrossRef
11.
Zurück zum Zitat Liu, A.A., Xu, N., Su, Y.T., Lin, H., Hao, T., Yang, Z.X.: Single/multi-view human action recognition via regularized multi-task learning. Neurocomputing 151, 544–553 (2015)CrossRef Liu, A.A., Xu, N., Su, Y.T., Lin, H., Hao, T., Yang, Z.X.: Single/multi-view human action recognition via regularized multi-task learning. Neurocomputing 151, 544–553 (2015)CrossRef
12.
Zurück zum Zitat Ma, B., Su, Y., Jurie, F.: Covariance descriptor based on bio-inspired features for person re-identification and face verification. Image Vis. Comput. 32(6), 379–390 (2014)CrossRef Ma, B., Su, Y., Jurie, F.: Covariance descriptor based on bio-inspired features for person re-identification and face verification. Image Vis. Comput. 32(6), 379–390 (2014)CrossRef
13.
Zurück zum Zitat Pennec, X., Fillard, P., Ayache, N.: A Riemannian framework for tensor computing. Int. J. Comput. Vis. 66(1), 41–66 (2006)CrossRef Pennec, X., Fillard, P., Ayache, N.: A Riemannian framework for tensor computing. Int. J. Comput. Vis. 66(1), 41–66 (2006)CrossRef
14.
Zurück zum Zitat Poppe, R.: A survey on vision-based human action recognition. Image Vis. Comput. 28(6), 976–990 (2010)CrossRef Poppe, R.: A survey on vision-based human action recognition. Image Vis. Comput. 28(6), 976–990 (2010)CrossRef
15.
Zurück zum Zitat Robertson, N., Reid, I.: A general method for human activity recognition in video. Comput. Vis. Image Underst. 104(2), 232–248 (2006)CrossRef Robertson, N., Reid, I.: A general method for human activity recognition in video. Comput. Vis. Image Underst. 104(2), 232–248 (2006)CrossRef
17.
Zurück zum Zitat Nour el houda Slimani, K., Benezeth, Y., Souami, F.: Human interaction recognition based on the co-occurence of visual words. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 455–460 (2014) Nour el houda Slimani, K., Benezeth, Y., Souami, F.: Human interaction recognition based on the co-occurence of visual words. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 455–460 (2014)
19.
Zurück zum Zitat Souvenir, R., Babbs, J.: Learning the viewpoint manifold for action recognition. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2008, pp. 1–7. IEEE (2008) Souvenir, R., Babbs, J.: Learning the viewpoint manifold for action recognition. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2008, pp. 1–7. IEEE (2008)
21.
Zurück zum Zitat Wang, Y., Huang, K., Tan, T.: Human activity recognition based on R transform. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2007, pp. 1–8. IEEE (2007) Wang, Y., Huang, K., Tan, T.: Human activity recognition based on R transform. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2007, pp. 1–8. IEEE (2007)
Metadaten
Titel
Frame-Level Covariance Descriptor for Action Recognition
verfasst von
Wilson Moreno
Gustavo Garzón
Fabio Martínez
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-98998-3_22

Premium Partner