Skip to main content

2016 | OriginalPaper | Buchkapitel

Improved GLOH Approach for One-Shot Learning Human Gesture Recognition

verfasst von : Nabin Kumar Karn, Feng Jiang

Erschienen in: Biometric Recognition

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

A method is presented for One-Shot Learning Human Gesture Recognition. Shi-Tomasi corner detector and sparse optical flow are used to quickly detect and track robust key-points around motion patterns in scale space. Then Improved Gradient Location and Orientation Histogram feature descriptor is applied to capture the description of robust key interest point. All the extracted features from the training samples are clustered with the k-means algorithm to learn a visual codebook. Subsequently, simulation orthogonal matching pursuit is applied to achieve descriptor coding which map each feature into a certain visual codeword. K-NN classifier is used to recognizing the gesture. The proposed approach has been evaluated on ChaLearn gesture database.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Mitra, S., Acharya, T.: Gesture recognition: a survey. EEE Trans. Syst. Man Cybern. Part C Appl. Rev. 37(3), 311–324 (2007)CrossRef Mitra, S., Acharya, T.: Gesture recognition: a survey. EEE Trans. Syst. Man Cybern. Part C Appl. Rev. 37(3), 311–324 (2007)CrossRef
2.
Zurück zum Zitat Guyon, I., Athitsos, V., Jangyodsuk, P.: ChaLearn gesture challenge: design and first results. In: IEEE Conference on CVPR Workshops, pp. 1–6 (2012) Guyon, I., Athitsos, V., Jangyodsuk, P.: ChaLearn gesture challenge: design and first results. In: IEEE Conference on CVPR Workshops, pp. 1–6 (2012)
3.
Zurück zum Zitat Malima, A., Ozgur, E., Çetin, M.: A fast algorithm for vision-based hand gesture recognition for robot control. In: IEEE Signal Processing and Communications Applications, pp. 1–4 (2006) Malima, A., Ozgur, E., Çetin, M.: A fast algorithm for vision-based hand gesture recognition for robot control. In: IEEE Signal Processing and Communications Applications, pp. 1–4 (2006)
4.
Zurück zum Zitat Guha, T., Ward, R.K.: Learning sparse representations for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 34(8), 1576–1588 (2012)CrossRef Guha, T., Ward, R.K.: Learning sparse representations for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 34(8), 1576–1588 (2012)CrossRef
6.
Zurück zum Zitat Han, J., Shao, L., Xu, D., Shotton, J.: Enhanced computer vision with Microsoft Kinect sensor: a review. IEEE Trans. Cybern. 43(5), 1318–1334 (2013)CrossRef Han, J., Shao, L., Xu, D., Shotton, J.: Enhanced computer vision with Microsoft Kinect sensor: a review. IEEE Trans. Cybern. 43(5), 1318–1334 (2013)CrossRef
7.
Zurück zum Zitat Oikonomidis, I., Kyriazis, N., Antonis, A.: Efficient model-based 3D tracking of hand articulations using Kinect. In: Proceedings of the BMVC, pp. 1–11 (2011) Oikonomidis, I., Kyriazis, N., Antonis, A.: Efficient model-based 3D tracking of hand articulations using Kinect. In: Proceedings of the BMVC, pp. 1–11 (2011)
8.
Zurück zum Zitat Wang, J., Liu, Z., Wu, Y., Yuan, J.: Mining: actionlet ensemble for action recognition with depth cameras. In: Proceedings of the IEEE Conference on CVPR, pp. 1290–1297 (2012) Wang, J., Liu, Z., Wu, Y., Yuan, J.: Mining: actionlet ensemble for action recognition with depth cameras. In: Proceedings of the IEEE Conference on CVPR, pp. 1290–1297 (2012)
9.
Zurück zum Zitat Ren, Z., Yuan, J., Meng, J., Zhang, Z.: Robust part-based hand gesture recognition using Kinect sensor. IEEE Trans. Multimedia 15(5), 1110–1120 (2013)CrossRef Ren, Z., Yuan, J., Meng, J., Zhang, Z.: Robust part-based hand gesture recognition using Kinect sensor. IEEE Trans. Multimedia 15(5), 1110–1120 (2013)CrossRef
10.
Zurück zum Zitat Cai, Q., Gallup, D., Zhang, C., Zhang, Z.: 3D deformable face tracking with a commodity depth camera. In: Proceedings of the 11th European Conference on Computer Vision, pp. 229–242 (2010) Cai, Q., Gallup, D., Zhang, C., Zhang, Z.: 3D deformable face tracking with a commodity depth camera. In: Proceedings of the 11th European Conference on Computer Vision, pp. 229–242 (2010)
11.
Zurück zum Zitat Shotton, J., Fitzgibbon, A.W., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., Blake, A.: Real-time human pose recognition in parts from single depth images. In: Proceedings of the IEEE Conference on CVPR, pp. 1297–1304 (2011) Shotton, J., Fitzgibbon, A.W., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., Blake, A.: Real-time human pose recognition in parts from single depth images. In: Proceedings of the IEEE Conference on CVPR, pp. 1297–1304 (2011)
12.
Zurück zum Zitat Laptev, I.: On space-time interest points. Int. J. Comput. Vision 64(2–3), 107–123 (2005)CrossRef Laptev, I.: On space-time interest points. Int. J. Comput. Vision 64(2–3), 107–123 (2005)CrossRef
13.
Zurück zum Zitat Hernández-Vela, A.: BoVDW: bag-of-visual-and-depth-words for gestur recognition. In: 21st International Conference on Pattern Recognition, pp. 449–452 (2012) Hernández-Vela, A.: BoVDW: bag-of-visual-and-depth-words for gestur recognition. In: 21st International Conference on Pattern Recognition, pp. 449–452 (2012)
14.
Zurück zum Zitat Ming, Y., Ruan, Q., Hauptmann, A.: Activity recognition from RGB-D camera with 3D local spatio-temporal features. In: IEEE Conference on Multimedia and Expo, pp. 344–349 (2012) Ming, Y., Ruan, Q., Hauptmann, A.: Activity recognition from RGB-D camera with 3D local spatio-temporal features. In: IEEE Conference on Multimedia and Expo, pp. 344–349 (2012)
15.
Zurück zum Zitat Wan, J., Ruan, Q., Li, W.: One-shot learning gesture recognition from RGB-D data using bag of features. J. Mach. Learn. Res. 14(9), 2549–2582 (2013) Wan, J., Ruan, Q., Li, W.: One-shot learning gesture recognition from RGB-D data using bag of features. J. Mach. Learn. Res. 14(9), 2549–2582 (2013)
16.
Zurück zum Zitat Wan, J., Ruan, Q., Li, W., An, G., Zhao, R.: 3D SMoSIFT three-dimensional sparse motion scale invariant feature transform for activity recognition from RGB-D videos. J. Electron. Imaging 23(2), 023017 (2014)CrossRef Wan, J., Ruan, Q., Li, W., An, G., Zhao, R.: 3D SMoSIFT three-dimensional sparse motion scale invariant feature transform for activity recognition from RGB-D videos. J. Electron. Imaging 23(2), 023017 (2014)CrossRef
17.
Zurück zum Zitat Wang, J., Yang, J., Yu, K., Lv, F., Huang, T., Gong, Y.: Locality-constrained linear coding for image classification. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 3360–3367 (2010) Wang, J., Yang, J., Yu, K., Lv, F., Huang, T., Gong, Y.: Locality-constrained linear coding for image classification. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 3360–3367 (2010)
18.
Zurück zum Zitat Liang, G., Chen, Y., Fang, G., Chen, X., Gao, W.: A vision-based sign language recognition system using tied-mixture density HMM. In: Proceedings of the 6th International Conference on Multimodal Interfaces, pp. 198–204 (2004) Liang, G., Chen, Y., Fang, G., Chen, X., Gao, W.: A vision-based sign language recognition system using tied-mixture density HMM. In: Proceedings of the 6th International Conference on Multimodal Interfaces, pp. 198–204 (2004)
19.
Zurück zum Zitat Ramamoorthy, A., Vaswani, N., Chaudhury, S., Banerjee, S.: Recognition of dynamic hand gestures. Pattern Recogn. 36(9), 2069–2081 (2003)CrossRefMATH Ramamoorthy, A., Vaswani, N., Chaudhury, S., Banerjee, S.: Recognition of dynamic hand gestures. Pattern Recogn. 36(9), 2069–2081 (2003)CrossRefMATH
20.
Zurück zum Zitat Mahbub, U., Roy, T., Rahman, Md.S., Imtiaz, H.: One-shot-learning gesture recognition using motion history based gesture silhouettes. In: Proceedings of the International Conference on Industrial Application Engineering, pp. 186–193 (2013) Mahbub, U., Roy, T., Rahman, Md.S., Imtiaz, H.: One-shot-learning gesture recognition using motion history based gesture silhouettes. In: Proceedings of the International Conference on Industrial Application Engineering, pp. 186–193 (2013)
21.
Zurück zum Zitat Wan, J., Ruan, Q., An, G., Li, W.: Hand tracking and segmentation via graph cuts and dynamic model in sign language videos. In: Proceedings of IEEE 11th International Conference on Signal Processing, vol. 2, pp. 1135–1138 (2012) Wan, J., Ruan, Q., An, G., Li, W.: Hand tracking and segmentation via graph cuts and dynamic model in sign language videos. In: Proceedings of IEEE 11th International Conference on Signal Processing, vol. 2, pp. 1135–1138 (2012)
22.
Zurück zum Zitat Shao, L., Ji, L.: Motion histogram analysis based key frame extraction for human action/activity representation. In: Proceedings of Canadian Conference on Computer and Robot Vision, pp. 88–92 (2009) Shao, L., Ji, L.: Motion histogram analysis based key frame extraction for human action/activity representation. In: Proceedings of Canadian Conference on Computer and Robot Vision, pp. 88–92 (2009)
23.
Zurück zum Zitat Zhou, R., Junsong, Y., Jingjing, M., Zhengyou, Z.: Robust part-based hand gesture recognition using Kinect sensor. IEEE Trans. Multimedia 15(5), 1110–1120 (2013)CrossRef Zhou, R., Junsong, Y., Jingjing, M., Zhengyou, Z.: Robust part-based hand gesture recognition using Kinect sensor. IEEE Trans. Multimedia 15(5), 1110–1120 (2013)CrossRef
24.
Zurück zum Zitat Reyes, M., Dominguez, G., Escalera, S.: Feature weighting in dynamic time warping for gesture recognition in depth data. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1182–1188 (2011) Reyes, M., Dominguez, G., Escalera, S.: Feature weighting in dynamic time warping for gesture recognition in depth data. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1182–1188 (2011)
25.
Zurück zum Zitat Fanello, S.R., Gori, I., Metta, G., Odone, F.: One-shot learning for real-time action recognition. In: Sanches, J.M., Micó, L., Cardoso, J.S. (eds.) IbPRIA 2013. LNCS, vol. 7887, pp. 31–40. Springer, Heidelberg (2013)CrossRef Fanello, S.R., Gori, I., Metta, G., Odone, F.: One-shot learning for real-time action recognition. In: Sanches, J.M., Micó, L., Cardoso, J.S. (eds.) IbPRIA 2013. LNCS, vol. 7887, pp. 31–40. Springer, Heidelberg (2013)CrossRef
26.
Zurück zum Zitat Shi, J., Tomasi, C.: Good features to track. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 593–600 (1994) Shi, J., Tomasi, C.: Good features to track. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 593–600 (1994)
28.
Zurück zum Zitat Mikolajczyk, K., Schmid, C.: A performance evaluation of local descriptors. IEEE Trans. on Pattern Anal. Mach. Intell. (S0162-8828), 1615–1630 (2005) Mikolajczyk, K., Schmid, C.: A performance evaluation of local descriptors. IEEE Trans. on Pattern Anal. Mach. Intell. (S0162-8828), 1615–1630 (2005)
29.
Zurück zum Zitat Farnebäck, G.: Two-Frame motion estimation based on polynomial expansion. In: Bigun, J., Gustavsson, T. (eds.) SCIA 2003. LNCS, vol. 2749, pp. 363–370. Springer, Heidelberg (2003)CrossRef Farnebäck, G.: Two-Frame motion estimation based on polynomial expansion. In: Bigun, J., Gustavsson, T. (eds.) SCIA 2003. LNCS, vol. 2749, pp. 363–370. Springer, Heidelberg (2003)CrossRef
Metadaten
Titel
Improved GLOH Approach for One-Shot Learning Human Gesture Recognition
verfasst von
Nabin Kumar Karn
Feng Jiang
Copyright-Jahr
2016
DOI
https://doi.org/10.1007/978-3-319-46654-5_49