Skip to main content
Erschienen in: Neural Computing and Applications 20/2020

02.03.2020 | S.I. : Applying Artificial Intelligence to the Internet of Things

Deep neural learning techniques with long short-term memory for gesture recognition

verfasst von: Deepak Kumar Jain, Aniket Mahanti, Pourya Shamsolmoali, Ramachandran Manikandan

Erschienen in: Neural Computing and Applications | Ausgabe 20/2020

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Gesture recognition is a kind of biometric which has assumed great significance in the field of computer vision for communicating information through human activities. To recognize the various gestures and achieve efficient classification, an efficient computational machine learning technique is required. The Shift Invariant Convolutional Deep Structured Neural Learning with Long Short-Term Memory (SICDSNL–LSTM) and Bivariate Fully Recurrent Deep Neural Network with Long Short-Term Memory (BFRDNN–LSTM) have been introduced for classifying human activities with efficient accuracy and minimal time complexity. The SICDSNL–LSTM technique collects gesture data (a kind of biometric) from the dataset and gives it to the input layers of Shift Invariant Convolutional Deep Structured Neural Learning. The SICDSNL–LSTM technique uses two hidden layers for performing regression and classification. In the regression process, dice similarity is used for measuring the relationship between data and output classes. In the second process, the input data are classified into dissimilar classes for each time step using LSTM unit with soft-step activation function. The soft-step activation function uses ‘forget gate’ for removing the less significant data from the cell state. This is also used to make a decision to display the output at a particular time step and to remove other class results. Then, LSTM output is given to the output layers, and convolutional deep neural learning is performed to classify the gesture. Based on the classification results, human activity and gesture are recognized with high accuracy. The BFRDNN–LSTM technique also performs regression in the first hidden layers using bivariate correlation to find relationship between data and classes. The LSTM unit in BFRDNN–LSTM technique uses Gaussian activation function in the second hidden layers for categorizing incoming data into various classes at each time step. In this proposed BFRDNN–LSTM method, fully recurrent deep neural network utilizes gradient descent function to minimize the error rate at the output layers and to increase the accuracy of the gesture recognition. Both SICDSNL–LSTM and BFRDNN–LSTM techniques automatically learn the features and the data to minimize time complexity in gesture recognition. Experimental evaluation is carried out using factors such as gesture recognition accuracy, false-positive rate and time complexity with a number of data.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Chong Y, Huang J, Pan S (2016) Hand gesture recognition using appearance features based on 3d point cloud. J Softw Eng Appl 9(04):103CrossRef Chong Y, Huang J, Pan S (2016) Hand gesture recognition using appearance features based on 3d point cloud. J Softw Eng Appl 9(04):103CrossRef
2.
Zurück zum Zitat Czuszyński K, Rumiński J, Kwaśniewska A (2018) Gesture recognition with the linear optical sensor and recurrent neural networks. IEEE Sens J 18(13):5429–5438CrossRef Czuszyński K, Rumiński J, Kwaśniewska A (2018) Gesture recognition with the linear optical sensor and recurrent neural networks. IEEE Sens J 18(13):5429–5438CrossRef
3.
Zurück zum Zitat Davila J, Cretu AM, Zaremba M (2017) Wearable sensor data classification for human activity recognition based on an iterative learning framework. Sensors 17(6):1287CrossRef Davila J, Cretu AM, Zaremba M (2017) Wearable sensor data classification for human activity recognition based on an iterative learning framework. Sensors 17(6):1287CrossRef
4.
Zurück zum Zitat Ding H, He Q, Zhou Y, Dan G, Cui S (2017) An individual finger gesture recognition system based on motion-intent analysis using mechanomyogram signal. Front Neurol 8:573CrossRef Ding H, He Q, Zhou Y, Dan G, Cui S (2017) An individual finger gesture recognition system based on motion-intent analysis using mechanomyogram signal. Front Neurol 8:573CrossRef
5.
Zurück zum Zitat Dominio F, Donadeo M, Zanuttigh P (2014) Combining multiple depth-based descriptors for hand gesture recognition. Pattern Recognit Lett 50:101–111CrossRef Dominio F, Donadeo M, Zanuttigh P (2014) Combining multiple depth-based descriptors for hand gesture recognition. Pattern Recognit Lett 50:101–111CrossRef
7.
Zurück zum Zitat Frolova D, Stern H, Berman S (2013) Most probable longest common subsequence for recognition of gesture character input. IEEE Trans Cybern 43(3):871–880CrossRef Frolova D, Stern H, Berman S (2013) Most probable longest common subsequence for recognition of gesture character input. IEEE Trans Cybern 43(3):871–880CrossRef
8.
Zurück zum Zitat Gao L, Bourke A, Nelson J (2014) Evaluation of accelerometer based multi-sensor versus single-sensor activity recognition systems. Med Eng Phys 36(6):779–785CrossRef Gao L, Bourke A, Nelson J (2014) Evaluation of accelerometer based multi-sensor versus single-sensor activity recognition systems. Med Eng Phys 36(6):779–785CrossRef
9.
Zurück zum Zitat Goyal M, Shahi B, Prema K, Reddy NS (2017) Performance analysis of human gesture recognition techniques. In: 2017 2nd IEEE international conference on recent trends in electronics, information and communication technology (RTEICT), IEEE, pp 111–115 Goyal M, Shahi B, Prema K, Reddy NS (2017) Performance analysis of human gesture recognition techniques. In: 2017 2nd IEEE international conference on recent trends in electronics, information and communication technology (RTEICT), IEEE, pp 111–115
10.
Zurück zum Zitat Gupta A, Sehrawat VK, Khosla M (2012) Fpga based real time human hand gesture recognition system. Proc Technol 6:98–107CrossRef Gupta A, Sehrawat VK, Khosla M (2012) Fpga based real time human hand gesture recognition system. Proc Technol 6:98–107CrossRef
11.
Zurück zum Zitat Hasan MM, Mishra PK (2012) Robust gesture recognition using Gaussian distribution for features fitting. Int J Mach Learn Comput 2(3):266CrossRef Hasan MM, Mishra PK (2012) Robust gesture recognition using Gaussian distribution for features fitting. Int J Mach Learn Comput 2(3):266CrossRef
13.
Zurück zum Zitat Joshi A, Monnier C, Betke M, Sclaroff S (2017) Comparing random forest approaches to segmenting and classifying gestures. Image Vis Comput 58:86–95CrossRef Joshi A, Monnier C, Betke M, Sclaroff S (2017) Comparing random forest approaches to segmenting and classifying gestures. Image Vis Comput 58:86–95CrossRef
14.
Zurück zum Zitat Kılıboz NÇ, Güdükbay U (2015) A hand gesture recognition technique for human–computer interaction. J Vis Commun Image Represent 28:97–104CrossRef Kılıboz NÇ, Güdükbay U (2015) A hand gesture recognition technique for human–computer interaction. J Vis Commun Image Represent 28:97–104CrossRef
15.
Zurück zum Zitat Kim JH, Hong GS, Kim BG, Dogra DP (2018) deepgesture: deep learning-based gesture recognition scheme using motion sensors. Displays 55:38–45CrossRef Kim JH, Hong GS, Kim BG, Dogra DP (2018) deepgesture: deep learning-based gesture recognition scheme using motion sensors. Displays 55:38–45CrossRef
16.
Zurück zum Zitat Kwon Y, Kang K, Bae C (2014) Unsupervised learning for human activity recognition using smartphone sensors. Expert Syst Appl 41(14):6067–6074CrossRef Kwon Y, Kang K, Bae C (2014) Unsupervised learning for human activity recognition using smartphone sensors. Expert Syst Appl 41(14):6067–6074CrossRef
17.
Zurück zum Zitat Li C, Xie C, Zhang B, Chen C, Han J (2018a) Deep fisher discriminant learning for mobile hand gesture recognition. Pattern Recognit 77:276–288CrossRef Li C, Xie C, Zhang B, Chen C, Han J (2018a) Deep fisher discriminant learning for mobile hand gesture recognition. Pattern Recognit 77:276–288CrossRef
18.
Zurück zum Zitat Li Y, Wang X, Liu W, Feng B (2018b) Deep attention network for joint hand gesture localization and recognition using static RGB-D images. Inf Sci 441:66–78MathSciNetCrossRef Li Y, Wang X, Liu W, Feng B (2018b) Deep attention network for joint hand gesture localization and recognition using static RGB-D images. Inf Sci 441:66–78MathSciNetCrossRef
19.
Zurück zum Zitat Li YT, Wachs JP (2014) Hegm: a hierarchical elastic graph matching for hand gesture recognition. Pattern Recognit 47(1):80–88CrossRef Li YT, Wachs JP (2014) Hegm: a hierarchical elastic graph matching for hand gesture recognition. Pattern Recognit 47(1):80–88CrossRef
20.
Zurück zum Zitat Liu K, Chen C, Jafari R, Kehtarnavaz N (2014) Fusion of inertial and depth sensor data for robust hand gesture recognition. IEEE Sens J 14(6):1898–1903CrossRef Liu K, Chen C, Jafari R, Kehtarnavaz N (2014) Fusion of inertial and depth sensor data for robust hand gesture recognition. IEEE Sens J 14(6):1898–1903CrossRef
21.
Zurück zum Zitat Liu YT, Zhang YA, Zeng M (2018) Novel algorithm for hand gesture recognition utilizing a wrist-worn inertial sensor. IEEE Sens J 18(24):10,085–10,095CrossRef Liu YT, Zhang YA, Zeng M (2018) Novel algorithm for hand gesture recognition utilizing a wrist-worn inertial sensor. IEEE Sens J 18(24):10,085–10,095CrossRef
22.
Zurück zum Zitat Madeo RCB, Peres SM, de Moraes Lima CA (2016) Gesture phase segmentation using support vector machines. Expert Syst Appl 56:100–115CrossRef Madeo RCB, Peres SM, de Moraes Lima CA (2016) Gesture phase segmentation using support vector machines. Expert Syst Appl 56:100–115CrossRef
23.
Zurück zum Zitat Nguyen-Duc-Thanh N, Lee S, Kim D (2012) Two-stage hidden markov model in gesture recognition for human robot interaction. Int J Adv Robot Syst 9(2):39CrossRef Nguyen-Duc-Thanh N, Lee S, Kim D (2012) Two-stage hidden markov model in gesture recognition for human robot interaction. Int J Adv Robot Syst 9(2):39CrossRef
24.
Zurück zum Zitat Nunez JC, Cabido R, Pantrigo JJ, Montemayor AS, Velez JF (2018) Convolutional neural networks and long short-term memory for skeleton-based human activity and hand gesture recognition. Pattern Recognit 76:80–94CrossRef Nunez JC, Cabido R, Pantrigo JJ, Montemayor AS, Velez JF (2018) Convolutional neural networks and long short-term memory for skeleton-based human activity and hand gesture recognition. Pattern Recognit 76:80–94CrossRef
25.
Zurück zum Zitat Ordóñez F, Roggen D (2016) Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1):115CrossRef Ordóñez F, Roggen D (2016) Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1):115CrossRef
26.
Zurück zum Zitat Rahman SU, Afroze Z, Tareq M (2014) Hand gesture recognition techniques for human computer interaction using open cv. Int J Sci Res Publ 4(12):1–6 Rahman SU, Afroze Z, Tareq M (2014) Hand gesture recognition techniques for human computer interaction using open cv. Int J Sci Res Publ 4(12):1–6
27.
Zurück zum Zitat Razzaq M, Cleland I, Nugent C, Lee S (2018) Multimodal sensor data fusion for activity recognition using filtered classifier. In: Multidisciplinary digital publishing institute proceedings, vol 2, p 1262 Razzaq M, Cleland I, Nugent C, Lee S (2018) Multimodal sensor data fusion for activity recognition using filtered classifier. In: Multidisciplinary digital publishing institute proceedings, vol 2, p 1262
28.
Zurück zum Zitat Ronao CA, Cho SB (2017) Recognizing human activities from smartphone sensors using hierarchical continuous hidden markov models. Int J Distrib Sens Netw 13(1):1550147716683,687CrossRef Ronao CA, Cho SB (2017) Recognizing human activities from smartphone sensors using hierarchical continuous hidden markov models. Int J Distrib Sens Netw 13(1):1550147716683,687CrossRef
29.
Zurück zum Zitat Sagayam KM, Hemanth DJ (2017) Hand posture and gesture recognition techniques for virtual reality applications: a survey. Virtual Real 21(2):91–107CrossRef Sagayam KM, Hemanth DJ (2017) Hand posture and gesture recognition techniques for virtual reality applications: a survey. Virtual Real 21(2):91–107CrossRef
30.
Zurück zum Zitat Santos D, Fernandes B, Bezerra B (2015) HAGR-D: a novel approach for gesture recognition with depth maps. Sensors 15(11):28,646–28,664CrossRef Santos D, Fernandes B, Bezerra B (2015) HAGR-D: a novel approach for gesture recognition with depth maps. Sensors 15(11):28,646–28,664CrossRef
31.
Zurück zum Zitat Soares AD, Apolinário AL Jr (2017) Real-time 3d gesture recognition using dynamic time warping and simplification methods. J WSCG 25(1):59–66 Soares AD, Apolinário AL Jr (2017) Real-time 3d gesture recognition using dynamic time warping and simplification methods. J WSCG 25(1):59–66
33.
Zurück zum Zitat Tai TM, Jhang YJ, Liao ZW, Teng KC, Hwang WJ (2018) Sensor-based continuous hand gesture recognition by long short-term memory. IEEE Sens Lett 2(3):1–4CrossRef Tai TM, Jhang YJ, Liao ZW, Teng KC, Hwang WJ (2018) Sensor-based continuous hand gesture recognition by long short-term memory. IEEE Sens Lett 2(3):1–4CrossRef
34.
Zurück zum Zitat Trabelsi D, Mohammed S, Chamroukhi F, Oukhellou L, Amirat Y (2013) An unsupervised approach for automatic activity recognition based on hidden markov model regression. IEEE Trans Autom Sci Eng 10(3):829–835CrossRef Trabelsi D, Mohammed S, Chamroukhi F, Oukhellou L, Amirat Y (2013) An unsupervised approach for automatic activity recognition based on hidden markov model regression. IEEE Trans Autom Sci Eng 10(3):829–835CrossRef
37.
Zurück zum Zitat Zhu J, San-Segundo R, Pardo JM (2017) Feature extraction for robust physical activity recognition. Hum Centric Comput Inf Sci 7(1):16CrossRef Zhu J, San-Segundo R, Pardo JM (2017) Feature extraction for robust physical activity recognition. Hum Centric Comput Inf Sci 7(1):16CrossRef
Metadaten
Titel
Deep neural learning techniques with long short-term memory for gesture recognition
verfasst von
Deepak Kumar Jain
Aniket Mahanti
Pourya Shamsolmoali
Ramachandran Manikandan
Publikationsdatum
02.03.2020
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 20/2020
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-020-04742-9

Weitere Artikel der Ausgabe 20/2020

Neural Computing and Applications 20/2020 Zur Ausgabe

S.I. : Applying Artificial Intelligence to the Internet of Things

A deep neural network-based model for named entity recognition for Hindi language

S.I. : Advances in Bio-Inspired Intelligent Systems

Monitoring ALS from speech articulation kinematics

Premium Partner