Skip to main content

2016 | OriginalPaper | Buchkapitel

Recognition of Social Touch Gestures Using 3D Convolutional Neural Networks

verfasst von : Nan Zhou, Jun Du

Erschienen in: Pattern Recognition

Verlag: Springer Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

This paper investigates on the deep learning approaches for the social touch gesture recognition. Several types of neural network architectures are studied with a comprehensive experiment design. First, recurrent neural network using long short-term memory (LSTM) is adopted for modeling the gesture sequence. However, for both handcrafted features using geometric moment and feature extraction using convolutional neural network (CNN), LSTM cannot achieve satisfactory performances. Therefore, we propose to use the 3D CNN to model a fixed length of touch gesture sequence. Experimental results show that the 3D CNN approach can achieve a recognition accuracy of 76.1 % on the human-animal affective robot touch (HAART) database in the recognition of social touch gestures challenge 2015, which significantly outperforms the best submitted system of the challenge with a recognition accuracy of 70.9 %.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Balli Altuglu, T., Altun, K.: Recognizing touch gestures for social human-robot interaction. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 407–413 (2015) Balli Altuglu, T., Altun, K.: Recognizing touch gestures for social human-robot interaction. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 407–413 (2015)
2.
Zurück zum Zitat Cooney, M.D., Nishio, S., Ishiguro, H.: Recognizing affection for a touch-based interaction with a humanoid robot. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1420–1427 (2012) Cooney, M.D., Nishio, S., Ishiguro, H.: Recognizing affection for a touch-based interaction with a humanoid robot. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1420–1427 (2012)
3.
Zurück zum Zitat Billard, A., Bonfiglio, A., Cannata, G., Cosseddu, P., Dahl, T., Dautenhahn, K., Mastrogiovanni, F., Metta, G., Natale, L., Robins, B., et al.: The ROBOSKIN project: challenges and results. In: Padois, V., Bidaud, P., Khatib, O. (eds.) Romansy 19–Robot Design, Dynamics and Control. CISM International Centre for Mechanical Sciences, pp. 351–358. Springer, Vienna (2013)CrossRef Billard, A., Bonfiglio, A., Cannata, G., Cosseddu, P., Dahl, T., Dautenhahn, K., Mastrogiovanni, F., Metta, G., Natale, L., Robins, B., et al.: The ROBOSKIN project: challenges and results. In: Padois, V., Bidaud, P., Khatib, O. (eds.) Romansy 19–Robot Design, Dynamics and Control. CISM International Centre for Mechanical Sciences, pp. 351–358. Springer, Vienna (2013)CrossRef
4.
Zurück zum Zitat Knight, H., Toscano, R., Stiehl, W.D., Chang, A., Wang, Y., Breazeal, C.: Real-time social touch gesture recognition for sensate robots. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3715–3720 (2009) Knight, H., Toscano, R., Stiehl, W.D., Chang, A., Wang, Y., Breazeal, C.: Real-time social touch gesture recognition for sensate robots. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3715–3720 (2009)
5.
Zurück zum Zitat Jung, M.M., Cang, X.L., Poel, M., MacLean, K.E.: Touch challenge ‘15: recognizing social touch gestures. In: Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (ICMI), pp. 387–390 (2015) Jung, M.M., Cang, X.L., Poel, M., MacLean, K.E.: Touch challenge ‘15: recognizing social touch gestures. In: Proceedings of the 2015 ACM on International Conference on Multimodal Interaction (ICMI), pp. 387–390 (2015)
6.
Zurück zum Zitat Ta, V.-C., Johal, W., Portaz, M., Castelli, E., Vaufreydaz, D.: The Grenoble system for the social touch challenge at ICMI 2015. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 391–398 (2015) Ta, V.-C., Johal, W., Portaz, M., Castelli, E., Vaufreydaz, D.: The Grenoble system for the social touch challenge at ICMI 2015. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 391–398 (2015)
7.
Zurück zum Zitat Hughes, D., Farrow, N., Profita, H., Correll, N.: Detecting and identifying tactile gestures using deep autoencoders, geometric moments and gesture level features. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 415–422 (2015) Hughes, D., Farrow, N., Profita, H., Correll, N.: Detecting and identifying tactile gestures using deep autoencoders, geometric moments and gesture level features. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 415–422 (2015)
8.
Zurück zum Zitat Falinie, Y. Gaus, A., Olugbade, T., Jan, A., Qin, R., Liu, J., Zhang, F., Meng, H., Bianchi-Berthouze, N.: Social touch gesture recognition using random forest and boosting on distinct feature sets. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 399–406 (2015) Falinie, Y. Gaus, A., Olugbade, T., Jan, A., Qin, R., Liu, J., Zhang, F., Meng, H., Bianchi-Berthouze, N.: Social touch gesture recognition using random forest and boosting on distinct feature sets. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 399–406 (2015)
9.
Zurück zum Zitat Altuglu, T.B., Altun, K.: Recognizing touch gestures for social human-robot interaction. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 407–413 (2015) Altuglu, T.B., Altun, K.: Recognizing touch gestures for social human-robot interaction. In: Proceedings of the 2015 ACM International Conference on Multimodal Interaction (ICMI), pp. 407–413 (2015)
10.
Zurück zum Zitat LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef
11.
Zurück zum Zitat Yu, K., Xu, W., Gong, Y.: Deep learning with kernel regularization for visual recognition. In: NIPS, pp. 1889–1896 (2008) Yu, K., Xu, W., Gong, Y.: Deep learning with kernel regularization for visual recognition. In: NIPS, pp. 1889–1896 (2008)
12.
Zurück zum Zitat Ahmed, A., Yu, K., Xu, W., Gong, Y., Xing, E.: Training hierarchical feed-forward visual recognition models using transfer learning from pseudo-tasks. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008. LNCS, vol. 5305, pp. 69–82. Springer, Heidelberg (2008). doi:10.1007/978-3-540-88690-7_6 CrossRef Ahmed, A., Yu, K., Xu, W., Gong, Y., Xing, E.: Training hierarchical feed-forward visual recognition models using transfer learning from pseudo-tasks. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008. LNCS, vol. 5305, pp. 69–82. Springer, Heidelberg (2008). doi:10.​1007/​978-3-540-88690-7_​6 CrossRef
13.
Zurück zum Zitat Mobahi, H., Collobert, R., Weston, J.: Deep learning from temporal coherence in video. In: ICML, pp. 737–744 (2009) Mobahi, H., Collobert, R., Weston, J.: Deep learning from temporal coherence in video. In: ICML, pp. 737–744 (2009)
14.
Zurück zum Zitat Bengio, Y., Simard, P., Frasconi, P.: Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 5(2), 157–166 (1994)CrossRef Bengio, Y., Simard, P., Frasconi, P.: Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 5(2), 157–166 (1994)CrossRef
15.
Zurück zum Zitat Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)CrossRef
16.
Zurück zum Zitat Teh, C.-H., Chin, R.T.: On image analysis by the methods of moments. IEEE Trans. Pattern Anal. Mach. Intell. 10(4), 496–513 (1988)CrossRefMATH Teh, C.-H., Chin, R.T.: On image analysis by the methods of moments. IEEE Trans. Pattern Anal. Mach. Intell. 10(4), 496–513 (1988)CrossRefMATH
17.
Zurück zum Zitat Donahue, J., Hendricks, L.A., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description (2014). CoRR, abs/1411.4389 Donahue, J., Hendricks, L.A., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description (2014). CoRR, abs/1411.4389
18.
Zurück zum Zitat Ji, S., Xu, W., Yang, M., Yu, K.: 3d convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221–231 (2013)CrossRef Ji, S., Xu, W., Yang, M., Yu, K.: 3d convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221–231 (2013)CrossRef
19.
Zurück zum Zitat Hinton, G.E., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Improving neural networks by preventing co-adaptation of feature detectors (2012). CoRR, abs/1207.0580 Hinton, G.E., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Improving neural networks by preventing co-adaptation of feature detectors (2012). CoRR, abs/1207.0580
20.
Zurück zum Zitat He, K., et al.: Delving deep into rectifiers: surpassing human-level performance on imagenet classification. arXiv preprint arXiv:1502.01852 (2015) He, K., et al.: Delving deep into rectifiers: surpassing human-level performance on imagenet classification. arXiv preprint arXiv:​1502.​01852 (2015)
Metadaten
Titel
Recognition of Social Touch Gestures Using 3D Convolutional Neural Networks
verfasst von
Nan Zhou
Jun Du
Copyright-Jahr
2016
Verlag
Springer Singapore
DOI
https://doi.org/10.1007/978-981-10-3002-4_14

Premium Partner