Skip to main content
Erschienen in: Artificial Life and Robotics 2/2019

19.12.2018 | Original Article

A deep unified framework for suspicious action recognition

verfasst von: Amine Ilidrissi, Joo Kooi Tan

Erschienen in: Artificial Life and Robotics | Ausgabe 2/2019

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

As action recognition undergoes change as a field under influence of the recent deep learning trend, and while research in areas such as background subtraction, object segmentation and action classification is steadily progressing, experiments devoted to evaluate a combination of the aforementioned fields, be it from a speed or a performance perspective, are far and few between. In this paper, we propose a deep, unified framework targeted towards suspicious action recognition that takes advantage of recent discoveries, fully leverages the power of convolutional neural networks and strikes a balance between speed and accuracy not accounted for in most research. We carry out performance evaluation on the KTH dataset and attain a 95.4% accuracy in 200 ms computational time, which compares favorably to other state-of-the-art methods. We also apply our framework to a video surveillance dataset and obtain 91.9% accuracy for suspicious actions in 205 ms computational time.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Baranwal M, Khan MT, De Silva CW (2011) Abnormal motion detection in real time using video surveillance and body sensors. Int J Inf Acquis 8:103–116CrossRef Baranwal M, Khan MT, De Silva CW (2011) Abnormal motion detection in real time using video surveillance and body sensors. Int J Inf Acquis 8:103–116CrossRef
2.
Zurück zum Zitat Setyawan FXA, Tan JK, Kim H et al (2017) Moving objects detection employing iterative update of the background. Artif Life Robot 22(2):168–174CrossRef Setyawan FXA, Tan JK, Kim H et al (2017) Moving objects detection employing iterative update of the background. Artif Life Robot 22(2):168–174CrossRef
3.
Zurück zum Zitat Braham M, Van Droogenbroeck M (2016) Deep background subtraction with scene-specific convolutional neural networks. In: international conference on systems, signals and image processing, pp 1–4 Braham M, Van Droogenbroeck M (2016) Deep background subtraction with scene-specific convolutional neural networks. In: international conference on systems, signals and image processing, pp 1–4
4.
Zurück zum Zitat Lucas BD, Kanade T (1981) An iterative image registration technique with an application to stereo vision. In: proceedings of imaging understanding workshop, pp 121–130 Lucas BD, Kanade T (1981) An iterative image registration technique with an application to stereo vision. In: proceedings of imaging understanding workshop, pp 121–130
5.
Zurück zum Zitat Farnebäck G (2003) Two-frame motion estimation based on polynomial expansion. Image Anal 2749:363–370CrossRefMATH Farnebäck G (2003) Two-frame motion estimation based on polynomial expansion. Image Anal 2749:363–370CrossRefMATH
6.
Zurück zum Zitat Zach C, Pock T, Bischof H (2007) A duality based approach for realtime TV-L1 optical flow. In: joint pattern recognition symposium, pp214–223 Zach C, Pock T, Bischof H (2007) A duality based approach for realtime TV-L1 optical flow. In: joint pattern recognition symposium, pp214–223
7.
Zurück zum Zitat Dosovitskiy A, Fischer P, Ilg E et al (2015) Flownet: learning optical flow with convolutional networks. In: proceedings of the IEEE international conference on computer vision, pp 2758–2766 Dosovitskiy A, Fischer P, Ilg E et al (2015) Flownet: learning optical flow with convolutional networks. In: proceedings of the IEEE international conference on computer vision, pp 2758–2766
8.
Zurück zum Zitat Ilg E, Mayer N, Saikia T et al (2017) FlowNet 2.0: evolution of optical flow estimation with deep networks. In: IEEE conference on computer vision and pattern recognition, pp 2462–2470 Ilg E, Mayer N, Saikia T et al (2017) FlowNet 2.0: evolution of optical flow estimation with deep networks. In: IEEE conference on computer vision and pattern recognition, pp 2462–2470
9.
Zurück zum Zitat Mayer N, Ilg E, Hausser P et al (2016) A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: proceedings of the IEEE conference on computer vision and pattern recognition, pp 4040–4048 Mayer N, Ilg E, Hausser P et al (2016) A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: proceedings of the IEEE conference on computer vision and pattern recognition, pp 4040–4048
10.
Zurück zum Zitat Bobick AF, Davis JW (2001) The recognition of human movement using temporal templates. IEEE Trans Pattern Anal Mach Intell 23(3):257–267CrossRef Bobick AF, Davis JW (2001) The recognition of human movement using temporal templates. IEEE Trans Pattern Anal Mach Intell 23(3):257–267CrossRef
11.
Zurück zum Zitat Canton-Ferrer C, Casas JR, Pardas M (2006) Human model and motion based 3D action recognition in multiple view scenarios. In: 14th European signal processing conference, pp 1–5 Canton-Ferrer C, Casas JR, Pardas M (2006) Human model and motion based 3D action recognition in multiple view scenarios. In: 14th European signal processing conference, pp 1–5
12.
Zurück zum Zitat Ahsan SMM, Tan JK, Kim H et al (2015) Human action representation and recognition: an approach to a histogram of spatiotemporal templates. Int J Innov Comput Inf Control 11(6):1855–1868 Ahsan SMM, Tan JK, Kim H et al (2015) Human action representation and recognition: an approach to a histogram of spatiotemporal templates. Int J Innov Comput Inf Control 11(6):1855–1868
13.
Zurück zum Zitat Ahad MAR, Ogata T, Tan JK et al (2008) A complex motion recognition technique employing directional motion templates. Int J Innov Comput Inf Control 4(8):1943–1954 Ahad MAR, Ogata T, Tan JK et al (2008) A complex motion recognition technique employing directional motion templates. Int J Innov Comput Inf Control 4(8):1943–1954
14.
Zurück zum Zitat Ahsan SMM, Tan JK, Kim H et al (2016) Spatiotemporal LBP and shape feature for human activity representation and recognition. Int J Innov Comput Inf Control 12(1):1–13 Ahsan SMM, Tan JK, Kim H et al (2016) Spatiotemporal LBP and shape feature for human activity representation and recognition. Int J Innov Comput Inf Control 12(1):1–13
15.
Zurück zum Zitat Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. Adv Neural Inf Process Syst 2014:568–576 Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. Adv Neural Inf Process Syst 2014:568–576
16.
Zurück zum Zitat Wang L, Xiong Y, Wang Z et al (2016) Temporal segment networks: towards good practices for deep action recognition. In: European conference on computer vision, pp 20–36 Wang L, Xiong Y, Wang Z et al (2016) Temporal segment networks: towards good practices for deep action recognition. In: European conference on computer vision, pp 20–36
18.
Zurück zum Zitat Jia Y, Shelhamer E, Donahue J et al (2014) Caffe: convolutional architecture for fast feature embedding. In: proceedings of the 22nd ACM international conference on Multimedia, pp 675–678 Jia Y, Shelhamer E, Donahue J et al (2014) Caffe: convolutional architecture for fast feature embedding. In: proceedings of the 22nd ACM international conference on Multimedia, pp 675–678
19.
Zurück zum Zitat LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278–2324CrossRef LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86:2278–2324CrossRef
20.
Zurück zum Zitat Wang Y, Jodoin PM, Porikli F et al (2014) CDnet 2014: an expanded change detection benchmark dataset. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 387–394 Wang Y, Jodoin PM, Porikli F et al (2014) CDnet 2014: an expanded change detection benchmark dataset. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 387–394
21.
22.
Zurück zum Zitat He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition. In: proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778 He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition. In: proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
23.
Zurück zum Zitat Schuldt C, Laptev I, Caputo B (2004) Recognizing human actions: a local SVM approach. In: proceedings of the 17th international conference on pattern recognition, pp 32–36 Schuldt C, Laptev I, Caputo B (2004) Recognizing human actions: a local SVM approach. In: proceedings of the 17th international conference on pattern recognition, pp 32–36
24.
Zurück zum Zitat Gao Z, Chen MY, Hauptmann A et al (2010) Comparing evaluation protocols on the KTH dataset. Hum Behav Underst 6219:88–100CrossRef Gao Z, Chen MY, Hauptmann A et al (2010) Comparing evaluation protocols on the KTH dataset. Hum Behav Underst 6219:88–100CrossRef
25.
Zurück zum Zitat Yi S, Li H, Wang X (2015) Understanding pedestrian behaviors from stationary crowd groups. In: IEEE conference on computer vision and pattern recognition, pp 3488–3496 Yi S, Li H, Wang X (2015) Understanding pedestrian behaviors from stationary crowd groups. In: IEEE conference on computer vision and pattern recognition, pp 3488–3496
26.
Zurück zum Zitat Han Y, Zhang P, Zhuo T et al (2017) Going deeper with two-stream ConvNets for action recognition in video surveillance. Pattern Recognit Lett 107:83–90CrossRef Han Y, Zhang P, Zhuo T et al (2017) Going deeper with two-stream ConvNets for action recognition in video surveillance. Pattern Recognit Lett 107:83–90CrossRef
27.
Zurück zum Zitat Kim TK, Wong SF, Cipolla R (2007) Tensor canonical correlation analysis for action classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–8 Kim TK, Wong SF, Cipolla R (2007) Tensor canonical correlation analysis for action classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1–8
28.
29.
Zurück zum Zitat Kuehne H, Jhuang H, Garrote E et al (2011) HMDB: a large video database for human motion recognition. In: IEEE international conference on computer vision, pp 2556–2563 Kuehne H, Jhuang H, Garrote E et al (2011) HMDB: a large video database for human motion recognition. In: IEEE international conference on computer vision, pp 2556–2563
30.
Metadaten
Titel
A deep unified framework for suspicious action recognition
verfasst von
Amine Ilidrissi
Joo Kooi Tan
Publikationsdatum
19.12.2018
Verlag
Springer Japan
Erschienen in
Artificial Life and Robotics / Ausgabe 2/2019
Print ISSN: 1433-5298
Elektronische ISSN: 1614-7456
DOI
https://doi.org/10.1007/s10015-018-0518-y

Weitere Artikel der Ausgabe 2/2019

Artificial Life and Robotics 2/2019 Zur Ausgabe

Neuer Inhalt