Skip to main content
Top

2016 | OriginalPaper | Chapter

Video Affective Content Analysis Based on Protagonist via Convolutional Neural Network

Authors : Yingying Zhu, Zhengbo Jiang, Jianfeng Peng, Sheng-hua Zhong

Published in: Advances in Multimedia Information Processing - PCM 2016

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Affective recognition is an important and challenging task for video content analysis. Affective information in videos is closely related to the viewer’s feelings and emotions. Thus, video affective content analysis has a great potential value. However, most of the previous methods are focused on how to effectively extract features from videos for affective analysis. There are several issues are worth to be investigated. For example, what information is used to express emotions in videos, and which information is useful to affect audiences’ emotions. Taking into account these issues, in this paper, we proposed a new video affective content analysis method based on protagonist information via Convolutional Neural Network (CNN). The proposed method is evaluated on the largest video emotion dataset and compared with some previous work. The experimental results show that our proposed affective analysis method based on protagonist information achieves best performance in emotion classification and prediction.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Zhu, Y., Huang, X., Huang, Q., Tian, Q.: Large-scale video copy retrieval with temporal-concentration sift. Neurocomputing 187, 83–91 (2016)CrossRef Zhu, Y., Huang, X., Huang, Q., Tian, Q.: Large-scale video copy retrieval with temporal-concentration sift. Neurocomputing 187, 83–91 (2016)CrossRef
2.
go back to reference Deng, C., Xu, J., Zhang, K., Tao, D., Gao, X., Li, X.: Similarity constraints-based structured output regression machine: an approach to image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. PP, 1 (2015)CrossRef Deng, C., Xu, J., Zhang, K., Tao, D., Gao, X., Li, X.: Similarity constraints-based structured output regression machine: an approach to image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. PP, 1 (2015)CrossRef
3.
go back to reference Zhong, S.h., Liu, Y., Ng, T.Y., Liu, Y.: Perception-oriented video saliency detection via spatio-temporal attention analysis. Neurocomputing 207, 178–188 (2016)CrossRef Zhong, S.h., Liu, Y., Ng, T.Y., Liu, Y.: Perception-oriented video saliency detection via spatio-temporal attention analysis. Neurocomputing 207, 178–188 (2016)CrossRef
4.
go back to reference Yuan, H., Kwong, S., Wang, X., Zhang, Y., Li, F.: A virtual view PSNR estimation method for 3-D videos. IEEE Trans. Broadcast. 62(1), 134–140 (2016)CrossRef Yuan, H., Kwong, S., Wang, X., Zhang, Y., Li, F.: A virtual view PSNR estimation method for 3-D videos. IEEE Trans. Broadcast. 62(1), 134–140 (2016)CrossRef
5.
go back to reference Wang, S., Ji, Q.: Video affective content analysis: a survey of state-of-the-art methods. IEEE Trans. Affect. Comput. 6(4), 410–430 (2015)CrossRef Wang, S., Ji, Q.: Video affective content analysis: a survey of state-of-the-art methods. IEEE Trans. Affect. Comput. 6(4), 410–430 (2015)CrossRef
6.
go back to reference Hanjalic, A.: Extracting moods from pictures and sounds: towards truly personalized TV. IEEE Sig. Process. Mag. 23(2), 90–100 (2006)CrossRef Hanjalic, A.: Extracting moods from pictures and sounds: towards truly personalized TV. IEEE Sig. Process. Mag. 23(2), 90–100 (2006)CrossRef
7.
go back to reference Zhao, S., Yao, H., Sun, X., Xu, P., Liu, X., Ji, R.: Video indexing and recommendation based on affective analysis of viewers. In: Proceedings of the 19th ACM International Conference on Multimedia, pp. 1473–1476. ACM (2011) Zhao, S., Yao, H., Sun, X., Xu, P., Liu, X., Ji, R.: Video indexing and recommendation based on affective analysis of viewers. In: Proceedings of the 19th ACM International Conference on Multimedia, pp. 1473–1476. ACM (2011)
8.
go back to reference Acar, E., Hopfgartner, F., Albayrak, S.: Understanding affective content of music videos through learned representations. In: Gurrin, C., Hopfgartner, F., Hurst, W., Johansen, H., Lee, H., O’Connor, N. (eds.) MMM 2014. LNCS, vol. 8325, pp. 303–314. Springer, Heidelberg (2014). doi:10.1007/978-3-319-04114-8_26 CrossRef Acar, E., Hopfgartner, F., Albayrak, S.: Understanding affective content of music videos through learned representations. In: Gurrin, C., Hopfgartner, F., Hurst, W., Johansen, H., Lee, H., O’Connor, N. (eds.) MMM 2014. LNCS, vol. 8325, pp. 303–314. Springer, Heidelberg (2014). doi:10.​1007/​978-3-319-04114-8_​26 CrossRef
9.
go back to reference Wang, H.L., Cheong, L.F.: Affective understanding in film. IEEE Trans. Circuits Syst. Video Technol. 16(6), 689–704 (2006)CrossRef Wang, H.L., Cheong, L.F.: Affective understanding in film. IEEE Trans. Circuits Syst. Video Technol. 16(6), 689–704 (2006)CrossRef
10.
go back to reference Cui, Y., Luo, S., Tian, Q., Zhang, S., Peng, Y., Jiang, L., Jin, J.S.: Mutual information-based emotion recognition. In: The Era of Interactive Media, pp. 471–479. Springer, New York (2014) Cui, Y., Luo, S., Tian, Q., Zhang, S., Peng, Y., Jiang, L., Jin, J.S.: Mutual information-based emotion recognition. In: The Era of Interactive Media, pp. 471–479. Springer, New York (2014)
11.
go back to reference Baveye, Y., Dellandrea, E., et al.: Deep learning vs. kernel methods: performance for emotion prediction in videos. In: Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 77–83. IEEE (2015) Baveye, Y., Dellandrea, E., et al.: Deep learning vs. kernel methods: performance for emotion prediction in videos. In: Proceedings of the 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 77–83. IEEE (2015)
12.
go back to reference Hubel, D.H., Wiesel, T.N.: Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 160(1), 106–154 (1962)CrossRef Hubel, D.H., Wiesel, T.N.: Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 160(1), 106–154 (1962)CrossRef
13.
go back to reference Felleman, D.J., Van Essen, D.C.: Distributed hierarchical processing in the primate cerebral cortex. Cereb. Cortex 1(1), 1–47 (1991)CrossRef Felleman, D.J., Van Essen, D.C.: Distributed hierarchical processing in the primate cerebral cortex. Cereb. Cortex 1(1), 1–47 (1991)CrossRef
14.
go back to reference Jin, C.-B., Li, S., Do, T.D., Kim, H.: Real-Time human action recognition using CNN over temporal images for static video surveillance cameras. In: Ho, Y.-S., Sang, J., Ro, Y.M., Kim, J., Wu, F. (eds.) PCM 2015. LNCS, vol. 9315, pp. 330–339. Springer, Heidelberg (2015). doi:10.1007/978-3-319-24078-7_33 CrossRef Jin, C.-B., Li, S., Do, T.D., Kim, H.: Real-Time human action recognition using CNN over temporal images for static video surveillance cameras. In: Ho, Y.-S., Sang, J., Ro, Y.M., Kim, J., Wu, F. (eds.) PCM 2015. LNCS, vol. 9315, pp. 330–339. Springer, Heidelberg (2015). doi:10.​1007/​978-3-319-24078-7_​33 CrossRef
15.
go back to reference He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. PAMI 37(9), 1904–1916 (2015)CrossRef He, K., Zhang, X., Ren, S., Sun, J.: Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. PAMI 37(9), 1904–1916 (2015)CrossRef
16.
go back to reference Kahou, S.E., Pal, C., Bouthillier, X., et al.: Combining modality specific deep neural networks for emotion recognition in video. In: Proceedings of the 15th ACM on International Conference on Multimodal Interaction, pp. 543–550. ACM (2013) Kahou, S.E., Pal, C., Bouthillier, X., et al.: Combining modality specific deep neural networks for emotion recognition in video. In: Proceedings of the 15th ACM on International Conference on Multimodal Interaction, pp. 543–550. ACM (2013)
17.
go back to reference Chen, T., Borth, D., Darrell, T.: Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. arXiv preprint arXiv:1410.8586 (2014) Chen, T., Borth, D., Darrell, T.: Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. arXiv preprint arXiv:​1410.​8586 (2014)
18.
go back to reference Baveye, Y., Dellandrea, E., Chamaret, C., Chen, L.: Liris-accede: a video database for affective content analysis. IEEE Trans. Affect. Comput. 6(1), 43–55 (2015)CrossRef Baveye, Y., Dellandrea, E., Chamaret, C., Chen, L.: Liris-accede: a video database for affective content analysis. IEEE Trans. Affect. Comput. 6(1), 43–55 (2015)CrossRef
19.
go back to reference Krizhevsky, A., Sutskever, I.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012) Krizhevsky, A., Sutskever, I.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012)
Metadata
Title
Video Affective Content Analysis Based on Protagonist via Convolutional Neural Network
Authors
Yingying Zhu
Zhengbo Jiang
Jianfeng Peng
Sheng-hua Zhong
Copyright Year
2016
DOI
https://doi.org/10.1007/978-3-319-48890-5_17