Skip to main content
Log in

Looking at the viewer: analysing facial activity to detect personal highlights of multimedia contents

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

This paper presents an approach to detect personal highlights in videos based on the analysis of facial activities of the viewer. Our facial activity analysis was based on the motion vectors tracked on twelve key points in the human face. In our approach, the magnitude of the motion vectors represented a degree of a viewer’s affective reaction to video contents. We examined 80 facial activity videos recorded for ten participants, each watching eight video clips in various genres. The experimental results suggest that useful motion vectors to detect personal highlights varied significantly across viewers. However, it was suggested that the activity in the upper part of face tended to be more indicative of personal highlights than the activity in the lower part.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Arifin S, Cheung P (2007) A computation method for video segmentation utilizing the pleasure-arousal-dominance emotional information. In: ACM international conference on multimedia

  2. Calvo R, D’Mello S (2010) Affect detection: an interdisciplinary review of models, methods, and their applications. IEEE Trans Aff Comp 1(1):18–37

    Article  Google Scholar 

  3. Chan CH, Jones GJF (2005) Affect-based indexing and retrieval of films. In: ACM international conference on multimedia, pp 427–430

  4. Chen L (2000) Joint processing of audio-visual information for the recognition of emotional expressions in human-computer interaction. PhD thesis, University of Illinois at Urbana-Champaign

  5. Cheon Y, Kim D (2009) Natural facial expression recognition using differential-AAM and manifold learning. Pattern Recogn 42(7):260–274

    Article  Google Scholar 

  6. Cohen I, Sebe N, Chen L, Garg A, Huang T (2003) Facial expression recognition from video sequences: temporal and static modeling. Comput Vis Image Underst 91(1–2):160–187

    Article  Google Scholar 

  7. Cohen I, Sebe N, Cozman F, Cirelo M, Huang T (2004) Semi-supervised learning of classifiers: theory, algorithms, and applications to human-computer interaction. IEEE Trans Pattern Anal Mach Intell 26(12):1553–1567

    Article  Google Scholar 

  8. Dietz R, Lang A (1999) Aefective agents: effects of agent affect on arousal, attention, liking and learning. In: Cognitive technology conference

  9. Ekman P, Friesen W (1978) Facial action coding system: investigator’s guide. Consulting Psychologists Press, Palo Alto

    Google Scholar 

  10. Hanjalic A (2006) Extracting moods from pictures and sounds: towards truly personalized TV. IEEE Signal Process Mag 2(23):90–100

    Article  Google Scholar 

  11. Hanjalic A, Xu LQ (2005) Affective video content representation and modeling. IEEE Trans Multimedia 7(1):143–154

    Article  Google Scholar 

  12. Hanjalic A, Lienhart R, Ma WY, Smith JR (2008) The holy grail of multimedia information retrieval: so close or yet so far away? Proc IEEE 96(4):541–547

    Article  Google Scholar 

  13. Huijsmans D, Sebe N (2005) How to complete performance graphs in content-based image retrieval: add generality and normalize scope. IEEE Trans Pattern Anal Mach Intell 27(2):245–251

    Article  Google Scholar 

  14. Jaimes A, Gatica-Perez D, Sebe N, Huang T (2007) Human-centered computing: towards a human revolution. IEEE Comp 5(40):30–34

    Google Scholar 

  15. Joho H, Jose JM, Valenti R, Sebe N (2009) Exploiting facial expressions for affective video summarisation. In: ACM international conference on image and video retrieval (CIVR)

  16. Kang H (2002) Analysis of scene context related with emotional events. In: ACM international conference on multimedia

  17. Madigan D, York J (1995) Bayesian graphical models for discrete data. Int Stat Rev 63:215–232

    Article  MATH  Google Scholar 

  18. Mehrabian A (1996) Pleasure-arousal-dominance: a general framework for describing and measuring individual differences in temperament. Curr Psychol 14(4):261–292

    Article  MathSciNet  Google Scholar 

  19. Moncrieff S, Dorai C, Venkatesh S (2001) Affect computing in film through sound energy dynamics. In: ACM international conference on multimedia

  20. Money A, Agius H (2008) Feasibility of personalized affective video summaries. In: Affect and emotion in human—computer interaction. Springer

  21. Money A, Agius H (2008) Video summarisation: a conceptual framework and survey of the state of the art. J Vis Commun Image Represent 19(2):121–143

    Article  Google Scholar 

  22. Mooney C, Scully M, Jones GJF, Smeaton AF (2006) Investigating biometric response for information retrieval applications. In: European conference on information retrieval, pp 570–574

  23. Over P, Smeaton AF, Kelly P (2007) The TRECVID 2007 BBC rushes summarization evaluation pilot. In: TVS ’07: Int. workshop on TRECVID video summarization, pp 1–15

  24. van Rijsbergen CJ (1979) Information retrieval, 2nd edn. Butterworths

  25. Sebe N, Cohen I, Cozman F, Huang T (2005) Learning probabilistic classifiers for human-computer interaction applications. Multimedia Syst 10(6):484–498

    Article  Google Scholar 

  26. Soleymani M, Chanel G, Kierkels JJ, Pun T (2008) Affective ranking of movie scenes using physiological signals and content analysis. In: ACM workshop on multimedia semantics, pp 32–39

  27. Sung J, Kanade T, Kim D (2008) Pose robust face tracking by combining active appearance models and cylinder head models. Int J Comput Vis 80(2):260–274

    Article  Google Scholar 

  28. Tao H, Huang T (1998) Connected vibrations: a modal analysis approach to non-rigid motion tracking. In: IEEE conference on compter vision and pattern recognition, pp 735–740

  29. Tjondronegoro D, Chen YP, Pham B (2004) Highlights for more complete sports video summarization. IEEE Multimed 11(4):22–37

    Article  Google Scholar 

  30. Wang H, Cheong L (2006) Affective understanding in film. IEEE Trans Circuits Syst Video Technol 16(6):689–704

    Article  Google Scholar 

  31. Xu M, Chia L, Jin J (2005) Affective content analysis in comedy and horror videos by audio emotional event detection. In: IEEE international conference on multimedia and expo

Download references

Acknowledgements

Funding was provided by the MIAUCE Project (EU IST-033715). Any opinions, findings, and conclusions described here are the authors and do not necessarily reflect those of the sponsor. The work of Jacopo Staiano and Nicu Sebe has been supported by the FP7 IP GLOCAL european project and by the FIRB S-PATTERNS project.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nicu Sebe.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Joho, H., Staiano, J., Sebe, N. et al. Looking at the viewer: analysing facial activity to detect personal highlights of multimedia contents. Multimed Tools Appl 51, 505–523 (2011). https://doi.org/10.1007/s11042-010-0632-x

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-010-0632-x

Keywords

Navigation