Skip to main content

2018 | OriginalPaper | Buchkapitel

RAMAS: Russian Multimodal Corpus of Dyadic Interaction for Affective Computing

verfasst von : Olga Perepelkina, Evdokia Kazimirova, Maria Konstantinova

Erschienen in: Speech and Computer

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Emotion expression encompasses various types of information, including face and eye movement, voice and body motion. Emotions collected from real conversations are difficult to classify using one channel. That is why multimodal techniques have recently become more popular in automatic emotion recognition. Multimodal databases that include audio, video, 3D motion capture and physiology data are quite rare. We collected The Russian Acted Multimodal Affective Set (RAMAS) − the first multimodal corpus in Russian language. Our database contains approximately 7 h of high-quality close-up video recordings of faces, speech, motion-capture data and such physiological signals as electro-dermal activity and photoplethysmogram. The subjects were 10 actors who played out interactive dyadic scenarios. Each scenario involved one of the basic emotions: Anger, Sadness, Disgust, Happiness, Fear or Surprise, and such characteristics of social interaction like Domination and Submission. In order to note emotions that subjects really felt during the process we asked them to fill in short questionnaires (self-reports) after each played scenario. The records were marked by 21 annotators (at least five annotators marked each scenario). We present our multimodal data collection, annotation process, inter-rater agreement analysis and the comparison between self-reports and received annotations. RAMAS is an open database that provides research community with multimodal data of faces, speech, gestures and physiology interrelation. Such material is useful for various investigations and automatic affective systems development.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
6.
Zurück zum Zitat Anderson, A., Hsiao, T., Metsis, V.: Classification of emotional arousal during multimedia exposure. In: Proceedings of the 10th International Conference on Pervasive Technologies Related to Assistive Environments, pp. 181–184. ACM (2017) Anderson, A., Hsiao, T., Metsis, V.: Classification of emotional arousal during multimedia exposure. In: Proceedings of the 10th International Conference on Pervasive Technologies Related to Assistive Environments, pp. 181–184. ACM (2017)
7.
Zurück zum Zitat Ayvaz, U., Gürüler, H., Devrim, M.O.: Use of facial emotion recognition in e-learning systems. Inf. Technol. Learn. Tools 60(4), 95–104 (2017) Ayvaz, U., Gürüler, H., Devrim, M.O.: Use of facial emotion recognition in e-learning systems. Inf. Technol. Learn. Tools 60(4), 95–104 (2017)
8.
Zurück zum Zitat Bänziger, T., Pirker, H., Scherer, K.: GEMEP-GEneva multimodal emotion portrayals: a corpus for the study of multimodal emotional expressions. In: Proceedings of LREC, vol. 6, pp. 15–19 (2006) Bänziger, T., Pirker, H., Scherer, K.: GEMEP-GEneva multimodal emotion portrayals: a corpus for the study of multimodal emotional expressions. In: Proceedings of LREC, vol. 6, pp. 15–19 (2006)
9.
Zurück zum Zitat Busso, C., Bulut, M., Lee, C.C., Kazemzadeh, A., Mower, E., Kim, S., Chang, J.N., Lee, S., Narayanan, S.S.: IEMOCAP: interactive emotional dyadic motion capture database. Lang. Resour. Eval. 42(4), 335 (2008)CrossRef Busso, C., Bulut, M., Lee, C.C., Kazemzadeh, A., Mower, E., Kim, S., Chang, J.N., Lee, S., Narayanan, S.S.: IEMOCAP: interactive emotional dyadic motion capture database. Lang. Resour. Eval. 42(4), 335 (2008)CrossRef
10.
Zurück zum Zitat Chaw, T.V., Khor, S.W., Lau, P.Y.: Facial expression recognition using correlation of eyes regions. In: The FICT Colloquium 2016, p. 34, December 2016 Chaw, T.V., Khor, S.W., Lau, P.Y.: Facial expression recognition using correlation of eyes regions. In: The FICT Colloquium 2016, p. 34, December 2016
11.
Zurück zum Zitat De Silva, L.C., Miyasato, T., Nakatsu, R.: Facial emotion recognition using multi-modal information. In: Proceedings of 1997 International Conference on Information, Communications and Signal Processing, ICICS 1997, vol. 1, pp. 397–401. IEEE (1997) De Silva, L.C., Miyasato, T., Nakatsu, R.: Facial emotion recognition using multi-modal information. In: Proceedings of 1997 International Conference on Information, Communications and Signal Processing, ICICS 1997, vol. 1, pp. 397–401. IEEE (1997)
13.
Zurück zum Zitat Douglas, M.: Purity and danger: an analysis of pollution and taboo London (1966) Douglas, M.: Purity and danger: an analysis of pollution and taboo London (1966)
14.
Zurück zum Zitat El Ayadi, M., Kamel, M.S., Karray, F.: Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recogn. 44(3), 572–587 (2011)CrossRef El Ayadi, M., Kamel, M.S., Karray, F.: Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recogn. 44(3), 572–587 (2011)CrossRef
15.
Zurück zum Zitat Gouizi, K., Bereksi Reguig, F., Maaoui, C.: Emotion recognition from physiological signals. J. Med. Eng. Technol. 35(6–7), 300–307 (2011)CrossRef Gouizi, K., Bereksi Reguig, F., Maaoui, C.: Emotion recognition from physiological signals. J. Med. Eng. Technol. 35(6–7), 300–307 (2011)CrossRef
16.
Zurück zum Zitat Hayes, A.F., Krippendorff, K.: Answering the call for a standard reliability measure for coding data. Commun. Methods Meas. 1(1), 77–89 (2007)CrossRef Hayes, A.F., Krippendorff, K.: Answering the call for a standard reliability measure for coding data. Commun. Methods Meas. 1(1), 77–89 (2007)CrossRef
17.
Zurück zum Zitat Karg, M., Samadani, A.A., Gorbet, R., Kühnlenz, K., Hoey, J., Kulić, D.: Body movements for affective expression: a survey of automatic recognition and generation. IEEE Trans. Affect. Comput. 4(4), 341–359 (2013)CrossRef Karg, M., Samadani, A.A., Gorbet, R., Kühnlenz, K., Hoey, J., Kulić, D.: Body movements for affective expression: a survey of automatic recognition and generation. IEEE Trans. Affect. Comput. 4(4), 341–359 (2013)CrossRef
18.
Zurück zum Zitat Krippendorff, K.: Estimating the reliability, systematic error and random error of interval data. Educ. Psychol. Meas. 30(1), 61–70 (1970)CrossRef Krippendorff, K.: Estimating the reliability, systematic error and random error of interval data. Educ. Psychol. Meas. 30(1), 61–70 (1970)CrossRef
19.
Zurück zum Zitat Mayer, J.D., Salovey, P., Caruso, D.R., Sitarenios, G.: Measuring emotional intelligence with the MSCEIT V2. 0. Emotion 3(1), 97 (2003)CrossRef Mayer, J.D., Salovey, P., Caruso, D.R., Sitarenios, G.: Measuring emotional intelligence with the MSCEIT V2. 0. Emotion 3(1), 97 (2003)CrossRef
20.
Zurück zum Zitat Metallinou, A., Lee, C.C., Busso, C., Carnicke, S., Narayanan, S.: The USC creativeIT database: a multimodal database of theatrical improvisation. Multimodal Corpora: Advances in Capturing, Coding and Analyzing Multimodality, p. 55 (2010) Metallinou, A., Lee, C.C., Busso, C., Carnicke, S., Narayanan, S.: The USC creativeIT database: a multimodal database of theatrical improvisation. Multimodal Corpora: Advances in Capturing, Coding and Analyzing Multimodality, p. 55 (2010)
21.
Zurück zum Zitat Rachman, S.: Anxiety. Psychology Press Ltd., Publishers, East Sussex (1998) Rachman, S.: Anxiety. Psychology Press Ltd., Publishers, East Sussex (1998)
23.
Zurück zum Zitat Ringeval, F., Sonderegger, A., Sauer, J., Lalanne, D.: Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In: 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–8. IEEE (2013) Ringeval, F., Sonderegger, A., Sauer, J., Lalanne, D.: Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions. In: 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–8. IEEE (2013)
24.
Zurück zum Zitat Russell, J.A., Fernández-Dols, J.M.: The Psychology of Facial Expression. Cambridge University Press, Cambridge (1997)CrossRef Russell, J.A., Fernández-Dols, J.M.: The Psychology of Facial Expression. Cambridge University Press, Cambridge (1997)CrossRef
25.
Zurück zum Zitat Sergienko, E.G., Vetrova, I.I., Volochkov, A.A., Popov, A.Y.: Adaptation of J. Mayer P. Salovey and D. Caruso emotional intelligence test on russian-speaking sample. Psikhologicheskii Zhurnal 31(1), 55–73 (2010) Sergienko, E.G., Vetrova, I.I., Volochkov, A.A., Popov, A.Y.: Adaptation of J. Mayer P. Salovey and D. Caruso emotional intelligence test on russian-speaking sample. Psikhologicheskii Zhurnal 31(1), 55–73 (2010)
26.
Zurück zum Zitat Sloetjes, H., Wittenburg, P.: Annotation by category: ELAN and ISO DCR. In: LREC (2008) Sloetjes, H., Wittenburg, P.: Annotation by category: ELAN and ISO DCR. In: LREC (2008)
27.
Zurück zum Zitat Tarnowski, P., Kołodziej, M., Majkowski, A., Rak, R.J.: Emotion recognition using facial expressions. Procedia Comput. Sci. 108, 1175–1184 (2017)CrossRef Tarnowski, P., Kołodziej, M., Majkowski, A., Rak, R.J.: Emotion recognition using facial expressions. Procedia Comput. Sci. 108, 1175–1184 (2017)CrossRef
28.
Zurück zum Zitat Tomkins, S.: Affect Imagery Consciousness: Volume II: The Negative Affects. Springer, New York (1963) Tomkins, S.: Affect Imagery Consciousness: Volume II: The Negative Affects. Springer, New York (1963)
29.
Zurück zum Zitat Volkova, E., De La Rosa, S., Bülthoff, H.H., Mohler, B.: The MPI emotional body expressions database for narrative scenarios. PloS one 9(12), e113647 (2014)CrossRef Volkova, E., De La Rosa, S., Bülthoff, H.H., Mohler, B.: The MPI emotional body expressions database for narrative scenarios. PloS one 9(12), e113647 (2014)CrossRef
30.
Zurück zum Zitat Wagner, J., Lingenfelser, F., Baur, T., Damian, I., Kistler, F., André, E.: The social signal interpretation (SSI) framework: multimodal signal processing and recognition in real-time. In: Proceedings of the 21st ACM International Conference on Multimedia, pp. 831–834. ACM (2013) Wagner, J., Lingenfelser, F., Baur, T., Damian, I., Kistler, F., André, E.: The social signal interpretation (SSI) framework: multimodal signal processing and recognition in real-time. In: Proceedings of the 21st ACM International Conference on Multimedia, pp. 831–834. ACM (2013)
Metadaten
Titel
RAMAS: Russian Multimodal Corpus of Dyadic Interaction for Affective Computing
verfasst von
Olga Perepelkina
Evdokia Kazimirova
Maria Konstantinova
Copyright-Jahr
2018
DOI
https://doi.org/10.1007/978-3-319-99579-3_52

Premium Partner