Skip to main content

2015 | OriginalPaper | Buchkapitel

Towards Classification of Engagement in Human Interaction with Talking Robots

verfasst von : Yuyun Huang, Christy Elias, João P. Cabral, Atul Nautiyal, Christian Saam, Nick Campbell

Erschienen in: HCI International 2015 - Posters’ Extended Abstracts

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

In this paper we describe ongoing work to develop an engagement classifier for human-computer interaction systems. We have successfully classified group and individual engagement in a corpus of a conversation among four people called TableTalk, by using a classifier trained with the Support Vector Machine method and audio-visual features. The goal in this paper is to extend that work for the classification of engagement in videos of interaction between an human and a talking robot. For that purpose we are using a corpus of dialogues between participants and a Lego robot named Herme, which was collected during an exhibition. We describe the techniques to improve the engagement detection by taking into account the differences between the characteristics of the videos between the two datasets. Currently we are also conducting an experiment to manually annotate the Herme videos with engagement labels. These annotations will be used for evaluation and further improvements to engagement detection.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Bohus, D., Horvitz, E.: Learning to predict engagement with a spoken dialog system in open-world settings. In: Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL), pp. 244–252. USA (2009) Bohus, D., Horvitz, E.: Learning to predict engagement with a spoken dialog system in open-world settings. In: Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL), pp. 244–252. USA (2009)
2.
Zurück zum Zitat Hernandez, J., Liu, Z., Hulten, G., DeBarr, D., Krum, K., Zhang, Z.: Measuring the engagement level of TV viewers. In: IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, pp. 1–7 (2013) Hernandez, J., Liu, Z., Hulten, G., DeBarr, D., Krum, K., Zhang, Z.: Measuring the engagement level of TV viewers. In: IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, pp. 1–7 (2013)
3.
Zurück zum Zitat Gustafson, J., Neiberg, D.: Prosodic cues to engagement in non-lexical response tokens in Swedish, In: DiSS-LPSS, Citeseer, pp. 63–66 (2010) Gustafson, J., Neiberg, D.: Prosodic cues to engagement in non-lexical response tokens in Swedish, In: DiSS-LPSS, Citeseer, pp. 63–66 (2010)
4.
Zurück zum Zitat Gatica-Perez, D., McCowan, I. A., Zhang, D., Bengio, S.: Detecting group interest-level in meetings. In: IEEE ICASSP, pp. 489–492 (2010) Gatica-Perez, D., McCowan, I. A., Zhang, D., Bengio, S.: Detecting group interest-level in meetings. In: IEEE ICASSP, pp. 489–492 (2010)
5.
Zurück zum Zitat Campbell, N.: An audio-visual approach to measuring discourse synchrony in multimodal conversation data. In: Henrichsen, P.J. (ed.) Linguistic Theory and Raw Sound. Samfundslitteratur, Frederiksberg (2010) Campbell, N.: An audio-visual approach to measuring discourse synchrony in multimodal conversation data. In: Henrichsen, P.J. (ed.) Linguistic Theory and Raw Sound. Samfundslitteratur, Frederiksberg (2010)
6.
Zurück zum Zitat Bonin, F., Bock, R., Campbell, N.: How do we react to context? annotation of individual and group engagement in a video corpus. In: Workshop on Context Based Affect Recognition, Held in conjunction with SocialCom, pp. 899–903 (2012) Bonin, F., Bock, R., Campbell, N.: How do we react to context? annotation of individual and group engagement in a video corpus. In: Workshop on Context Based Affect Recognition, Held in conjunction with SocialCom, pp. 899–903 (2012)
7.
Zurück zum Zitat Cabral, J. P., Renals, S., Richmond, K., Yamagishi, J.: Towards an improved modeling of the glottal source in statistical parametric speech synthesis. In: Workshop on Speech Synthesis, Germany (2007) Cabral, J. P., Renals, S., Richmond, K., Yamagishi, J.: Towards an improved modeling of the glottal source in statistical parametric speech synthesis. In: Workshop on Speech Synthesis, Germany (2007)
8.
Zurück zum Zitat Douxchamps, D., Campbell, N.: Robust real time face tracking for the analysis of human behaviour. In: Popescu-Belis, A., Renals, S., Bourlard, H. (eds.) MLMI 2007. LNCS, vol. 4892, pp. 1–10. Springer, Heidelberg (2008) CrossRef Douxchamps, D., Campbell, N.: Robust real time face tracking for the analysis of human behaviour. In: Popescu-Belis, A., Renals, S., Bourlard, H. (eds.) MLMI 2007. LNCS, vol. 4892, pp. 1–10. Springer, Heidelberg (2008) CrossRef
9.
Zurück zum Zitat Viola, P., Jones, M.: Robust real-time face detection. Int. J. Comput. Vis. 57(2), 137–154 (2004)CrossRef Viola, P., Jones, M.: Robust real-time face detection. Int. J. Comput. Vis. 57(2), 137–154 (2004)CrossRef
10.
Zurück zum Zitat Vaughan, B., Han, J.G., Gilmartin, E., Campbell, N.: Designing and implementing a platform for collecting multi-modal data of human-robot interaction. Acta Polytech. Hung. 9(1), 7–17 (2012) Vaughan, B., Han, J.G., Gilmartin, E., Campbell, N.: Designing and implementing a platform for collecting multi-modal data of human-robot interaction. Acta Polytech. Hung. 9(1), 7–17 (2012)
11.
Zurück zum Zitat Mandatory speech codec speech processing functions; Adaptive Multi-Rate (AMR) speech codec; Voice Activity Detector (VAD), ETSI Standard TS 126 194 V12.0.0 (2014) Mandatory speech codec speech processing functions; Adaptive Multi-Rate (AMR) speech codec; Voice Activity Detector (VAD), ETSI Standard TS 126 194 V12.0.0 (2014)
12.
Zurück zum Zitat Hang, J.G., Gilmartin, E., De Looze, C., Vaughan, B., Campbell, N.: Speech and multimodal resources: the herme database of spontaneous multimodal human-robot dialogues. In: LREC, pp. 1328–1331. Turkey (2012) Hang, J.G., Gilmartin, E., De Looze, C., Vaughan, B., Campbell, N.: Speech and multimodal resources: the herme database of spontaneous multimodal human-robot dialogues. In: LREC, pp. 1328–1331. Turkey (2012)
13.
Zurück zum Zitat Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA data mining software: an update. SIGKDD Explor. Newslett. 11, 10–18 (2009)CrossRef Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA data mining software: an update. SIGKDD Explor. Newslett. 11, 10–18 (2009)CrossRef
Metadaten
Titel
Towards Classification of Engagement in Human Interaction with Talking Robots
verfasst von
Yuyun Huang
Christy Elias
João P. Cabral
Atul Nautiyal
Christian Saam
Nick Campbell
Copyright-Jahr
2015
Verlag
Springer International Publishing
DOI
https://doi.org/10.1007/978-3-319-21380-4_125

Neuer Inhalt