Skip to main content
Erschienen in: Cognitive Computation 2/2021

29.06.2020

Affect Recognition for Multimodal Natural Language Processing

verfasst von: Soujanya Poria, Ong Yew Soon, Bing Liu, Lidong Bing

Erschienen in: Cognitive Computation | Ausgabe 2/2021

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Excerpt

Language is inherently multimodal. It has many forms of appearance, like speech, gestures, facial expressions, and head-nods. In an ideal human-machine conversational system, machines should understand this multimodal language. Understanding human language also largely depends on the machines’ ability to interpret emotions. Emotional sensitivity can prevent desultory answers provided by these machines, thus making conversations more natural and engaging. For us humans, emotions aid our learning, communication, and decision-making. Hence, over the past two decades, there has been a significant effort to incorporate cognitive capabilities into machines so that they can interpret, comprehend and express emotions. Computational analysis of human multimodal language is an emerging research area in natural language processing (NLP). It expands the horizons of NLP to study language used in face to face communication and in online multimedia. This form of language contains modalities of language (in terms of spoken text), visual (in terms of gestures and facial expressions) and acoustic (in terms of changes in the voice tone). At its core, this research area is focused on modeling the three modalities and their complex interactions. This special issue on Affect Recognition in Human Multimodal Language aims to facilitate the growth of this new research direction in the community. The challenges of modeling human multimodal language can be split into two major categories: (1) studying each modality individually and modeling each in a manner that can be linked to other modalities (also known as intramodal dynamics, (2) linking the modalities by modeling the interactions between them (also known as intermodal dynamics). Common forms of these interactions include complementary or correlated information across modes. Intrinsic to each modality, modeling human multimodal language is complex due to factors such as idiosyncrasy in communicative styles, non-trivial alignment between modalities and unreliable or contradictory information across modalities. Therefore, computational analysis of multimodal language becomes a challenging research area. This special issue aimed at bringing together contributions from both academics and practitioners in the context of affect recognition in multimodal language with a primary focus on:
1.
Multimodal affect recognition in monologues,
 
2.
Effective multimodal fusion.
 
3.
Detecting affect in dyadic and multiparty multimodal conversations. Detecting affect in conversations is more challenging than monologues. This is primarily due to the presence of complex inter-dependency between speaker states in the conversation,
 

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Metadaten
Titel
Affect Recognition for Multimodal Natural Language Processing
verfasst von
Soujanya Poria
Ong Yew Soon
Bing Liu
Lidong Bing
Publikationsdatum
29.06.2020
Verlag
Springer US
Erschienen in
Cognitive Computation / Ausgabe 2/2021
Print ISSN: 1866-9956
Elektronische ISSN: 1866-9964
DOI
https://doi.org/10.1007/s12559-020-09738-0

Weitere Artikel der Ausgabe 2/2021

Cognitive Computation 2/2021 Zur Ausgabe