skip to main content
10.1145/2663204.2663248acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
poster

Predicting Influential Statements in Group Discussions using Speech and Head Motion Information

Authors Info & Claims
Published:12 November 2014Publication History

ABSTRACT

Group discussions are used widely when generating new ideas and forming decisions as a group. Therefore, it is assumed that giving social influence to other members through facilitating the discussion is an important part of discussion skill. This study focuses on influential statements that affect discussion flow and highly related to facilitation, and aims to establish a model that predicts influential statements in group discussions. First, we collected a multimodal corpus using different group discussion tasks; in-basket and case-study. Based on schemes for analyzing arguments, each utterance was annotated as being influential or not. Then, we created classification models for predicting influential utterances using prosodic features as well as attention and head motion information from the speaker and other members of the group. In our model evaluation, we discovered that the assessment of each participant in terms of discussion facilitation skills by experienced observers correlated highly to the number of influential utterances by a given participant. This suggests that the proposed model can predict influential statements with considerable accuracy, and the prediction results can be a good predictor of facilitators in group discussions.

References

  1. Toulmin, S. E. 2003. The Uses of Argument: Cambridge University Press.Google ScholarGoogle Scholar
  2. Motowidlo, T. D. a. S. J. 1999. Why visual and vocal interview cues can affect interviewers' judgment and predict job performance, Journal of Applied Psychology, 84(6): p. 986--993.Google ScholarGoogle ScholarCross RefCross Ref
  3. Hareli, S. and Hess, U. 2010. What emotional reactions can tell us about the nature of others: An appraisal perspective on person perception, Cognition & Emotion, 24(1): p. 128--140.Google ScholarGoogle ScholarCross RefCross Ref
  4. Nguyen, L. S., et al. 2013, Multimodal analysis of body communication cues in employment interviews, in Proceedings of the 15th ACM on International conference on multimodal interaction, ACM: Sydney, Australia, p. 437--444. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Anderson, K., et al. 2013. The TARDIS Framework: Intelligent Virtual Agents for Social Coaching in Job Interviews, in Advances in Computer Entertainment. p. 476491.Google ScholarGoogle Scholar
  6. Jovanovic, N., Akker, R. o. d., and Nijholt, A. 2006 A corpus for studying addressing behavior in multi-party dialogues, in Language Resources and Evaluation. p. 5--23.Google ScholarGoogle Scholar
  7. Burger, S., MacLaren, V., and Yu, H. 2002, The ISL meeting corpus: the impact of meeting type on speech style, in International conference on spoken language processing, Interspeech-ICSLP.Google ScholarGoogle Scholar
  8. Sanchez-Cortes, D., et al. 2013. Emergent leaders through looking and speaking: from audio-visual data to multimodal recognition, Journal on Multimodal User Interfaces, 7(1--2): p. 39--53.Google ScholarGoogle ScholarCross RefCross Ref
  9. Otsuka, K., et al. 2006. Quantifying interpersonal influence in face-to-face conversations based on visual attention patterns, in CHI '06 Extended Abstracts on Human Factors in Computing Systems (CHI EA '06). ACM, New York, NY, USA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Basu, S., et al. 2001. Towards measuring human interactions in conversational settings, in IEEE CVPR Workshop on Cues in Communication.Google ScholarGoogle Scholar
  11. Dong, W., et al. 2007. Using the influence model to recognize functional roles in meetings, in the 9th ACM international conference on Multimodal interfaces. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Bales, R. F. 1970. Personality and interpersonal behavior.: New York: Holt, Rinehart and Winston.Google ScholarGoogle Scholar
  13. Rienks, R., et al. 2006. Detection and application of influence rankings in small group meetings, in ICMI '06 Proceedings of the 8th international conference on Multimodal interfaces (ICMI06), p. 257 - 264. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Hung, H., et al. 2008. Investigating Automatic Dominance Estimation in Groups From Visual Attention and Speaking Activity, in the 10th international conference on Multimodal interface (ICMI '08), p. 233--236. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Jayagopi, D. B., et al. 2009. Modeling Dominance in Group Conversations Using Nonverbal Activity Cues, IEEE Transactions on Audio, Speech, and Language Processing, 17(3): p. 501 - 513. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Escalera, S., et al. 2010. Automatic Detection of Dominance and Expected Interest, EURASIP Journal on Advances in Signal Processing, Volume 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Lepri, B., et al. 2012. Connecting meeting behavior with extraversion - a systematic study, IEEE Transactions on Affective Computing, 3(4): p. 443--455. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Staiano, J., et al. 2011, Automatic modeling of personality states in small group interactions, in ACM Multimedia, p. 989--992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Aran, O. and Gatica-Perez, D. 2013, One of a Kind: Inferring Personality Impressions in Meetings, in the 15th ACM on International conference on multimodal interaction (ICMI2013), p. 11--18. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Jayagopi, D. B., et al. 2012. Linking Speaking and Looking Behavior Patterns with Group Composition, Perception, and Performance, in the 14th ACM international conference on Multimodal interaction (ICMI2012), p. 433--440. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Gordon, T. F. and Karacapilidis, N. I. 1997, The Zeno Argumentation Framework, in International Conference on Artificial Intelligence and Law, p. 10--18. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Verheij, B. 2003. Artificial argument assistance for defeasible argumentation, Artificial Intelligence, 150(1--2): p. 291--324. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Prakken, H. 2008. A formal model of adjudication dialogues, Artificial Intelligence and Law, 16(3): p. 305--328. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Georgila, K., et al. 2011, An Annotation Scheme for CrossCultural Argumentation and Persuasion Dialogues, in the SIGDIAL 2011 Conference, p. 272--278. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Kickul, J. and Neuman, G. 2000. Emergent Leadership Behaviors: The Function of Personality and Cognitive Ability in Determining Teamwork Performance and KSAs, Journal of Business and Psychology, 15(1): p. 27--51.Google ScholarGoogle ScholarCross RefCross Ref
  26. Julius http://julius.sourceforge.jp/.Google ScholarGoogle Scholar
  27. Pianesi, F., et al. 2008. Multimodal support to group dynamics, Personal and Ubiquitous Computing, 12: p. 181195. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Mast, M. S. 2002. Dominance as expressed and inferred through speaking time, Human Communication Research, 28(3): p. 420--450.Google ScholarGoogle Scholar
  29. Praat http://www.fon.hum.uva.nl/praat/.Google ScholarGoogle Scholar
  30. Sanchez-Cortes, D., et al. 2010. Identifying Emergent Leadership in Small Groups using Nonverbal Communicative Cues, in International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction (ICMI-MLMI'10). Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Predicting Influential Statements in Group Discussions using Speech and Head Motion Information

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        ICMI '14: Proceedings of the 16th International Conference on Multimodal Interaction
        November 2014
        558 pages
        ISBN:9781450328852
        DOI:10.1145/2663204

        Copyright © 2014 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 12 November 2014

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • poster

        Acceptance Rates

        ICMI '14 Paper Acceptance Rate51of127submissions,40%Overall Acceptance Rate453of1,080submissions,42%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader