Skip to main content
Erschienen in: International Journal on Interactive Design and Manufacturing (IJIDeM) 4/2023

15.03.2023 | Original Paper

Using RGBD cameras for classifying learning and teacher interaction through postural attitude

verfasst von: Mauricio Hincapié, Christian Andrés Díaz, Alejandro Valencia-Arias, David Güemes-Castorena, Manuel Contero

Erschienen in: International Journal on Interactive Design and Manufacturing (IJIDeM) | Ausgabe 4/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Globally the improvement and evaluation of the academic performance of students has been a priority, however the way in which the quality of learning process is evaluated within the classroom is based on traditional methods such as grades or perception surveys. Additionally, measuring continuously the performance of the student, teacher and its interaction in the classroom is difficult because there are several internal and external factors that can affect the pedagogical practice in the classroom or e-learning environments, and currently, their effects are not completely understood. Currently, advances in motion tracking through low cost devices such RGBD cameras allows the real-time monitoring of persons posture inside closed spaces such a classroom. Some research projects have associated posture with affective and cognitive state, but as far as we know none have proposed an approach to classify learning and teacher interaction using posture. An approach that uses a set of performance metrics of the student and teacher, in order to classify whether learning and teacher-student interaction was successful is developed and tested. This was an experimental design using an experimental and control group in order to evaluate if it is possible to classify between poor and good interaction between teacher and student. The results showed that it is possible to classify between poor and good interaction between teacher and student, besides the best method of classification is the approach based on neural networks with an accuracy of 76%. The proposed approach could classify whether an interaction between the student and the teacher was good or not. The results showed that the best method of classification was the approach based on neural networks with an accuracy of 78%.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
2.
Zurück zum Zitat Ministerio de Educación Nacional (2008) ¿Qué y cómo mejorar a partir de las pruebas PISA?(No.44) [Online]. Available: mineducacion.gov.co/1621/article-162392.html Ministerio de Educación Nacional (2008) ¿Qué y cómo mejorar a partir de las pruebas PISA?(No.44) [Online]. Available: mineducacion.gov.co/1621/article-162392.html
10.
Zurück zum Zitat Salcedo, A.: Deserción Universitaria en Colombia. Acad. Virtual. 3(1), 1–9 (2010)MathSciNet Salcedo, A.: Deserción Universitaria en Colombia. Acad. Virtual. 3(1), 1–9 (2010)MathSciNet
11.
Zurück zum Zitat Gaviria, A. and Barrientos, J.: Determinantes de la calidad de la educación en Colombia. Archivos de Economía, Departamento Nacional de Planeación, no. 159 (2001) Gaviria, A. and Barrientos, J.: Determinantes de la calidad de la educación en Colombia. Archivos de Economía, Departamento Nacional de Planeación, no. 159 (2001)
12.
Zurück zum Zitat Hiebert, J., Morris, A., Berk, D., Jansen, A.: Preparing teachers to learn from teaching. J. Teach. Educ. 58(1), 47–54 (2007)CrossRef Hiebert, J., Morris, A., Berk, D., Jansen, A.: Preparing teachers to learn from teaching. J. Teach. Educ. 58(1), 47–54 (2007)CrossRef
13.
Zurück zum Zitat Ernest, P., Guitert, M., Hampel, R., Heiser, S., Hopkins, J., Murphy, L., Stickler, U.: Online teacher development: collaborating in a virtual learning environment. Comput. Assist. Lang. Learn. 26(4), 311–333 (2013)CrossRef Ernest, P., Guitert, M., Hampel, R., Heiser, S., Hopkins, J., Murphy, L., Stickler, U.: Online teacher development: collaborating in a virtual learning environment. Comput. Assist. Lang. Learn. 26(4), 311–333 (2013)CrossRef
16.
Zurück zum Zitat Chen, C.-C., Huang, T.-C.: Learning in a u-museum: developing a context-aware ubiquitous learning environment. Comput. Educ. 59(3), 873–883 (2012)CrossRef Chen, C.-C., Huang, T.-C.: Learning in a u-museum: developing a context-aware ubiquitous learning environment. Comput. Educ. 59(3), 873–883 (2012)CrossRef
17.
Zurück zum Zitat Papatheodorou, C., Antoniou, G., and Bikakis, A.: On the deployment of contextual reasoning in ambient intelligence environments. In: 6th International conference on intelligent environments (EI), pp. 13–18, 2010. Papatheodorou, C., Antoniou, G., and Bikakis, A.: On the deployment of contextual reasoning in ambient intelligence environments. In: 6th International conference on intelligent environments (EI), pp. 13–18, 2010.
23.
Zurück zum Zitat Torres, D.M.A.: Los smart classroom y ambientes de aprendizaje híbridos ¿dónde está la innovación? Sol de Aquino 21, 58–61 (2022) Torres, D.M.A.: Los smart classroom y ambientes de aprendizaje híbridos ¿dónde está la innovación? Sol de Aquino 21, 58–61 (2022)
24.
Zurück zum Zitat Kapoor, A., Picard, R. W., and Ivanov, Y.: Probabilistic combination of multiple modalities to detect interest. In: 17th ICPR, vol. 3, pp. 969–972 (2004) Kapoor, A., Picard, R. W., and Ivanov, Y.: Probabilistic combination of multiple modalities to detect interest. In: 17th ICPR, vol. 3, pp. 969–972 (2004)
25.
Zurück zum Zitat Craig, S. D., D'Mello, S. K., Gholson, B., Witherspoon, A., Sullins, J., and Graesser, A. C.: Emotions during learning: the first steps toward an affect sensitive intelligent tutoring system. In: J. Nall & R. (2004) Craig, S. D., D'Mello, S. K., Gholson, B., Witherspoon, A., Sullins, J., and Graesser, A. C.: Emotions during learning: the first steps toward an affect sensitive intelligent tutoring system. In: J. Nall & R. (2004)
26.
Zurück zum Zitat Stevenson, A., Bailenson, J., Stathatos, S.: Automatically detected nonverbal behavior predicts creativity in collaborating dyads. J. Nonverbal Behav. 38, 389–408 (2014)CrossRef Stevenson, A., Bailenson, J., Stathatos, S.: Automatically detected nonverbal behavior predicts creativity in collaborating dyads. J. Nonverbal Behav. 38, 389–408 (2014)CrossRef
27.
Zurück zum Zitat Mota, S.: Automated posture analysis for detecting learner's affective state. In: PhD thesis, School of Architecture and Planning, MIT (2002) Mota, S.: Automated posture analysis for detecting learner's affective state. In: PhD thesis, School of Architecture and Planning, MIT (2002)
29.
Zurück zum Zitat Margetis, G. et al.: Towards ambient intelligence in the classroom. In: Proceedings 6th UAHCI - Volume Part IV, pp. 577–586 (2011) Margetis, G. et al.: Towards ambient intelligence in the classroom. In: Proceedings 6th UAHCI - Volume Part IV, pp. 577–586 (2011)
30.
Zurück zum Zitat Antona, M. et al.: A student-centric intelligent classroom. J Ambient Intell pp. 248–252 (2011) Antona, M. et al.: A student-centric intelligent classroom. J Ambient Intell pp. 248–252 (2011)
31.
Zurück zum Zitat Xie, W. et al.: Smart classroom - an intelligent environment for tele-education. In: Proceedings 2nd IEEE PCM, pp. 662–668 (2001) Xie, W. et al.: Smart classroom - an intelligent environment for tele-education. In: Proceedings 2nd IEEE PCM, pp. 662–668 (2001)
32.
Zurück zum Zitat Diaz, C., Hincapié, M., Moreno, G., and Guerra, W.: Descripción de una Aula de Clase Inteligente e Implementación de su Componente de Interacción. In: Proceedings SICC (2015) Diaz, C., Hincapié, M., Moreno, G., and Guerra, W.: Descripción de una Aula de Clase Inteligente e Implementación de su Componente de Interacción. In: Proceedings SICC (2015)
33.
Zurück zum Zitat Niebles, J., Chen, C.-W., and Fei-Fei, L.: Modeling temporal structure of decomposable motion segments for activity classification. In: Proceedings 11th ECCV, Part II, pp. 392–405 (2010) Niebles, J., Chen, C.-W., and Fei-Fei, L.: Modeling temporal structure of decomposable motion segments for activity classification. In: Proceedings 11th ECCV, Part II, pp. 392–405 (2010)
34.
Zurück zum Zitat Gligoric, N., Uzelac, A., and Krco, S.: Smart classroom: real-time feedback on lecture quality. In: IEEE PERCOM, pp. 391–394 (2012) Gligoric, N., Uzelac, A., and Krco, S.: Smart classroom: real-time feedback on lecture quality. In: IEEE PERCOM, pp. 391–394 (2012)
36.
Zurück zum Zitat Huang, L.-S., Su, J.-Y., Pao, T.-L.: A context aware smart classroom architecture for smart campuses. Appl. Sci. 9(9), 1837 (2019)CrossRef Huang, L.-S., Su, J.-Y., Pao, T.-L.: A context aware smart classroom architecture for smart campuses. Appl. Sci. 9(9), 1837 (2019)CrossRef
37.
Zurück zum Zitat Hong, J., Suh, E., Kim, S.: Context-aware systems: a literature review and classification. Expert Syst. Appl. 36(4), 8509–8522 (2009)CrossRef Hong, J., Suh, E., Kim, S.: Context-aware systems: a literature review and classification. Expert Syst. Appl. 36(4), 8509–8522 (2009)CrossRef
38.
Zurück zum Zitat Zhang, Y., Li, X., Zhu, L., Dong, X., and Hao, Q.: What is a smart classroom? A literature review. In: Shaping Future Schools with Digital Technology: An International Handbook, pp. 25–40 (2019) Zhang, Y., Li, X., Zhu, L., Dong, X., and Hao, Q.: What is a smart classroom? A literature review. In: Shaping Future Schools with Digital Technology: An International Handbook, pp. 25–40 (2019)
41.
Zurück zum Zitat Fischer, I. A., Pereira, E. P., Lopes, V., and Medina, R. D.: Improving efficiency and availability in smart classroom environments. In: 2019 IEEE 16th ICNSC pp. 52–56, (2019) Fischer, I. A., Pereira, E. P., Lopes, V., and Medina, R. D.: Improving efficiency and availability in smart classroom environments. In: 2019 IEEE 16th ICNSC pp. 52–56, (2019)
43.
Zurück zum Zitat Martin, J.-C., Devillers, L.: A Multimodal Corpus Approach for the Study of Spontaneous Emotions. Springer, Cham (2009)CrossRef Martin, J.-C., Devillers, L.: A Multimodal Corpus Approach for the Study of Spontaneous Emotions. Springer, Cham (2009)CrossRef
44.
Zurück zum Zitat Zeng, Z., Pantic, M., et al.: A survey of affect recognition methods: audio, visual and spontaneous expressions. IEEE PAMI 31(1), 39–58 (2009)CrossRef Zeng, Z., Pantic, M., et al.: A survey of affect recognition methods: audio, visual and spontaneous expressions. IEEE PAMI 31(1), 39–58 (2009)CrossRef
45.
Zurück zum Zitat Ekman, P. and Friesen, W.: Manual for the facial action coding system. Environ. Psychol. Nonverbal Behav. (1978) Ekman, P. and Friesen, W.: Manual for the facial action coding system. Environ. Psychol. Nonverbal Behav. (1978)
46.
Zurück zum Zitat Kleinsmith, A., Bianchi-Berthouze, N., Steed, A.: Automatic recognition of non-acted affective postures. IEEE Trans. Syst. Man Cybern. Syst. 41(4), 837–847 (2011) Kleinsmith, A., Bianchi-Berthouze, N., Steed, A.: Automatic recognition of non-acted affective postures. IEEE Trans. Syst. Man Cybern. Syst. 41(4), 837–847 (2011)
47.
Zurück zum Zitat Kleinsmith, A., and Bianchi-Berthouze, N.: Recognizing affective dimensions from body posture. In: Proceedings 2nd ACII, pp. 48–58 (2007) Kleinsmith, A., and Bianchi-Berthouze, N.: Recognizing affective dimensions from body posture. In: Proceedings 2nd ACII, pp. 48–58 (2007)
48.
Zurück zum Zitat Glowinski, D., Dael, N., Camurri, A., Volpe, G., Mortillaro, M., Scherer, K.: Towards a minimal representation of affective gestures. IEEE Trans. Affect. Comput. 2(2), 106–118 (2011)CrossRef Glowinski, D., Dael, N., Camurri, A., Volpe, G., Mortillaro, M., Scherer, K.: Towards a minimal representation of affective gestures. IEEE Trans. Affect. Comput. 2(2), 106–118 (2011)CrossRef
49.
Zurück zum Zitat Kleinsmith, A., De Silva, P., Bianchi-Berthouze, N.: Cross-cultural differences in recognizing affect from body posture. Interact. Comput. 18(6), 1371–1389 (2006)CrossRef Kleinsmith, A., De Silva, P., Bianchi-Berthouze, N.: Cross-cultural differences in recognizing affect from body posture. Interact. Comput. 18(6), 1371–1389 (2006)CrossRef
50.
Zurück zum Zitat Castellano, G., Pereira, A., Paiva, A.: Automatic analysis of affective postures and body motion to detect engagement with a game companion. IEEE Trans. Affect. Comput. 2(3), 119–128 (2011) Castellano, G., Pereira, A., Paiva, A.: Automatic analysis of affective postures and body motion to detect engagement with a game companion. IEEE Trans. Affect. Comput. 2(3), 119–128 (2011)
51.
Zurück zum Zitat Gunes, H., and Piccardi, M.: Fusing face and body display for bi-modal emotion recognition: single frame analysis and multi-frame post-integration. In: Proceedings 1st ACII, pp. 102–111 (2005) Gunes, H., and Piccardi, M.: Fusing face and body display for bi-modal emotion recognition: single frame analysis and multi-frame post-integration. In: Proceedings 1st ACII, pp. 102–111 (2005)
52.
Zurück zum Zitat Omlor, L., and Giese, M.: Unsupervised Learning of Spatio-temporal Primitives of Emotional Gait. In: Perception and interactive technologies: international tutorial and research workshop, PIT 2006 Kloster Irsee, Germany, June 19-21, Proceedings (pp. 188-192), Springer (2006) Omlor, L., and Giese, M.: Unsupervised Learning of Spatio-temporal Primitives of Emotional Gait. In: Perception and interactive technologies: international tutorial and research workshop, PIT 2006 Kloster Irsee, Germany, June 19-21, Proceedings (pp. 188-192), Springer (2006)
53.
Zurück zum Zitat Clark, R., Pua, Y.-H., Oliveira, C., Bower, K., Thilarajah, S., McGaw, R., Hasanki, K., Mentiplay, B.: Reliability and concurrent validity of the Microsoft Xbox One Kinect for assessment of standing balance and postural control. Gait Posture 42(2), 210–213 (2015)CrossRef Clark, R., Pua, Y.-H., Oliveira, C., Bower, K., Thilarajah, S., McGaw, R., Hasanki, K., Mentiplay, B.: Reliability and concurrent validity of the Microsoft Xbox One Kinect for assessment of standing balance and postural control. Gait Posture 42(2), 210–213 (2015)CrossRef
54.
Zurück zum Zitat Diego-Mas, J., Alocaide-Marzal, J.: Using Kinect sensor in observational methods for assessing postures at work. Appl. Ergon 25(4), 976–985 (2014)CrossRef Diego-Mas, J., Alocaide-Marzal, J.: Using Kinect sensor in observational methods for assessing postures at work. Appl. Ergon 25(4), 976–985 (2014)CrossRef
55.
Zurück zum Zitat Xu, X., McGorry, R.: The validity of the first and second generation Microsoft Kinect for identifying joint center locations during static postures. Appl. Ergon 49, 47–54 (2015)CrossRef Xu, X., McGorry, R.: The validity of the first and second generation Microsoft Kinect for identifying joint center locations during static postures. Appl. Ergon 49, 47–54 (2015)CrossRef
56.
Zurück zum Zitat Su, C.-J., Chiang, C.-Y., Huang, J.-Y.: Kinect-enabled home-based rehabilitation system using dynamic time warping and fuzzy logic. Appl. Soft Comput. 22, 652–666 (2014)CrossRef Su, C.-J., Chiang, C.-Y., Huang, J.-Y.: Kinect-enabled home-based rehabilitation system using dynamic time warping and fuzzy logic. Appl. Soft Comput. 22, 652–666 (2014)CrossRef
57.
Zurück zum Zitat Ojuno, K., Yamashita, T., Fukui, H., Noridomi, S., Arata, K., and Yamauchi, Y.: Body posture and face orientation estimation by convolutional network with heterogenous learning. In: Proceedings IWAIT (2018) Ojuno, K., Yamashita, T., Fukui, H., Noridomi, S., Arata, K., and Yamauchi, Y.: Body posture and face orientation estimation by convolutional network with heterogenous learning. In: Proceedings IWAIT (2018)
58.
Zurück zum Zitat Mahata, J., and Phadikar, A.: Recent advances in human behaviour understanding: a survey. In: Proceedings DevIC (2017) Mahata, J., and Phadikar, A.: Recent advances in human behaviour understanding: a survey. In: Proceedings DevIC (2017)
59.
Zurück zum Zitat Kobayashi, H. and Hara, F.: The recognition of basic facial expressions by neural network. In: Proceedings IJCNN, pp. 460–466 (1991) Kobayashi, H. and Hara, F.: The recognition of basic facial expressions by neural network. In: Proceedings IJCNN, pp. 460–466 (1991)
60.
Zurück zum Zitat Kessous, A., Amir, S., Dubnov, G., and Reichart, D.: Multimodal emotion recognition from expressive faces, body gestures and speech. In: Artificial intelligence and innovations 2007: from theory to applications, vol. 247, Boston: Springer, pp. 375–388 (2007) Kessous, A., Amir, S., Dubnov, G., and Reichart, D.: Multimodal emotion recognition from expressive faces, body gestures and speech. In: Artificial intelligence and innovations 2007: from theory to applications, vol. 247, Boston: Springer, pp. 375–388 (2007)
61.
Zurück zum Zitat Xavier-Junior et al.: Introducing affective agents in recommendation systems based on relational data clustering. In: Proceedings 22nd DEXA, Part II, Springer-Verlag Berlin Heidelberg, LNCS 6861, pp. 303–310 (2011) Xavier-Junior et al.: Introducing affective agents in recommendation systems based on relational data clustering. In: Proceedings 22nd DEXA, Part II, Springer-Verlag Berlin Heidelberg, LNCS 6861, pp. 303–310 (2011)
62.
Zurück zum Zitat Samad, A. et al.: Adaptive-neuro fuzzy inference system for human posture classification using a simplified shock graph. In: Proceedings 4th IVIC, Springer-Verlag Berlin Heidelberg, LNCS 5857, pp. 585–595 (2009) Samad, A. et al.: Adaptive-neuro fuzzy inference system for human posture classification using a simplified shock graph. In: Proceedings 4th IVIC, Springer-Verlag Berlin Heidelberg, LNCS 5857, pp. 585–595 (2009)
63.
Zurück zum Zitat Rezazadeh, I., Firoozabadi, S.: A novel human–machine interface based on recognition of multi-channel facial bioelectric signals. Aust. Phys. Eng. Sci. Med. 34, 497–513 (2011)CrossRef Rezazadeh, I., Firoozabadi, S.: A novel human–machine interface based on recognition of multi-channel facial bioelectric signals. Aust. Phys. Eng. Sci. Med. 34, 497–513 (2011)CrossRef
64.
Zurück zum Zitat Ren, H. and Xu, G.: Human action recognition in smart room. In: IEEE FG, Washington, DC, USA (2002) Ren, H. and Xu, G.: Human action recognition in smart room. In: IEEE FG, Washington, DC, USA (2002)
65.
Zurück zum Zitat Di Mitri, D., Schneider, J., Specht, M., Drachsler, H.: From signals to knowledge: a conceptual model for multimodal learning analytics. J. Comput. Assist. Learn. 34(4), 338–349 (2018)CrossRef Di Mitri, D., Schneider, J., Specht, M., Drachsler, H.: From signals to knowledge: a conceptual model for multimodal learning analytics. J. Comput. Assist. Learn. 34(4), 338–349 (2018)CrossRef
66.
Zurück zum Zitat Raca, M. and Dillenbourg, P.: Holistic analysis of the classroom. In: Proceedings of the 2014 ACM, pp. 13–20 (2014) Raca, M. and Dillenbourg, P.: Holistic analysis of the classroom. In: Proceedings of the 2014 ACM, pp. 13–20 (2014)
67.
Zurück zum Zitat Bosch, N.: Multimodal affect detection in the wild: accuracy, availability, and generalizability. In: Proceedings of the 2015 ACM ICMI, pp. 645–649 (2015) Bosch, N.: Multimodal affect detection in the wild: accuracy, availability, and generalizability. In: Proceedings of the 2015 ACM ICMI, pp. 645–649 (2015)
68.
Zurück zum Zitat Sanghvi, J., Castellano, G., Leite, L., Pereira, A., McOwan, P.W., Paiva, A.: Automatic analysis of affective postures and body motion to detect engagement with a game companion. IEEE Trans. Syst. Man Cybern. 41(6), 1064–1077 (2011) Sanghvi, J., Castellano, G., Leite, L., Pereira, A., McOwan, P.W., Paiva, A.: Automatic analysis of affective postures and body motion to detect engagement with a game companion. IEEE Trans. Syst. Man Cybern. 41(6), 1064–1077 (2011)
69.
Zurück zum Zitat Piana, S., and Staglianò, A.: A set of full-body movement features for emotion recognition to help children affected by autism spectrum condition. InfoMus Lab (2009) Piana, S., and Staglianò, A.: A set of full-body movement features for emotion recognition to help children affected by autism spectrum condition. InfoMus Lab (2009)
70.
Zurück zum Zitat Nixon, M. and Howard, A.: Applying gaming principles to virtual environments for upper extremity therapy games. In: Proceedings of the 7th ICDVRAT (2013) Nixon, M. and Howard, A.: Applying gaming principles to virtual environments for upper extremity therapy games. In: Proceedings of the 7th ICDVRAT (2013)
71.
Zurück zum Zitat Castellano, G., Villalba, S., and Camurri, A.: Recognising human emotions from body movement and gesture dynamics. In: Proceedings ACII (2007) Castellano, G., Villalba, S., and Camurri, A.: Recognising human emotions from body movement and gesture dynamics. In: Proceedings ACII (2007)
72.
Zurück zum Zitat Castellano, G., Villalba, S., and Camurri, A.: Recognising human emotions from body movement and gesture dynamics, affective computing and intelligent interaction. In: Proceedings ACII 2007, pp. 71-82 (2007) Castellano, G., Villalba, S., and Camurri, A.: Recognising human emotions from body movement and gesture dynamics, affective computing and intelligent interaction. In: Proceedings ACII 2007, pp. 71-82 (2007)
73.
Zurück zum Zitat Mancini, M., Castellano, G., Bevacqua, E., and Peters, C.: Copying behaviour of expressive motion. In: Computer vision/computer graphics collaboration techniques: third international conference, MIRAGE 2007, Rocquencourt, France, March 28-30, Proceedings 3, pp. 549–553 (2007) Mancini, M., Castellano, G., Bevacqua, E., and Peters, C.: Copying behaviour of expressive motion. In: Computer vision/computer graphics collaboration techniques: third international conference, MIRAGE 2007, Rocquencourt, France, March 28-30, Proceedings 3, pp. 549–553 (2007)
Metadaten
Titel
Using RGBD cameras for classifying learning and teacher interaction through postural attitude
verfasst von
Mauricio Hincapié
Christian Andrés Díaz
Alejandro Valencia-Arias
David Güemes-Castorena
Manuel Contero
Publikationsdatum
15.03.2023
Verlag
Springer Paris
Erschienen in
International Journal on Interactive Design and Manufacturing (IJIDeM) / Ausgabe 4/2023
Print ISSN: 1955-2513
Elektronische ISSN: 1955-2505
DOI
https://doi.org/10.1007/s12008-023-01262-3

Weitere Artikel der Ausgabe 4/2023

International Journal on Interactive Design and Manufacturing (IJIDeM) 4/2023 Zur Ausgabe