Skip to main content
Top

2021 | OriginalPaper | Chapter

Applying Delaunay Triangulation Augmentation for Deep Learning Facial Expression Generation and Recognition

Authors : Hristo Valev, Alessio Gallucci, Tim Leufkens, Joyce Westerink, Corina Sas

Published in: Pattern Recognition. ICPR International Workshops and Challenges

Publisher: Springer International Publishing

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Generating and recognizing facial expressions has numerous applications, however, those are limited by the scarcity of datasets containing labeled nuanced expressions. In this paper, we describe the use of Delaunay triangulation combined with simple morphing techniques to blend images of faces, which allows us to create and automatically label facial expressions portraying controllable intensities of emotion. We have applied this approach on the RafD dataset consisting of 67 participants and 8 categorical emotions and evaluated the augmentation in a facial expression generation and recognition tasks using deep learning models. For the generation task, we used a deconvolution neural network which learns to encode the input images in a high-dimensional feature space and generate realistic expressions at varying intensities. The augmentation significantly improves the quality of images compared to previous comparable experiments and it allows to create images with a higher resolution. For the recognition task, we evaluated pre-trained Densenet121 and Resnet50 networks with either the original or augmented dataset. Our results indicate that the augmentation alone has a similar or better performance compared to the original. Implications of this method and its role in improving existing facial expression generation and recognition approaches are discussed.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Cogsdill, E.J., Todorov, A.T., Spelke, E.S., Banaji, M.R.: Inferring character from faces: a developmental study. Psychol. Sci. 25(5), 1132–1139 (2014)CrossRef Cogsdill, E.J., Todorov, A.T., Spelke, E.S., Banaji, M.R.: Inferring character from faces: a developmental study. Psychol. Sci. 25(5), 1132–1139 (2014)CrossRef
2.
go back to reference Buck, R.: Social and emotional functions in facial expression and communication: the readout hypothesis. Biol. Psychol. 38(2–3), 95–115 (1994)CrossRef Buck, R.: Social and emotional functions in facial expression and communication: the readout hypothesis. Biol. Psychol. 38(2–3), 95–115 (1994)CrossRef
3.
go back to reference Kruszka, P., et al.: 22q11. 2 deletion syndrome in diverse populations. Am. J. Med. Genet. Part A 173(4), 879–888 (2017)CrossRef Kruszka, P., et al.: 22q11. 2 deletion syndrome in diverse populations. Am. J. Med. Genet. Part A 173(4), 879–888 (2017)CrossRef
4.
go back to reference Chang, K.I., Bowyer, K.W., Flynn, P.J.: Multiple nose region matching for 3D face recognition under varying facial expression. IEEE Trans. Pattern Anal. Mach. Intell. 28(10), 1695–1700 (2006)CrossRef Chang, K.I., Bowyer, K.W., Flynn, P.J.: Multiple nose region matching for 3D face recognition under varying facial expression. IEEE Trans. Pattern Anal. Mach. Intell. 28(10), 1695–1700 (2006)CrossRef
6.
go back to reference Sanches, P., et al.: HCI and affective health: taking stock of a decade of studies and charting future research directions. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–17 (2019) Sanches, P., et al.: HCI and affective health: taking stock of a decade of studies and charting future research directions. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–17 (2019)
7.
go back to reference Colombo, D., et al.: The need for change: understanding emotion regulation antecedents and consequences using ecological momentary assessment. Emotion 20(1), 30 (2020)CrossRef Colombo, D., et al.: The need for change: understanding emotion regulation antecedents and consequences using ecological momentary assessment. Emotion 20(1), 30 (2020)CrossRef
8.
go back to reference Alfaras, M., et al.: From biodata to somadata. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–14 (2020) Alfaras, M., et al.: From biodata to somadata. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–14 (2020)
9.
go back to reference Hu, L., et al.: Avatar digitization from a single image for real-time rendering. ACM Trans. Graph. (ToG) 36(6), 1–14 (2017)CrossRef Hu, L., et al.: Avatar digitization from a single image for real-time rendering. ACM Trans. Graph. (ToG) 36(6), 1–14 (2017)CrossRef
10.
go back to reference Gallucci, A., Znamenskiy, D., Petkovic, M.: Prediction of 3D body parts from face shape and anthropometric measurements. J. Image Graph. 8(3), 67–74 (2020)CrossRef Gallucci, A., Znamenskiy, D., Petkovic, M.: Prediction of 3D body parts from face shape and anthropometric measurements. J. Image Graph. 8(3), 67–74 (2020)CrossRef
11.
go back to reference Lombardi, S., Saragih, J., Simon, T., Sheikh, Y.: Deep appearance models for face rendering. ACM Trans. Graph. (TOG) 37(4), 1–13 (2018)CrossRef Lombardi, S., Saragih, J., Simon, T., Sheikh, Y.: Deep appearance models for face rendering. ACM Trans. Graph. (TOG) 37(4), 1–13 (2018)CrossRef
13.
go back to reference Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 8789–8797 (2018) Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 8789–8797 (2018)
14.
go back to reference Ding, H., Sricharan, K., Chellappa, R.: Exprgan: Facial expression editing with controllable expression intensity. arXiv preprint arXiv:1709.03842 (2017) Ding, H., Sricharan, K., Chellappa, R.: Exprgan: Facial expression editing with controllable expression intensity. arXiv preprint arXiv:​1709.​03842 (2017)
15.
go back to reference Pumarola, A., Agudo, A., Martinez, A.M., Sanfeliu, A., Moreno-Noguer, F.: Ganimation: anatomically-aware facial animation from a single image. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 818–833 (2018) Pumarola, A., Agudo, A., Martinez, A.M., Sanfeliu, A., Moreno-Noguer, F.: Ganimation: anatomically-aware facial animation from a single image. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 818–833 (2018)
16.
go back to reference Rychlowska, M., Jack, R.E., Garrod, O.G., Schyns, P.G., Martin, J.D., Niedenthal, P.M.: Functional smiles: tools for love, sympathy, and war. Psychol. Sci. 28(9), 1259–1270 (2017)CrossRef Rychlowska, M., Jack, R.E., Garrod, O.G., Schyns, P.G., Martin, J.D., Niedenthal, P.M.: Functional smiles: tools for love, sympathy, and war. Psychol. Sci. 28(9), 1259–1270 (2017)CrossRef
17.
go back to reference Yitzhak, N., et al.: Gently does it: humans outperform a software classifier in recognizing subtle, nonstereotypical facial expressions. Emotion 17(8), 1187 (2017)CrossRef Yitzhak, N., et al.: Gently does it: humans outperform a software classifier in recognizing subtle, nonstereotypical facial expressions. Emotion 17(8), 1187 (2017)CrossRef
19.
go back to reference Langner, O., Dotsch, R., Bijlstra, G., Wigboldus, D.H., Hawk, S.T., Van Knippenberg, A.: Presentation and validation of the radboud faces database. Cogn. Emotion 24(8), 1377–1388 (2010)CrossRef Langner, O., Dotsch, R., Bijlstra, G., Wigboldus, D.H., Hawk, S.T., Van Knippenberg, A.: Presentation and validation of the radboud faces database. Cogn. Emotion 24(8), 1377–1388 (2010)CrossRef
20.
go back to reference Lee, D.-T., Schachter, B.J.: Two algorithms for constructing a delaunay triangulation. Int. J. Comput. Inf. Sci. 9(3), 219–242 (1980)MathSciNetCrossRef Lee, D.-T., Schachter, B.J.: Two algorithms for constructing a delaunay triangulation. Int. J. Comput. Inf. Sci. 9(3), 219–242 (1980)MathSciNetCrossRef
21.
go back to reference Dosovitskiy, A., Springenberg, J.T., Tatarchenko, M., Brox, T.: Learning to generate chairs, tables and cars with convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(4), 692–705 (2016) Dosovitskiy, A., Springenberg, J.T., Tatarchenko, M., Brox, T.: Learning to generate chairs, tables and cars with convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(4), 692–705 (2016)
22.
go back to reference Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017) Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017)
23.
go back to reference He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016) He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
24.
go back to reference Ravuri, S., Vinyals, O.: Classification accuracy score for conditional generative models. In: Advances in Neural Information Processing Systems, pp. 12268–12279 (2019) Ravuri, S., Vinyals, O.: Classification accuracy score for conditional generative models. In: Advances in Neural Information Processing Systems, pp. 12268–12279 (2019)
25.
go back to reference Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211–252 (2015)MathSciNetCrossRef Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211–252 (2015)MathSciNetCrossRef
Metadata
Title
Applying Delaunay Triangulation Augmentation for Deep Learning Facial Expression Generation and Recognition
Authors
Hristo Valev
Alessio Gallucci
Tim Leufkens
Joyce Westerink
Corina Sas
Copyright Year
2021
DOI
https://doi.org/10.1007/978-3-030-68796-0_53

Premium Partner