Skip to main content
Top
Published in: Neural Computing and Applications 14/2020

21-10-2019 | Original Article

Generative adversarial networks with denoising penalty and sample augmentation

Authors: Yan Gan, Kedi Liu, Mao Ye, Yuxiao Zhang, Yang Qian

Published in: Neural Computing and Applications | Issue 14/2020

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

For the original generative adversarial networks (GANs) model, there are three problems that (1) the generator is not robust to the input random noise; (2) the discriminating ability of discriminator gradually reduces in the later stage of training; and (3) it is difficult to reach at the theoretical Nash equilibrium point in the process of training. To solve the above problems, in this paper, a GANs model with denoising penalty and sample augmentation is proposed. In this model, a denoising constraint is firstly designed as the penalty term of the generator, which minimizes the F-norm between the input noise and the encoding of the image generated by the corresponding perturbed noise, respectively. The generator is forced to learn more robust invariant characteristics. Secondly, we put forward a sample augmentation discriminator to improve the ability of discriminator, which is trained by mixing the generated and real images as training samples. Thirdly, in order to achieve the theoretical optimization as far as possible, our model combines denoising penalty and sample augmentation discriminator. Then, denoising penalty and sample augmentation discriminator are applied to five different GANs models whose loss functions include the original GANs, Hinge and least squares loss. Finally, experimental results on the LSUN and CelebA datasets show that our proposed method can help the baseline models improve the quality of generated images.

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Goodfellow I, Pouget-Abadie J, Mirza M (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672–2680 Goodfellow I, Pouget-Abadie J, Mirza M (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672–2680
3.
4.
go back to reference Saito Y, Takamichi S, Saruwatari H (2018) Statistical parametric speech synthesis incorporating generative adversarial networks. IEEE/ACM Trans Audio Speech Lang Process 26(1):84–96CrossRef Saito Y, Takamichi S, Saruwatari H (2018) Statistical parametric speech synthesis incorporating generative adversarial networks. IEEE/ACM Trans Audio Speech Lang Process 26(1):84–96CrossRef
5.
go back to reference Bollepalli B, Lauri J, Paavo A (2019) Generative adversarial network-based glottal waveform model for statistical parametric speech synthesis. arXiv preprint arXiv:1903.05955 Bollepalli B, Lauri J, Paavo A (2019) Generative adversarial network-based glottal waveform model for statistical parametric speech synthesis. arXiv preprint arXiv:​1903.​05955
6.
go back to reference Dai B, Fidler S, Urtasun R (2017) Towards diverse and natural imagedescriptions via a conditional GAN. In: Proceedings of the IEEE international conference on computer vision, pp 2970–2979 Dai B, Fidler S, Urtasun R (2017) Towards diverse and natural imagedescriptions via a conditional GAN. In: Proceedings of the IEEE international conference on computer vision, pp 2970–2979
7.
9.
go back to reference Pan Y, Zhaofan Q, Ting Y, Houqiang L, Tao M (2017) To create what you tell: generating videos from captions. In: Proceedings of the 25th ACM international conference on Multimedia, pp 1789–1798 Pan Y, Zhaofan Q, Ting Y, Houqiang L, Tao M (2017) To create what you tell: generating videos from captions. In: Proceedings of the 25th ACM international conference on Multimedia, pp 1789–1798
10.
go back to reference Zhang J, Shu Y, Xu S (2018) Sparsely grouped multi-task generative adversarial networks for facial attribute manipulation. arXiv preprint arXiv:1805.07509 Zhang J, Shu Y, Xu S (2018) Sparsely grouped multi-task generative adversarial networks for facial attribute manipulation. arXiv preprint arXiv:​1805.​07509
11.
go back to reference Pumarola A, Antonio A, Martinez AM, Sanfeliu A, Moreno-Noguer F (2018) Ganimation: anatomically-aware facial animation from a single image. In: Proceedings of the European conference on computer vision, pp 818–833 Pumarola A, Antonio A, Martinez AM, Sanfeliu A, Moreno-Noguer F (2018) Ganimation: anatomically-aware facial animation from a single image. In: Proceedings of the European conference on computer vision, pp 818–833
12.
go back to reference Li T, Qian R, Dong C (2018) Beautygan: instance-level facial makeup transfer with deep generative adversarial network. In: ACM multimedia conference on multimedia conference, pp 645–653 Li T, Qian R, Dong C (2018) Beautygan: instance-level facial makeup transfer with deep generative adversarial network. In: ACM multimedia conference on multimedia conference, pp 645–653
13.
go back to reference Schlegl T, Seebock P, Waldstein SM (2017) Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In: International conference on information processing in medical imaging, pp 146–157 Schlegl T, Seebock P, Waldstein SM (2017) Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In: International conference on information processing in medical imaging, pp 146–157
14.
go back to reference Nagarajan V, Kolter JZ (2017) Gradient descent GAN optimization is locally stable. In: Advances in neural information processing systems, pp 5585–5595 Nagarajan V, Kolter JZ (2017) Gradient descent GAN optimization is locally stable. In: Advances in neural information processing systems, pp 5585–5595
15.
go back to reference Odena A, Buckman J, Olsson C (2018) Is generator conditioning causally related to GAN performance. In: International conference on machine learning, pp 3846–3855 Odena A, Buckman J, Olsson C (2018) Is generator conditioning causally related to GAN performance. In: International conference on machine learning, pp 3846–3855
16.
go back to reference Warde-Farley D, Bengio Y (2016) Improving generative adversarial networks with denoising feature matching Warde-Farley D, Bengio Y (2016) Improving generative adversarial networks with denoising feature matching
17.
go back to reference Zhang H, Sun Y, Liu L, Wang X, Li L, Liu W (2018) ClothingOut: a category-supervised GAN model for clothing segmentation and retrieval. In: Neural computing and applications, pp 1–12 Zhang H, Sun Y, Liu L, Wang X, Li L, Liu W (2018) ClothingOut: a category-supervised GAN model for clothing segmentation and retrieval. In: Neural computing and applications, pp 1–12
18.
go back to reference Gulrajani I, Ahmed F, Arjovsky M (2017) Improved training of wasserstein GANs. In: Advances in neural information processing systems, pp 5767–5777 Gulrajani I, Ahmed F, Arjovsky M (2017) Improved training of wasserstein GANs. In: Advances in neural information processing systems, pp 5767–5777
19.
21.
go back to reference Salimans T, Goodfellow I, Zaremba W (2016) Improved techniques for training GANs. In: Advances in neural information processing systems, pp 2234–2242 Salimans T, Goodfellow I, Zaremba W (2016) Improved techniques for training GANs. In: Advances in neural information processing systems, pp 2234–2242
22.
go back to reference Nam S, Kim Y, Kim SJ (2018) Text-adaptive generative adversarial networks: manipulating images with natural language. In: Advances in neural information processing systems, pp 42–51 Nam S, Kim Y, Kim SJ (2018) Text-adaptive generative adversarial networks: manipulating images with natural language. In: Advances in neural information processing systems, pp 42–51
23.
go back to reference Liu L, Zhang H, Ji Y, Wu QJ (2019) Toward AI fashion design: an attribute-GAN model for clothing match. Neurocomputing 341:156–167CrossRef Liu L, Zhang H, Ji Y, Wu QJ (2019) Toward AI fashion design: an attribute-GAN model for clothing match. Neurocomputing 341:156–167CrossRef
25.
go back to reference Wang X, Girshick R, Gupta A (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7794–7803 Wang X, Girshick R, Gupta A (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7794–7803
26.
go back to reference Xu T, Zhang P, Huang Q (2018) AttnGAN: fine-grained text to image generation with attentional generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1316–1324 Xu T, Zhang P, Huang Q (2018) AttnGAN: fine-grained text to image generation with attentional generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1316–1324
29.
go back to reference Salimans T, Karpathy A, Chen X (2017) Pixelcnn++: improving the pixelcnn with discretized logistic mixture likelihood and other modifications. arXiv preprint arXiv:1701.05517 Salimans T, Karpathy A, Chen X (2017) Pixelcnn++: improving the pixelcnn with discretized logistic mixture likelihood and other modifications. arXiv preprint arXiv:​1701.​05517
30.
go back to reference Radford A, Metz L, Chintala S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 Radford A, Metz L, Chintala S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:​1511.​06434
32.
go back to reference Zhang H, Xu T, Li H (2018) StackGAN++: Realistic image synthesis with stacked generative adversarial networks. In: IEEE transactions on pattern analysis and machine intelligence Zhang H, Xu T, Li H (2018) StackGAN++: Realistic image synthesis with stacked generative adversarial networks. In: IEEE transactions on pattern analysis and machine intelligence
33.
go back to reference Zhu J Y, Park T, Isola P (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223–2232 Zhu J Y, Park T, Isola P (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223–2232
34.
go back to reference Gan Y, Gong J, Ye M (2018) Unpaired cross domain image translation with augmented auxiliary domain information. Neurocomputing 316:112–123CrossRef Gan Y, Gong J, Ye M (2018) Unpaired cross domain image translation with augmented auxiliary domain information. Neurocomputing 316:112–123CrossRef
35.
go back to reference Yi Z, Zhang H, Tan P (2017) DualGAN: unsupervised dual learning for image-to-image translation. In: Proceedings of the IEEE international conference on computer vision, pp 2849–2857 Yi Z, Zhang H, Tan P (2017) DualGAN: unsupervised dual learning for image-to-image translation. In: Proceedings of the IEEE international conference on computer vision, pp 2849–2857
36.
go back to reference Bau D, Zhu JY, Strobelt H (2018) GAN dissection: visualizing and understanding generative adversarial networks. arXiv preprint arXiv:1811.10597 Bau D, Zhu JY, Strobelt H (2018) GAN dissection: visualizing and understanding generative adversarial networks. arXiv preprint arXiv:​1811.​10597
37.
go back to reference Yu J, Lin Z, Yang J (2018) Generative image inpainting with contextual attention. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5505–5514 Yu J, Lin Z, Yang J (2018) Generative image inpainting with contextual attention. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5505–5514
38.
go back to reference Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Cheung X (2016) Improved techniques for training GANs. In: Advances in neural information processing systems, pp 2234–2242 Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Cheung X (2016) Improved techniques for training GANs. In: Advances in neural information processing systems, pp 2234–2242
39.
go back to reference Hou M, Chaib-Draa B, Li C, Zhao Q (2018) Generative adversarial positive-unlabeled learning. In: Proceedings of the 27th international joint conference on artificial intelligence, pp 2255–2261 Hou M, Chaib-Draa B, Li C, Zhao Q (2018) Generative adversarial positive-unlabeled learning. In: Proceedings of the 27th international joint conference on artificial intelligence, pp 2255–2261
40.
go back to reference Mao X, Li Q, Xie H (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794–2802 Mao X, Li Q, Xie H (2017) Least squares generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2794–2802
42.
go back to reference Yu F, Seff A, Zhang Y (2015) LSUN: construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365 Yu F, Seff A, Zhang Y (2015) LSUN: construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:​1506.​03365
43.
go back to reference Liu Z, Luo P, Wang X (2015) Deep learning face attributes in the wild. In: Proceedings of the IEEE international conference on computer vision, pp 3730–3738 Liu Z, Luo P, Wang X (2015) Deep learning face attributes in the wild. In: Proceedings of the IEEE international conference on computer vision, pp 3730–3738
44.
go back to reference Heusel M, Ramsauer H, Unterthiner T (2017) Gans trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in neural information processing systems, pp 6626–6637 Heusel M, Ramsauer H, Unterthiner T (2017) Gans trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in neural information processing systems, pp 6626–6637
45.
go back to reference Choi Y, Choi M, Kim M (2018) StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8789–8797 Choi Y, Choi M, Kim M (2018) StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8789–8797
Metadata
Title
Generative adversarial networks with denoising penalty and sample augmentation
Authors
Yan Gan
Kedi Liu
Mao Ye
Yuxiao Zhang
Yang Qian
Publication date
21-10-2019
Publisher
Springer London
Published in
Neural Computing and Applications / Issue 14/2020
Print ISSN: 0941-0643
Electronic ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-019-04526-w

Other articles of this Issue 14/2020

Neural Computing and Applications 14/2020 Go to the issue

Premium Partner