Skip to main content
Top
Published in: Neural Processing Letters 1/2023

06-06-2022

Dual Attention Mechanism Based Outline Loss for Image Stylization

Authors: Pengqi Tu, Nong Sang

Published in: Neural Processing Letters | Issue 1/2023

Log in

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

Image stylization has attracted considerable attention from various fields. Although impressive results have been achieved, existing methods pay less attention to the preservation of outline and putting constraint on it when training, which makes generated images suffering from different degrees of distortion. To address this issue, we propose a dual attention mechanism based outline loss to enhance the restriction of outline consistency by incorporating an outline detection module and a dual attention module. Specifically, an outline detection module is used to detect outlines of the source image and the stylized image, which are further compared and enforced to be consistent with each other by a carefully-elaborated outline loss. Additionally, the dual attention module first guides the model to focus on regions of the source image whose style has the biggest difference from the target image during stylization based on the style attention feature map obtained by the auxiliary classifier. Then, an outline attention map is predicted to highlight regions where the outlines are prone to distort during stylization, which further facilitates the outline loss to execute stronger constraint on these regions. Experimental results show the superiority of our method compared to the existing state-of-the-art methods

Dont have a licence yet? Then find out more about our products and how to get one now:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literature
1.
go back to reference Arjovsky M, Chintala S, Bottou L (2017) “Wasserstein generative adversarial networks,” in International Conference on Machine Learning, pp. 214–223 Arjovsky M, Chintala S, Bottou L (2017) “Wasserstein generative adversarial networks,” in International Conference on Machine Learning, pp. 214–223
3.
go back to reference Brock A, Donahue J, Simonyan K (2019) “Large scale GAN training for high fidelity natural image synthesis,” in International Conference on Learning Representations Brock A, Donahue J, Simonyan K (2019) “Large scale GAN training for high fidelity natural image synthesis,” in International Conference on Learning Representations
5.
6.
go back to reference Cho W, Bahng H, Park DK, Yoo S, Wu Z, Ma X, Choo J (2018) “Text2colors: Guiding image colorization through text-driven palette generation,” in European Conference on Computer Vision, pp. 431–447 Cho W, Bahng H, Park DK, Yoo S, Wu Z, Ma X, Choo J (2018) “Text2colors: Guiding image colorization through text-driven palette generation,” in European Conference on Computer Vision, pp. 431–447
7.
go back to reference Choi Y, Choi M, Kim M, Ha J, Kim S, Choo J (2018) “Stargan: Unified generative adversarial networks for multi-domain image-to- image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789–8797 Choi Y, Choi M, Kim M, Ha J, Kim S, Choo J (2018) “Stargan: Unified generative adversarial networks for multi-domain image-to- image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789–8797
8.
go back to reference Deshpande A, Lu J, Yeh M, Forsyth DA (2017) “Learning diverse image colorization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 6837–6845 Deshpande A, Lu J, Yeh M, Forsyth DA (2017) “Learning diverse image colorization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 6837–6845
9.
go back to reference Dong C, Loy CC, He K, Tang X (2014)“Learning a deep convolu- tional network for image super-resolution,” in European Conference on Computer Vision, pp. 184–199 Dong C, Loy CC, He K, Tang X (2014)“Learning a deep convolu- tional network for image super-resolution,” in European Conference on Computer Vision, pp. 184–199
10.
go back to reference Gatys LA, Ecker AS, Bethge M (2015) “Texture synthesis using convolutional neural networks,” in Neural Information Processing Systems, pp. 262–270 Gatys LA, Ecker AS, Bethge M (2015) “Texture synthesis using convolutional neural networks,” in Neural Information Processing Systems, pp. 262–270
11.
go back to reference Gatys LA, Ecker AS, Bethge M (2016) “Image style transfer using convolutional neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414–2423 Gatys LA, Ecker AS, Bethge M (2016) “Image style transfer using convolutional neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414–2423
12.
go back to reference Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, D. Warde-Farley, Ozair S, Courville AC, Bengio Y (2014) “Generative adversarial nets,” in Neural Information Processing Systems, pp. 2672–2680 Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, D. Warde-Farley, Ozair S, Courville AC, Bengio Y (2014) “Generative adversarial nets,” in Neural Information Processing Systems, pp. 2672–2680
13.
go back to reference Gorijala M, Dukkipati A (2017) “Image generation and editing with variational info generative adversarial networks,” arXiv:1701.04568 Gorijala M, Dukkipati A (2017) “Image generation and editing with variational info generative adversarial networks,” arXiv:​1701.​04568
14.
go back to reference Gu S, Chen C, Liao J, Yuan L (2018) “Arbitrary style transfer with deep feature reshuffle,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8222–8231 Gu S, Chen C, Liao J, Yuan L (2018) “Arbitrary style transfer with deep feature reshuffle,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8222–8231
15.
go back to reference Hoffman J, Tzeng E, Park T, Zhu J, Isola P, Saenko K, Efros AA, Darrell T (2018) “Cycada: Cycle-consistent adversarial domain adaptation,” in International Conference on Machine Learnin, Hoffman J, Tzeng E, Park T, Zhu J, Isola P, Saenko K, Efros AA, Darrell T (2018) “Cycada: Cycle-consistent adversarial domain adaptation,” in International Conference on Machine Learnin,
16.
go back to reference Huang X, Belongie SJ (2017) “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, pp. 1510–1519 Huang X, Belongie SJ (2017) “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, pp. 1510–1519
17.
go back to reference Huang X, Liu M, Belongie SJ, Kautz J (2018) “Multimodal unsupervised image-to-image translation,” in European Conference on Computer Vision, pp. 179–196 Huang X, Liu M, Belongie SJ, Kautz J (2018) “Multimodal unsupervised image-to-image translation,” in European Conference on Computer Vision, pp. 179–196
18.
go back to reference Isola P, Zhu J, Zhou T, Efros AA (2017) “Image-to-image translation with conditional adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5967–5976 Isola P, Zhu J, Zhou T, Efros AA (2017) “Image-to-image translation with conditional adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5967–5976
20.
go back to reference Johnson J, Alahi A, Li F (2016) “Perceptual losses for real-time style transfer and super-resolution,” in European Conference on Computer Vision, pp. 694–711 Johnson J, Alahi A, Li F (2016) “Perceptual losses for real-time style transfer and super-resolution,” in European Conference on Computer Vision, pp. 694–711
21.
go back to reference Karras T, Aila T, Laine S, Lehtinen J (2018) “Progressive growing of gans for improved quality, stability, and variation,” in International Conference on Learning Representations Karras T, Aila T, Laine S, Lehtinen J (2018) “Progressive growing of gans for improved quality, stability, and variation,” in International Conference on Learning Representations
22.
go back to reference Kim J, Kim M, Kang H, Lee K (2020) “U-GAT-IT:unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation,” in International Conference on Learning Representations Kim J, Kim M, Kang H, Lee K (2020) “U-GAT-IT:unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation,” in International Conference on Learning Representations
23.
go back to reference Kotovenko D, Sanakoyeu A, Ma P, Lang S, and Ommer B (2019) “A content transformation block for image style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 10032–10041 Kotovenko D, Sanakoyeu A, Ma P, Lang S, and Ommer B (2019) “A content transformation block for image style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 10032–10041
24.
go back to reference Ledig C, Theis L, Huszar F, Caballero J, Aitken AP, Tejani A, Totz J, Wang Z, Shi W (2017) “Photo-realistic single image super- resolution using a generative adversarial network,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681–4690 Ledig C, Theis L, Huszar F, Caballero J, Aitken AP, Tejani A, Totz J, Wang Z, Shi W (2017) “Photo-realistic single image super- resolution using a generative adversarial network,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681–4690
25.
go back to reference Lee H, Tseng H, Huang J, Singh M, Yang M (2018) “Diverse image- to-image translation via disentangled representations,” in European Conference on Computer Vision, pp. 36–52 Lee H, Tseng H, Huang J, Singh M, Yang M (2018) “Diverse image- to-image translation via disentangled representations,” in European Conference on Computer Vision, pp. 36–52
26.
go back to reference Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M (2017) “Universal style transfer via feature transforms,” in Neural Information Processing Systems, pp. 386–396 Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M (2017) “Universal style transfer via feature transforms,” in Neural Information Processing Systems, pp. 386–396
27.
go back to reference Li X, Liu S, Kautz J, Yang M (2019) “Learning linear transformations for fast arbitrary style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 3809–3817 Li X, Liu S, Kautz J, Yang M (2019) “Learning linear transformations for fast arbitrary style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 3809–3817
28.
go back to reference Liu M, Breuel T, Kautz J (2017) “Unsupervised image-to-image translation networks,” in Neural Information Processing Systems, pp. 700–708 Liu M, Breuel T, Kautz J (2017) “Unsupervised image-to-image translation networks,” in Neural Information Processing Systems, pp. 700–708
29.
go back to reference Li C, Wand M (2016) “Combining markov random fields and convolu- tional neural networks for image synthesis,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2479–2486 Li C, Wand M (2016) “Combining markov random fields and convolu- tional neural networks for image synthesis,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2479–2486
31.
go back to reference Nizan O, Tal A (2020) “Breaking the cycle - colleagues are all you need,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 7860–7869 Nizan O, Tal A (2020) “Breaking the cycle - colleagues are all you need,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 7860–7869
32.
33.
go back to reference Shao X, Zhang W (2021) “SPatchGAN: a statistical feature based discriminator for unsupervised image-to-Image translation,” in IEEE International Conference on Computer Vision, pp. 6546-6555 Shao X, Zhang W (2021) “SPatchGAN: a statistical feature based discriminator for unsupervised image-to-Image translation,” in IEEE International Conference on Computer Vision, pp. 6546-6555
34.
go back to reference Sheng L, Lin Z, Shao J, Wang X (2018)“Avatar-net: Multi-scale zero-shot style transfer by feature decoration,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8242–8250 Sheng L, Lin Z, Shao J, Wang X (2018)“Avatar-net: Multi-scale zero-shot style transfer by feature decoration,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8242–8250
35.
go back to reference Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) “Rethinking the inception architecture for computer vision,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818–2826 Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) “Rethinking the inception architecture for computer vision,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818–2826
36.
go back to reference Tomei M, Cornia M, Baraldi L, Cucchiara R (2019) “Art2real: Unfolding the reality of artworks via semantically-aware image-to-image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5849–5859 Tomei M, Cornia M, Baraldi L, Cucchiara R (2019) “Art2real: Unfolding the reality of artworks via semantically-aware image-to-image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5849–5859
37.
go back to reference Ulyanov D, Lebedev V, Vedaldi A, Lempitsky VS (2016) “Texture networks: Feed-forward synthesis of textures and stylized images,” in International Conference on Machine Learning, pp. 1349–1357 Ulyanov D, Lebedev V, Vedaldi A, Lempitsky VS (2016) “Texture networks: Feed-forward synthesis of textures and stylized images,” in International Conference on Machine Learning, pp. 1349–1357
38.
go back to reference Wang T, Liu M, Zhu J, Tao A, Kautz J, Catanzaro B (2018) “High- resolution image synthesis and semantic manipulation with conditional gans,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798–8807 Wang T, Liu M, Zhu J, Tao A, Kautz J, Catanzaro B (2018) “High- resolution image synthesis and semantic manipulation with conditional gans,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798–8807
39.
go back to reference Xie S, Tu Z (2015) “Holistically-nested edge detection,” in IEEE International Conference on Computer Vision, pp. 1395–1403 Xie S, Tu Z (2015) “Holistically-nested edge detection,” in IEEE International Conference on Computer Vision, pp. 1395–1403
40.
go back to reference Yao Y, Ren J, Xie X, Liu W, Liu Y, Wang J (2019) “Attention-aware multi-stroke style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467–1475 Yao Y, Ren J, Xie X, Liu W, Liu Y, Wang J (2019) “Attention-aware multi-stroke style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467–1475
41.
go back to reference Yoo S, Bahng H, Chung S, Lee J, Chang J, Choo J (2019) “Col- oring with limited data: Few-shot colorization via memory-augmented networks,” in IEEE Conference on Computer Vision and Pattern Recog- nition, pp. 11283–11292 Yoo S, Bahng H, Chung S, Lee J, Chang J, Choo J (2019) “Col- oring with limited data: Few-shot colorization via memory-augmented networks,” in IEEE Conference on Computer Vision and Pattern Recog- nition, pp. 11283–11292
42.
go back to reference Yu J, Rui Y, Tao D (2014) “Click prediction for web image reranking using multimodal sparse coding,” in IEEE Transactions on Image Processing, pp. 2019-2032 Yu J, Rui Y, Tao D (2014) “Click prediction for web image reranking using multimodal sparse coding,” in IEEE Transactions on Image Processing, pp. 2019-2032
43.
go back to reference Yu J, Tan M, Zhang H, et al. (2019) “Hierarchical deep click feature prediction for fine-grained image recognition,” in IEEE transactions on pattern analysis and machine intelligence Yu J, Tan M, Zhang H, et al. (2019) “Hierarchical deep click feature prediction for fine-grained image recognition,” in IEEE transactions on pattern analysis and machine intelligence
44.
go back to reference Zhao JJ, Mathieu M, LeCun Y (2017) “Energy-based generative adversarial networks,” in International Conference on Learning Representations Zhao JJ, Mathieu M, LeCun Y (2017) “Energy-based generative adversarial networks,” in International Conference on Learning Representations
45.
go back to reference Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) “Learning deep features for discriminative localization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929 Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) “Learning deep features for discriminative localization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929
46.
go back to reference Zhu J, Park T, Isola P, Efros AA (2017) “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE Inter- national Conference on Computer Vision, pp. 2242–2251 Zhu J, Park T, Isola P, Efros AA (2017) “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE Inter- national Conference on Computer Vision, pp. 2242–2251
Metadata
Title
Dual Attention Mechanism Based Outline Loss for Image Stylization
Authors
Pengqi Tu
Nong Sang
Publication date
06-06-2022
Publisher
Springer US
Published in
Neural Processing Letters / Issue 1/2023
Print ISSN: 1370-4621
Electronic ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-022-10896-5

Other articles of this Issue 1/2023

Neural Processing Letters 1/2023 Go to the issue