Skip to main content
Erschienen in: Neural Processing Letters 1/2023

06.06.2022

Dual Attention Mechanism Based Outline Loss for Image Stylization

verfasst von: Pengqi Tu, Nong Sang

Erschienen in: Neural Processing Letters | Ausgabe 1/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Image stylization has attracted considerable attention from various fields. Although impressive results have been achieved, existing methods pay less attention to the preservation of outline and putting constraint on it when training, which makes generated images suffering from different degrees of distortion. To address this issue, we propose a dual attention mechanism based outline loss to enhance the restriction of outline consistency by incorporating an outline detection module and a dual attention module. Specifically, an outline detection module is used to detect outlines of the source image and the stylized image, which are further compared and enforced to be consistent with each other by a carefully-elaborated outline loss. Additionally, the dual attention module first guides the model to focus on regions of the source image whose style has the biggest difference from the target image during stylization based on the style attention feature map obtained by the auxiliary classifier. Then, an outline attention map is predicted to highlight regions where the outlines are prone to distort during stylization, which further facilitates the outline loss to execute stronger constraint on these regions. Experimental results show the superiority of our method compared to the existing state-of-the-art methods

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Arjovsky M, Chintala S, Bottou L (2017) “Wasserstein generative adversarial networks,” in International Conference on Machine Learning, pp. 214–223 Arjovsky M, Chintala S, Bottou L (2017) “Wasserstein generative adversarial networks,” in International Conference on Machine Learning, pp. 214–223
2.
3.
Zurück zum Zitat Brock A, Donahue J, Simonyan K (2019) “Large scale GAN training for high fidelity natural image synthesis,” in International Conference on Learning Representations Brock A, Donahue J, Simonyan K (2019) “Large scale GAN training for high fidelity natural image synthesis,” in International Conference on Learning Representations
5.
Zurück zum Zitat Cheng Y, Gan Z, Li Y, Liu J, Gao J (2018) “Sequential attention GAN for interactive image editing via dialogue,” arXiv:1812.08352 Cheng Y, Gan Z, Li Y, Liu J, Gao J (2018) “Sequential attention GAN for interactive image editing via dialogue,” arXiv:​1812.​08352
6.
Zurück zum Zitat Cho W, Bahng H, Park DK, Yoo S, Wu Z, Ma X, Choo J (2018) “Text2colors: Guiding image colorization through text-driven palette generation,” in European Conference on Computer Vision, pp. 431–447 Cho W, Bahng H, Park DK, Yoo S, Wu Z, Ma X, Choo J (2018) “Text2colors: Guiding image colorization through text-driven palette generation,” in European Conference on Computer Vision, pp. 431–447
7.
Zurück zum Zitat Choi Y, Choi M, Kim M, Ha J, Kim S, Choo J (2018) “Stargan: Unified generative adversarial networks for multi-domain image-to- image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789–8797 Choi Y, Choi M, Kim M, Ha J, Kim S, Choo J (2018) “Stargan: Unified generative adversarial networks for multi-domain image-to- image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789–8797
8.
Zurück zum Zitat Deshpande A, Lu J, Yeh M, Forsyth DA (2017) “Learning diverse image colorization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 6837–6845 Deshpande A, Lu J, Yeh M, Forsyth DA (2017) “Learning diverse image colorization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 6837–6845
9.
Zurück zum Zitat Dong C, Loy CC, He K, Tang X (2014)“Learning a deep convolu- tional network for image super-resolution,” in European Conference on Computer Vision, pp. 184–199 Dong C, Loy CC, He K, Tang X (2014)“Learning a deep convolu- tional network for image super-resolution,” in European Conference on Computer Vision, pp. 184–199
10.
Zurück zum Zitat Gatys LA, Ecker AS, Bethge M (2015) “Texture synthesis using convolutional neural networks,” in Neural Information Processing Systems, pp. 262–270 Gatys LA, Ecker AS, Bethge M (2015) “Texture synthesis using convolutional neural networks,” in Neural Information Processing Systems, pp. 262–270
11.
Zurück zum Zitat Gatys LA, Ecker AS, Bethge M (2016) “Image style transfer using convolutional neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414–2423 Gatys LA, Ecker AS, Bethge M (2016) “Image style transfer using convolutional neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414–2423
12.
Zurück zum Zitat Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, D. Warde-Farley, Ozair S, Courville AC, Bengio Y (2014) “Generative adversarial nets,” in Neural Information Processing Systems, pp. 2672–2680 Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, D. Warde-Farley, Ozair S, Courville AC, Bengio Y (2014) “Generative adversarial nets,” in Neural Information Processing Systems, pp. 2672–2680
13.
Zurück zum Zitat Gorijala M, Dukkipati A (2017) “Image generation and editing with variational info generative adversarial networks,” arXiv:1701.04568 Gorijala M, Dukkipati A (2017) “Image generation and editing with variational info generative adversarial networks,” arXiv:​1701.​04568
14.
Zurück zum Zitat Gu S, Chen C, Liao J, Yuan L (2018) “Arbitrary style transfer with deep feature reshuffle,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8222–8231 Gu S, Chen C, Liao J, Yuan L (2018) “Arbitrary style transfer with deep feature reshuffle,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8222–8231
15.
Zurück zum Zitat Hoffman J, Tzeng E, Park T, Zhu J, Isola P, Saenko K, Efros AA, Darrell T (2018) “Cycada: Cycle-consistent adversarial domain adaptation,” in International Conference on Machine Learnin, Hoffman J, Tzeng E, Park T, Zhu J, Isola P, Saenko K, Efros AA, Darrell T (2018) “Cycada: Cycle-consistent adversarial domain adaptation,” in International Conference on Machine Learnin,
16.
Zurück zum Zitat Huang X, Belongie SJ (2017) “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, pp. 1510–1519 Huang X, Belongie SJ (2017) “Arbitrary style transfer in real-time with adaptive instance normalization,” in IEEE International Conference on Computer Vision, pp. 1510–1519
17.
Zurück zum Zitat Huang X, Liu M, Belongie SJ, Kautz J (2018) “Multimodal unsupervised image-to-image translation,” in European Conference on Computer Vision, pp. 179–196 Huang X, Liu M, Belongie SJ, Kautz J (2018) “Multimodal unsupervised image-to-image translation,” in European Conference on Computer Vision, pp. 179–196
18.
Zurück zum Zitat Isola P, Zhu J, Zhou T, Efros AA (2017) “Image-to-image translation with conditional adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5967–5976 Isola P, Zhu J, Zhou T, Efros AA (2017) “Image-to-image translation with conditional adversarial networks,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5967–5976
20.
Zurück zum Zitat Johnson J, Alahi A, Li F (2016) “Perceptual losses for real-time style transfer and super-resolution,” in European Conference on Computer Vision, pp. 694–711 Johnson J, Alahi A, Li F (2016) “Perceptual losses for real-time style transfer and super-resolution,” in European Conference on Computer Vision, pp. 694–711
21.
Zurück zum Zitat Karras T, Aila T, Laine S, Lehtinen J (2018) “Progressive growing of gans for improved quality, stability, and variation,” in International Conference on Learning Representations Karras T, Aila T, Laine S, Lehtinen J (2018) “Progressive growing of gans for improved quality, stability, and variation,” in International Conference on Learning Representations
22.
Zurück zum Zitat Kim J, Kim M, Kang H, Lee K (2020) “U-GAT-IT:unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation,” in International Conference on Learning Representations Kim J, Kim M, Kang H, Lee K (2020) “U-GAT-IT:unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation,” in International Conference on Learning Representations
23.
Zurück zum Zitat Kotovenko D, Sanakoyeu A, Ma P, Lang S, and Ommer B (2019) “A content transformation block for image style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 10032–10041 Kotovenko D, Sanakoyeu A, Ma P, Lang S, and Ommer B (2019) “A content transformation block for image style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 10032–10041
24.
Zurück zum Zitat Ledig C, Theis L, Huszar F, Caballero J, Aitken AP, Tejani A, Totz J, Wang Z, Shi W (2017) “Photo-realistic single image super- resolution using a generative adversarial network,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681–4690 Ledig C, Theis L, Huszar F, Caballero J, Aitken AP, Tejani A, Totz J, Wang Z, Shi W (2017) “Photo-realistic single image super- resolution using a generative adversarial network,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681–4690
25.
Zurück zum Zitat Lee H, Tseng H, Huang J, Singh M, Yang M (2018) “Diverse image- to-image translation via disentangled representations,” in European Conference on Computer Vision, pp. 36–52 Lee H, Tseng H, Huang J, Singh M, Yang M (2018) “Diverse image- to-image translation via disentangled representations,” in European Conference on Computer Vision, pp. 36–52
26.
Zurück zum Zitat Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M (2017) “Universal style transfer via feature transforms,” in Neural Information Processing Systems, pp. 386–396 Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M (2017) “Universal style transfer via feature transforms,” in Neural Information Processing Systems, pp. 386–396
27.
Zurück zum Zitat Li X, Liu S, Kautz J, Yang M (2019) “Learning linear transformations for fast arbitrary style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 3809–3817 Li X, Liu S, Kautz J, Yang M (2019) “Learning linear transformations for fast arbitrary style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 3809–3817
28.
Zurück zum Zitat Liu M, Breuel T, Kautz J (2017) “Unsupervised image-to-image translation networks,” in Neural Information Processing Systems, pp. 700–708 Liu M, Breuel T, Kautz J (2017) “Unsupervised image-to-image translation networks,” in Neural Information Processing Systems, pp. 700–708
29.
Zurück zum Zitat Li C, Wand M (2016) “Combining markov random fields and convolu- tional neural networks for image synthesis,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2479–2486 Li C, Wand M (2016) “Combining markov random fields and convolu- tional neural networks for image synthesis,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2479–2486
31.
Zurück zum Zitat Nizan O, Tal A (2020) “Breaking the cycle - colleagues are all you need,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 7860–7869 Nizan O, Tal A (2020) “Breaking the cycle - colleagues are all you need,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 7860–7869
32.
33.
Zurück zum Zitat Shao X, Zhang W (2021) “SPatchGAN: a statistical feature based discriminator for unsupervised image-to-Image translation,” in IEEE International Conference on Computer Vision, pp. 6546-6555 Shao X, Zhang W (2021) “SPatchGAN: a statistical feature based discriminator for unsupervised image-to-Image translation,” in IEEE International Conference on Computer Vision, pp. 6546-6555
34.
Zurück zum Zitat Sheng L, Lin Z, Shao J, Wang X (2018)“Avatar-net: Multi-scale zero-shot style transfer by feature decoration,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8242–8250 Sheng L, Lin Z, Shao J, Wang X (2018)“Avatar-net: Multi-scale zero-shot style transfer by feature decoration,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8242–8250
35.
Zurück zum Zitat Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) “Rethinking the inception architecture for computer vision,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818–2826 Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) “Rethinking the inception architecture for computer vision,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818–2826
36.
Zurück zum Zitat Tomei M, Cornia M, Baraldi L, Cucchiara R (2019) “Art2real: Unfolding the reality of artworks via semantically-aware image-to-image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5849–5859 Tomei M, Cornia M, Baraldi L, Cucchiara R (2019) “Art2real: Unfolding the reality of artworks via semantically-aware image-to-image translation,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 5849–5859
37.
Zurück zum Zitat Ulyanov D, Lebedev V, Vedaldi A, Lempitsky VS (2016) “Texture networks: Feed-forward synthesis of textures and stylized images,” in International Conference on Machine Learning, pp. 1349–1357 Ulyanov D, Lebedev V, Vedaldi A, Lempitsky VS (2016) “Texture networks: Feed-forward synthesis of textures and stylized images,” in International Conference on Machine Learning, pp. 1349–1357
38.
Zurück zum Zitat Wang T, Liu M, Zhu J, Tao A, Kautz J, Catanzaro B (2018) “High- resolution image synthesis and semantic manipulation with conditional gans,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798–8807 Wang T, Liu M, Zhu J, Tao A, Kautz J, Catanzaro B (2018) “High- resolution image synthesis and semantic manipulation with conditional gans,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798–8807
39.
Zurück zum Zitat Xie S, Tu Z (2015) “Holistically-nested edge detection,” in IEEE International Conference on Computer Vision, pp. 1395–1403 Xie S, Tu Z (2015) “Holistically-nested edge detection,” in IEEE International Conference on Computer Vision, pp. 1395–1403
40.
Zurück zum Zitat Yao Y, Ren J, Xie X, Liu W, Liu Y, Wang J (2019) “Attention-aware multi-stroke style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467–1475 Yao Y, Ren J, Xie X, Liu W, Liu Y, Wang J (2019) “Attention-aware multi-stroke style transfer,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 1467–1475
41.
Zurück zum Zitat Yoo S, Bahng H, Chung S, Lee J, Chang J, Choo J (2019) “Col- oring with limited data: Few-shot colorization via memory-augmented networks,” in IEEE Conference on Computer Vision and Pattern Recog- nition, pp. 11283–11292 Yoo S, Bahng H, Chung S, Lee J, Chang J, Choo J (2019) “Col- oring with limited data: Few-shot colorization via memory-augmented networks,” in IEEE Conference on Computer Vision and Pattern Recog- nition, pp. 11283–11292
42.
Zurück zum Zitat Yu J, Rui Y, Tao D (2014) “Click prediction for web image reranking using multimodal sparse coding,” in IEEE Transactions on Image Processing, pp. 2019-2032 Yu J, Rui Y, Tao D (2014) “Click prediction for web image reranking using multimodal sparse coding,” in IEEE Transactions on Image Processing, pp. 2019-2032
43.
Zurück zum Zitat Yu J, Tan M, Zhang H, et al. (2019) “Hierarchical deep click feature prediction for fine-grained image recognition,” in IEEE transactions on pattern analysis and machine intelligence Yu J, Tan M, Zhang H, et al. (2019) “Hierarchical deep click feature prediction for fine-grained image recognition,” in IEEE transactions on pattern analysis and machine intelligence
44.
Zurück zum Zitat Zhao JJ, Mathieu M, LeCun Y (2017) “Energy-based generative adversarial networks,” in International Conference on Learning Representations Zhao JJ, Mathieu M, LeCun Y (2017) “Energy-based generative adversarial networks,” in International Conference on Learning Representations
45.
Zurück zum Zitat Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) “Learning deep features for discriminative localization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929 Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) “Learning deep features for discriminative localization,” in IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929
46.
Zurück zum Zitat Zhu J, Park T, Isola P, Efros AA (2017) “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE Inter- national Conference on Computer Vision, pp. 2242–2251 Zhu J, Park T, Isola P, Efros AA (2017) “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in IEEE Inter- national Conference on Computer Vision, pp. 2242–2251
Metadaten
Titel
Dual Attention Mechanism Based Outline Loss for Image Stylization
verfasst von
Pengqi Tu
Nong Sang
Publikationsdatum
06.06.2022
Verlag
Springer US
Erschienen in
Neural Processing Letters / Ausgabe 1/2023
Print ISSN: 1370-4621
Elektronische ISSN: 1573-773X
DOI
https://doi.org/10.1007/s11063-022-10896-5

Weitere Artikel der Ausgabe 1/2023

Neural Processing Letters 1/2023 Zur Ausgabe

Neuer Inhalt