Skip to main content
Erschienen in: Medical & Biological Engineering & Computing 12/2022

20.10.2022 | Original Article

MsRAN: a multi-scale residual attention network for multi-model image fusion

verfasst von: Jing Wang, Long Yu, Shengwei Tian

Erschienen in: Medical & Biological Engineering & Computing | Ausgabe 12/2022

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Fusion is a critical step in image processing tasks. Recently, deep learning networks have been considerably applied in information fusion. But the significant limitation of existing image fusion methods is the inability to highlight typical regions of the source image and retain sufficient useful information. To address the problem, the paper proposes a multi-scale residual attention network (MsRAN) to fully exploit the image feature. Its generator network contains two information refinement networks and one information integration network. The information refinement network extracts feature at different scales using convolution kernels of different sizes. The information integration network, with a merging block and an attention block added, prevents the underutilization of information in the intermediate layers and forces the generator to focus on salient regions in multi-modal source images. Furthermore, in the phase of model training, we add an information loss function and adopt a dual adversarial structure, enabling the model to capture more details. Qualitative and quantitative experiments on publicly available datasets validate that the proposed method provides better visual results than other methods and retains more detail information.

Graphical abstract

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Dogra A, Goyal B, Agrawal S (2017) From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE Access 5:16040–16067CrossRef Dogra A, Goyal B, Agrawal S (2017) From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE Access 5:16040–16067CrossRef
2.
Zurück zum Zitat Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Information Fusion 45:153–178CrossRef Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Information Fusion 45:153–178CrossRef
3.
Zurück zum Zitat Li W, Peng X, Fu J, Wang G, Huang Y, Chao F (2022) A multiscale double-branch residual attention network for anatomical–functional medical image fusion. Comp Biol Med 141:105005CrossRef Li W, Peng X, Fu J, Wang G, Huang Y, Chao F (2022) A multiscale double-branch residual attention network for anatomical–functional medical image fusion. Comp Biol Med 141:105005CrossRef
4.
Zurück zum Zitat Li Q, Lu L, Li Z, Wu W, Liu Z, Jeon G, Yang X (2019) Coupled GAN with relativistic discriminators for infrared and visible images fusion. IEEE Sensors J 21(6):7458–7467CrossRef Li Q, Lu L, Li Z, Wu W, Liu Z, Jeon G, Yang X (2019) Coupled GAN with relativistic discriminators for infrared and visible images fusion. IEEE Sensors J 21(6):7458–7467CrossRef
5.
Zurück zum Zitat Li J et al (2019) Poisson reconstruction-based fusion of infrared and visible images via saliency detection. IEEE Access 7:20676–20688CrossRef Li J et al (2019) Poisson reconstruction-based fusion of infrared and visible images via saliency detection. IEEE Access 7:20676–20688CrossRef
6.
Zurück zum Zitat Xiang T, Yan Li, Gao R (2015) A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain. Infrared Phys Technol 69:53–61CrossRef Xiang T, Yan Li, Gao R (2015) A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain. Infrared Phys Technol 69:53–61CrossRef
7.
Zurück zum Zitat Naidu VPS (2011) Image fusion technique using multi-resolution singular value decomposition. Def Sci J 61(5):479CrossRef Naidu VPS (2011) Image fusion technique using multi-resolution singular value decomposition. Def Sci J 61(5):479CrossRef
8.
Zurück zum Zitat Zhang Q et al (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review. Information Fusion 40:57–75CrossRef Zhang Q et al (2018) Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: a review. Information Fusion 40:57–75CrossRef
9.
Zurück zum Zitat Mou J, Gao W, Song Z (2013) Image fusion based on non-negative matrix factorization and infrared feature extraction. 2013 6th International Congress on Image and Signal Processing (CISP). Vol 2. IEEE Mou J, Gao W, Song Z (2013) Image fusion based on non-negative matrix factorization and infrared feature extraction. 2013 6th International Congress on Image and Signal Processing (CISP). Vol 2. IEEE
10.
Zurück zum Zitat Yang Y et al (2020) Infrared and visible image fusion using visual saliency sparse representation and detail injection model. IEEE Trans Instrum Meas 70:1–15CrossRef Yang Y et al (2020) Infrared and visible image fusion using visual saliency sparse representation and detail injection model. IEEE Trans Instrum Meas 70:1–15CrossRef
11.
Zurück zum Zitat Singh S, Anand RS (2019) Multimodal medical image sensor fusion model using sparse K-SVD dictionary learning in nonsubsampled shearlet domain. IEEE Trans Instrum Meas 69(2):593–607CrossRef Singh S, Anand RS (2019) Multimodal medical image sensor fusion model using sparse K-SVD dictionary learning in nonsubsampled shearlet domain. IEEE Trans Instrum Meas 69(2):593–607CrossRef
12.
Zurück zum Zitat Ma J et al (2020) DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980–4995CrossRef Ma J et al (2020) DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980–4995CrossRef
13.
Zurück zum Zitat Liu Y et al (2018) Deep learning for pixel-level image fusion: recent advances and future prospects. Information Fusion 42:158–173CrossRef Liu Y et al (2018) Deep learning for pixel-level image fusion: recent advances and future prospects. Information Fusion 42:158–173CrossRef
14.
Zurück zum Zitat Xu H, Liang P, Yu W, Jiang J, Ma J (2019) Learning a generative model for fusing infrared and visible images via conditional generative adversarial network with dual discriminators. In: IJCAI, pp 3954–3960 Xu H, Liang P, Yu W, Jiang J, Ma J (2019) Learning a generative model for fusing infrared and visible images via conditional generative adversarial network with dual discriminators. In: IJCAI, pp 3954–3960
15.
Zurück zum Zitat Goodfellow I (2016) Nips 2016 tutorial: generative adversarial networks. arXiv preprint arXiv:1701.00160 Goodfellow I (2016) Nips 2016 tutorial: generative adversarial networks. arXiv preprint arXiv:1701.00160
16.
Zurück zum Zitat Ma J et al (2019) FusionGAN: a generative adversarial network for infrared and visible image fusion. Information Fusion 48:11–26CrossRef Ma J et al (2019) FusionGAN: a generative adversarial network for infrared and visible image fusion. Information Fusion 48:11–26CrossRef
17.
Zurück zum Zitat Ma J et al (2020) Infrared and visible image fusion via detail preserving adversarial learning. Information Fusion 54:85–98CrossRef Ma J et al (2020) Infrared and visible image fusion via detail preserving adversarial learning. Information Fusion 54:85–98CrossRef
18.
Zurück zum Zitat Xu X (2020) Multifocus image fusion algorithm based on rough set and neural network. IEEE Sensors J 99:1–1 Xu X (2020) Multifocus image fusion algorithm based on rough set and neural network. IEEE Sensors J 99:1–1
20.
Zurück zum Zitat Liu Y et al (2017) A medical image fusion method based on convolutional neural networks. 2017 20th international conference on information fusion (Fusion). IEEE Liu Y et al (2017) A medical image fusion method based on convolutional neural networks. 2017 20th international conference on information fusion (Fusion). IEEE
21.
Zurück zum Zitat Li X, Zhang X, Ding M (2019) A sum-modified-Laplacian and sparse representation based multimodal medical image fusion in Laplacian pyramid domain. Med Biol Eng Compu 57(10):2265–2275CrossRef Li X, Zhang X, Ding M (2019) A sum-modified-Laplacian and sparse representation based multimodal medical image fusion in Laplacian pyramid domain. Med Biol Eng Compu 57(10):2265–2275CrossRef
22.
Zurück zum Zitat Liu S et al (2019) Multi-focus image fusion based on residual network in non-subsampled shearlet domain. IEEE Access 7:152043–152063CrossRef Liu S et al (2019) Multi-focus image fusion based on residual network in non-subsampled shearlet domain. IEEE Access 7:152043–152063CrossRef
23.
Zurück zum Zitat Huang J et al (2020) MGMDcGAN: medical image fusion using multi-generator multi-discriminator conditional generative adversarial network. IEEE Access 99:1–1 Huang J et al (2020) MGMDcGAN: medical image fusion using multi-generator multi-discriminator conditional generative adversarial network. IEEE Access 99:1–1
24.
Zurück zum Zitat Chan W et al (2016) Listen, attend and spell: a neural network for large vocabulary conversational speech recognition. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE Chan W et al (2016) Listen, attend and spell: a neural network for large vocabulary conversational speech recognition. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE
25.
Zurück zum Zitat Galassi A, Lippi M, Torroni P (2020) Attention in natural language processing. IEEE Trans Neural Netw Learn Syst 32(10):4291–4308CrossRef Galassi A, Lippi M, Torroni P (2020) Attention in natural language processing. IEEE Trans Neural Netw Learn Syst 32(10):4291–4308CrossRef
26.
Zurück zum Zitat Xu K et al (2015) Show, attend and tell: neural image caption generation with visual attention. International conference on machine learning. PMLR Xu K et al (2015) Show, attend and tell: neural image caption generation with visual attention. International conference on machine learning. PMLR
27.
Zurück zum Zitat Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141 Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141
28.
Zurück zum Zitat Woo S et al (2018) Cbam: Convolutional block attention module. Proceedings of the European conference on computer vision (ECCV) Woo S et al (2018) Cbam: Convolutional block attention module. Proceedings of the European conference on computer vision (ECCV)
29.
Zurück zum Zitat Zhao B et al (2017) Diversified visual attention networks for fine-grained object classification. IEEE Trans Multimedia 19(6):1245–1256CrossRef Zhao B et al (2017) Diversified visual attention networks for fine-grained object classification. IEEE Trans Multimedia 19(6):1245–1256CrossRef
30.
Zurück zum Zitat Wang F et al (2017) Residual attention network for image classification. Proceedings of the IEEE conference on computer vision and pattern recognition Wang F et al (2017) Residual attention network for image classification. Proceedings of the IEEE conference on computer vision and pattern recognition
31.
Zurück zum Zitat Newell A, Yang K, Deng J (2016) Stacked hourglass networks for human pose estimation. European conference on computer vision. Springer, Cham Newell A, Yang K, Deng J (2016) Stacked hourglass networks for human pose estimation. European conference on computer vision. Springer, Cham
32.
Zurück zum Zitat Yan Q et al (2019) Attention-guided network for ghost-free high dynamic range imaging. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Yan Q et al (2019) Attention-guided network for ghost-free high dynamic range imaging. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
33.
Zurück zum Zitat Ganasala P, Kumar V, Prasad A D (2016) Performance evaluation of color models in the fusion of functional and anatomical images. J Med Syst 40(5):122CrossRefPubMed Ganasala P, Kumar V, Prasad A D (2016) Performance evaluation of color models in the fusion of functional and anatomical images. J Med Syst 40(5):122CrossRefPubMed
34.
Zurück zum Zitat Roberts JW, Van Aardt JA, Ahmed FB (2008) Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J Appl Remote Sens 2(1):023522CrossRef Roberts JW, Van Aardt JA, Ahmed FB (2008) Assessment of image fusion procedures using entropy, image quality, and multispectral classification. J Appl Remote Sens 2(1):023522CrossRef
35.
Zurück zum Zitat Han Y et al (2013) A new image fusion performance metric based on visual information fidelity. Information Fusion 14(2):127–135CrossRef Han Y et al (2013) A new image fusion performance metric based on visual information fidelity. Information Fusion 14(2):127–135CrossRef
36.
Zurück zum Zitat Wang Z et al (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612CrossRefPubMed Wang Z et al (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612CrossRefPubMed
37.
Zurück zum Zitat Naidu VPS (2014) Hybrid DDCT-PCA based multi sensor image fusion. J Opt 43(1):48–61CrossRef Naidu VPS (2014) Hybrid DDCT-PCA based multi sensor image fusion. J Opt 43(1):48–61CrossRef
38.
Zurück zum Zitat Yin M et al (2018) Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas 68(1):49–64CrossRef Yin M et al (2018) Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas 68(1):49–64CrossRef
39.
Zurück zum Zitat Lewis JJ et al (2007) Pixel-and region-based image fusion with complex wavelets. Information Fusion 8(2):119–130CrossRef Lewis JJ et al (2007) Pixel-and region-based image fusion with complex wavelets. Information Fusion 8(2):119–130CrossRef
40.
Zurück zum Zitat Li J et al (2020) Multigrained attention network for infrared and visible image fusion. IEEE Trans Instrum Meas 70:1–12 Li J et al (2020) Multigrained attention network for infrared and visible image fusion. IEEE Trans Instrum Meas 70:1–12
Metadaten
Titel
MsRAN: a multi-scale residual attention network for multi-model image fusion
verfasst von
Jing Wang
Long Yu
Shengwei Tian
Publikationsdatum
20.10.2022
Verlag
Springer Berlin Heidelberg
Erschienen in
Medical & Biological Engineering & Computing / Ausgabe 12/2022
Print ISSN: 0140-0118
Elektronische ISSN: 1741-0444
DOI
https://doi.org/10.1007/s11517-022-02690-1

Weitere Artikel der Ausgabe 12/2022

Medical & Biological Engineering & Computing 12/2022 Zur Ausgabe

Premium Partner