Skip to main content
Erschienen in: Neural Computing and Applications 4/2024

18.11.2023 | Original Article

Lightweight transformer and multi-head prediction network for no-reference image quality assessment

verfasst von: Zhenjun Tang, Yihua Chen, Zhiyuan Chen, Xiaoping Liang, Xianquan Zhang

Erschienen in: Neural Computing and Applications | Ausgabe 4/2024

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

No-reference (NR) image quality assessment (IQA) is an important task of computer vision. Most NR-IQA methods via deep neural networks do not reach desirable IQA performance and have bulky models which make them difficult to be used in the practical scenarios. This paper proposes a lightweight transformer and multi-head prediction network for NR-IQA. The proposed method consists of two lightweight modules: feature extraction and multi-head prediction. The module of feature extraction exploits lightweight transformer blocks to learn features at different scales for measuring different image distortions. The module of multi-head prediction uses three weighted prediction blocks and an FC layer to aggregate the learned features for predicting image quality score. The weighted prediction block can measure the importance of different elements of input feature at the same scale. Since the importance of feature elements at the same scale and the importance of the features at different scales are both considered, the module of multi-head prediction can provide more accurate prediction results. Extensive experiments on the standard IQA datasets are conducted. The results show that the proposed method outperforms some baseline NR-IQA methods in IQA performance on the large image datasets. For the model complexity, the proposed method is also superior to several recent NR-IQA methods.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Lu F, Zhao Q, Yang G (2015) A no-reference image quality assessment approach based on steerable pyramid decomposition using natural scene statistics. Neural Comput Appl 26(1):77–90CrossRef Lu F, Zhao Q, Yang G (2015) A no-reference image quality assessment approach based on steerable pyramid decomposition using natural scene statistics. Neural Comput Appl 26(1):77–90CrossRef
2.
Zurück zum Zitat Chan KY, Lam H-K, Jiang H (2022) A genetic programming-based convolutional neural network for image quality evaluations. Neural Comput Appl 34(18):15409–15427CrossRef Chan KY, Lam H-K, Jiang H (2022) A genetic programming-based convolutional neural network for image quality evaluations. Neural Comput Appl 34(18):15409–15427CrossRef
3.
Zurück zum Zitat Liang X, Tang Z, Huang Z, Zhang X, Zhang S (2023) Efficient hashing method using 2d–2d pca for image copy detection. IEEE Trans Knowl Data Eng 35(4):3765–3778CrossRef Liang X, Tang Z, Huang Z, Zhang X, Zhang S (2023) Efficient hashing method using 2d–2d pca for image copy detection. IEEE Trans Knowl Data Eng 35(4):3765–3778CrossRef
4.
Zurück zum Zitat Muthusamy D, Sathyamoorthy S (2022) Deep belief network for solving the image quality assessment in full reference and no reference model. Neural Comput Appl 34(24):21809–21833CrossRef Muthusamy D, Sathyamoorthy S (2022) Deep belief network for solving the image quality assessment in full reference and no reference model. Neural Comput Appl 34(24):21809–21833CrossRef
5.
Zurück zum Zitat Tang Z, Zhang X, Li X, Zhang S (2016) Robust image hashing with ring partition and invariant vector distance. IEEE Trans Inf Foren Secur 11(1):200–214CrossRef Tang Z, Zhang X, Li X, Zhang S (2016) Robust image hashing with ring partition and invariant vector distance. IEEE Trans Inf Foren Secur 11(1):200–214CrossRef
6.
Zurück zum Zitat Zheng H, Yang H, Fu J, Zha Z-J, Luo J (2021) Learning conditional knowledge distillation for degraded-reference image quality assessment. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV 2021), pp. 10222–10231 Zheng H, Yang H, Fu J, Zha Z-J, Luo J (2021) Learning conditional knowledge distillation for degraded-reference image quality assessment. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV 2021), pp. 10222–10231
7.
Zurück zum Zitat Tang Z, Huang Z, Yao H, Zhang X, Chen L, Yu C (2018) Perceptual image hashing with weighted dwt features for reduced-reference image quality assessment. Comput J 61:1695–1709CrossRef Tang Z, Huang Z, Yao H, Zhang X, Chen L, Yu C (2018) Perceptual image hashing with weighted dwt features for reduced-reference image quality assessment. Comput J 61:1695–1709CrossRef
8.
Zurück zum Zitat Chen Z, Che Y, Liang X, Tang Z (2022) Multi-level feature aggregaton network for full-reference image quality assessment. In: Proceedings of the 34th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 2022), pp. 861–866 Chen Z, Che Y, Liang X, Tang Z (2022) Multi-level feature aggregaton network for full-reference image quality assessment. In: Proceedings of the 34th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 2022), pp. 861–866
9.
Zurück zum Zitat Chen Y, Chen Z, Yu M, Tang Z (2023) Dual-feature aggregation network for no-reference image quality assessment. In: Proceedings of the 29th International Conference on MultiMedia Modeling (MMM 2023), pp. 149–161 Chen Y, Chen Z, Yu M, Tang Z (2023) Dual-feature aggregation network for no-reference image quality assessment. In: Proceedings of the 29th International Conference on MultiMedia Modeling (MMM 2023), pp. 149–161
10.
Zurück zum Zitat Saad MA, Bovik AC, Charrier C (2012) Blind image quality assessment: a natural scene statistics approach in the dct domain. IEEE Trans Image Process 21(8):3339–3352MathSciNetCrossRef Saad MA, Bovik AC, Charrier C (2012) Blind image quality assessment: a natural scene statistics approach in the dct domain. IEEE Trans Image Process 21(8):3339–3352MathSciNetCrossRef
11.
Zurück zum Zitat Mittal A, Moorthy AK, Bovik AC (2012) No-reference image quality assessment in the spatial domain. IEEE Trans Image Process 21(12):4695–4708MathSciNetCrossRef Mittal A, Moorthy AK, Bovik AC (2012) No-reference image quality assessment in the spatial domain. IEEE Trans Image Process 21(12):4695–4708MathSciNetCrossRef
12.
Zurück zum Zitat Zhang L, Zhang L, Bovik AC (2015) A feature-enriched completely blind image quality evaluator. IEEE Trans Image Process 24(8):2579–2591MathSciNetCrossRef Zhang L, Zhang L, Bovik AC (2015) A feature-enriched completely blind image quality evaluator. IEEE Trans Image Process 24(8):2579–2591MathSciNetCrossRef
13.
Zurück zum Zitat Kang L, Ye P, Li Y, Doermann D (2014) Convolutional neural networks for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2014), pp. 1733–1740 Kang L, Ye P, Li Y, Doermann D (2014) Convolutional neural networks for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2014), pp. 1733–1740
14.
Zurück zum Zitat Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: Transformers for image recognition at scale. In: Proceedings of the International Conference on Learning Representations(ICLR 2021), pp. 1–22 Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: Transformers for image recognition at scale. In: Proceedings of the International Conference on Learning Representations(ICLR 2021), pp. 1–22
15.
Zurück zum Zitat Ferzli R, Karam LJ (2009) A no-reference objective image sharpness metric based on the notion of just noticeable blur (jnb). IEEE Trans Image Process 18(4):717–728MathSciNetCrossRef Ferzli R, Karam LJ (2009) A no-reference objective image sharpness metric based on the notion of just noticeable blur (jnb). IEEE Trans Image Process 18(4):717–728MathSciNetCrossRef
16.
Zurück zum Zitat Sheikh HR, Bovik AC, Cormack L (2005) No-reference quality assessment using natural scene statistics: Jpeg 2000. IEEE Trans Image Process 14(11):1918–1927CrossRef Sheikh HR, Bovik AC, Cormack L (2005) No-reference quality assessment using natural scene statistics: Jpeg 2000. IEEE Trans Image Process 14(11):1918–1927CrossRef
17.
Zurück zum Zitat Suthaharan S (2009) No-reference visually significant blocking artifact metric for natural scene images. Signal Process 89:1647–1652CrossRef Suthaharan S (2009) No-reference visually significant blocking artifact metric for natural scene images. Signal Process 89:1647–1652CrossRef
18.
Zurück zum Zitat Kim J, Lee S (2017) Fully deep blind image quality predictor. IEEE J Sel Topics Signal Process 11(1):206–220CrossRef Kim J, Lee S (2017) Fully deep blind image quality predictor. IEEE J Sel Topics Signal Process 11(1):206–220CrossRef
19.
Zurück zum Zitat Ma K, Liu W, Zhang K, Duanmu Z, Wang Z, Zuo W (2018) End-to-end blind image quality assessment using deep neural networks. IEEE Trans Image Process 27(3):1202–1213MathSciNetCrossRef Ma K, Liu W, Zhang K, Duanmu Z, Wang Z, Zuo W (2018) End-to-end blind image quality assessment using deep neural networks. IEEE Trans Image Process 27(3):1202–1213MathSciNetCrossRef
20.
Zurück zum Zitat Mittal A, Soundararajan R, Bovik AC (2013) Making a completely blindimage quality analyzer. IEEE Signal Process Lett 20(3):209–212CrossRef Mittal A, Soundararajan R, Bovik AC (2013) Making a completely blindimage quality analyzer. IEEE Signal Process Lett 20(3):209–212CrossRef
21.
Zurück zum Zitat Xue W, Mou X, Zhang L, Bovik AC, Feng X (2014) Blind image quality assessment using joint statistics of gradient magnitude and laplacian features. IEEE Trans Image Process 23(11):4850–4862MathSciNetCrossRef Xue W, Mou X, Zhang L, Bovik AC, Feng X (2014) Blind image quality assessment using joint statistics of gradient magnitude and laplacian features. IEEE Trans Image Process 23(11):4850–4862MathSciNetCrossRef
22.
Zurück zum Zitat Li Q, Lin W, Xu J, Fang Y (2016) Blind image quality assessment using statistical structural and luminance features. IEEE Trans Multim 18(12):2457–2469CrossRef Li Q, Lin W, Xu J, Fang Y (2016) Blind image quality assessment using statistical structural and luminance features. IEEE Trans Multim 18(12):2457–2469CrossRef
23.
Zurück zum Zitat Feng P, Tang Z (2022) A survey of visual neural networks: current trends, challenges and opportunities. Multim Syst 29(2):693–724MathSciNetCrossRef Feng P, Tang Z (2022) A survey of visual neural networks: current trends, challenges and opportunities. Multim Syst 29(2):693–724MathSciNetCrossRef
24.
Zurück zum Zitat Huang H, Zeng H, Tian Y, Chen J, Zhu J, Ma K-K (2020) Light field image quality assessment: An overview. In: Proceedings of the IEEE Conference on Multimedia Information Processing and Retrieval (MIPR 2020), pp. 348–353 Huang H, Zeng H, Tian Y, Chen J, Zhu J, Ma K-K (2020) Light field image quality assessment: An overview. In: Proceedings of the IEEE Conference on Multimedia Information Processing and Retrieval (MIPR 2020), pp. 348–353
25.
Zurück zum Zitat Niu Y, Zhong Y, Guo W, Shi Y, Chen P (2019) 2d and 3d image quality assessment: a survey of metrics and challenges. IEEE Access 7:782–801CrossRef Niu Y, Zhong Y, Guo W, Shi Y, Chen P (2019) 2d and 3d image quality assessment: a survey of metrics and challenges. IEEE Access 7:782–801CrossRef
26.
Zurück zum Zitat Yang X, Li F, Liu H (2021) Ttl-iqa: Transitive transfer learning based no-reference image quality assessment. IEEE Trans Multim 23:4326–4340CrossRef Yang X, Li F, Liu H (2021) Ttl-iqa: Transitive transfer learning based no-reference image quality assessment. IEEE Trans Multim 23:4326–4340CrossRef
27.
Zurück zum Zitat Golestaneh SA, Dadsetan S, Kitani KM (2022) No-reference image quality assessment via transformers, relative ranking, and self-consistency. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2022), pp. 3989–3999 Golestaneh SA, Dadsetan S, Kitani KM (2022) No-reference image quality assessment via transformers, relative ranking, and self-consistency. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2022), pp. 3989–3999
28.
Zurück zum Zitat Li F, Zhang Y, Cosman PC (2021) Mmmnet: An end-to-end multi-task deep convolution neural network with multi-scale and multi-hierarchy fusion for blind image quality assessment. IEEE Trans Circuits Syst Video Technol 31(12):4798–4811CrossRef Li F, Zhang Y, Cosman PC (2021) Mmmnet: An end-to-end multi-task deep convolution neural network with multi-scale and multi-hierarchy fusion for blind image quality assessment. IEEE Trans Circuits Syst Video Technol 31(12):4798–4811CrossRef
29.
Zurück zum Zitat Ma J, Wu J, Li L, Dong W, Xie X, Shi G, Lin W (2021) Blind image quality assessment with active inference. IEEE Trans Image Process 30:3650–3663CrossRef Ma J, Wu J, Li L, Dong W, Xie X, Shi G, Lin W (2021) Blind image quality assessment with active inference. IEEE Trans Image Process 30:3650–3663CrossRef
30.
Zurück zum Zitat Pan Z, Yuan F, Lei J, Fang Y, Shao X, Kwong S (2022) Vcrnet: Visual compensation restoration network for no-reference image quality assessment. IEEE Trans Image Process 31:1613–1627CrossRef Pan Z, Yuan F, Lei J, Fang Y, Shao X, Kwong S (2022) Vcrnet: Visual compensation restoration network for no-reference image quality assessment. IEEE Trans Image Process 31:1613–1627CrossRef
31.
Zurück zum Zitat Zhou Z, Xu Y, Quan Y, Xu R (2022) Deep blind image quality assessment using dual-order statistics. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME 2022), pp. 1–6 Zhou Z, Xu Y, Quan Y, Xu R (2022) Deep blind image quality assessment using dual-order statistics. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME 2022), pp. 1–6
32.
Zurück zum Zitat Pan Z, Yuan F, Wang X, Xu L, Shao X, Kwong S (2023) No-reference image quality assessment via multibranch convolutional neural networks. IEEE Trans Artif Intell 4(1):148–160CrossRef Pan Z, Yuan F, Wang X, Xu L, Shao X, Kwong S (2023) No-reference image quality assessment via multibranch convolutional neural networks. IEEE Trans Artif Intell 4(1):148–160CrossRef
33.
Zurück zum Zitat Tan M, Le QV (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: Proceedings of the 36th International Conference on Machine Learning (ICML2019), pp. 10691–10700 Tan M, Le QV (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: Proceedings of the 36th International Conference on Machine Learning (ICML2019), pp. 10691–10700
34.
Zurück zum Zitat Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), pp. 1–14 Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), pp. 1–14
35.
Zurück zum Zitat Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), pp. 4510–4520 Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), pp. 4510–4520
36.
Zurück zum Zitat Howard A, Sandler M, Chen B, Wang W, Chen L-C, Tan M, Chu G, Vasudevan V, Zhu Y, Pang R, Adam H, Le Q (2019) Searching for mobilenetv3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV 2019), pp. 1314–1324 Howard A, Sandler M, Chen B, Wang W, Chen L-C, Tan M, Chu G, Vasudevan V, Zhu Y, Pang R, Adam H, Le Q (2019) Searching for mobilenetv3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV 2019), pp. 1314–1324
37.
Zurück zum Zitat Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: Efficient convolutional neural networks for mobile vision applications. CoRR abs/1704.04861 Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: Efficient convolutional neural networks for mobile vision applications. CoRR abs/1704.04861
38.
Zurück zum Zitat Shi W, Caballero J, Huszár F, Totz J, Aitken AP, Bishop R, Rueckert D, Wang Z (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2016), pp. 1874–1883 Shi W, Caballero J, Huszár F, Totz J, Aitken AP, Bishop R, Rueckert D, Wang Z (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2016), pp. 1874–1883
39.
Zurück zum Zitat Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI 2015), pp. 234–241 Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI 2015), pp. 234–241
40.
Zurück zum Zitat Yang S, Wu T, Shi S, Lao S, Gong Y, Cao M, Wang J, Yang Y (2022) Maniqa: Multi-dimension attention network for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1190–1199 Yang S, Wu T, Shi S, Lao S, Gong Y, Cao M, Wang J, Yang Y (2022) Maniqa: Multi-dimension attention network for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1190–1199
41.
Zurück zum Zitat Sheikh HR, Sabir MF, Bovik AC (2006) A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans Image Process 15(11):3440–3451CrossRef Sheikh HR, Sabir MF, Bovik AC (2006) A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans Image Process 15(11):3440–3451CrossRef
42.
Zurück zum Zitat Larson EC, Chandler DM (2010) Most apparent distortion: full-reference image quality assessment and the role of strategy. J Electron Imag 19(1):011006CrossRef Larson EC, Chandler DM (2010) Most apparent distortion: full-reference image quality assessment and the role of strategy. J Electron Imag 19(1):011006CrossRef
43.
Zurück zum Zitat Ponomarenko N, Ieremeiev O, Lukin V, Egiazarian K, Jin L, Astola J, Vozel B, Chehdi K, Carli M, Battisti F, Kuo C-CJ (2013) Color image database tid2013: Peculiarities and preliminary results. In: Proceedings of the European Workshop on Visual Information Processing (EUVIP 2013), pp. 106–111 Ponomarenko N, Ieremeiev O, Lukin V, Egiazarian K, Jin L, Astola J, Vozel B, Chehdi K, Carli M, Battisti F, Kuo C-CJ (2013) Color image database tid2013: Peculiarities and preliminary results. In: Proceedings of the European Workshop on Visual Information Processing (EUVIP 2013), pp. 106–111
44.
Zurück zum Zitat Lin H, Hosu V, Saupe D (2019) Kadid-10k: A large-scale artificially distorted iqa database. In: Proceedings of the 11th International Conference on Quality of Multimedia Experience (QoMEX 2019), pp. 1–3 Lin H, Hosu V, Saupe D (2019) Kadid-10k: A large-scale artificially distorted iqa database. In: Proceedings of the 11th International Conference on Quality of Multimedia Experience (QoMEX 2019), pp. 1–3
45.
Zurück zum Zitat Hosu V, Lin H, Sziranyi T, Saupe D (2020) Koniq-10k: an ecologically valid database for deep learning of blind image quality assessment. IEEE Trans Image Process 29:4041–4056CrossRef Hosu V, Lin H, Sziranyi T, Saupe D (2020) Koniq-10k: an ecologically valid database for deep learning of blind image quality assessment. IEEE Trans Image Process 29:4041–4056CrossRef
46.
Zurück zum Zitat Su S, Yan Q, Zhu Y, Zhang C, Ge X, Sun J, Zhang Y (2020) Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 3664–3673 Su S, Yan Q, Zhu Y, Zhang C, Ge X, Sun J, Zhang Y (2020) Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 3664–3673
47.
Zurück zum Zitat Yu M, Tang Z, Zhang X, Zhong B, Zhang X (2022) Perceptual hashing with complementary color wavelet transform and compressed sensing for reduced-reference image quality assessment. IEEE Trans Circuits Syst Video Technol 32(11):7559–7574CrossRef Yu M, Tang Z, Zhang X, Zhong B, Zhang X (2022) Perceptual hashing with complementary color wavelet transform and compressed sensing for reduced-reference image quality assessment. IEEE Trans Circuits Syst Video Technol 32(11):7559–7574CrossRef
48.
Zurück zum Zitat Ye P, Kumar J, Kang L, Doermann D (2012) Unsupervised feature learning framework for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2012), pp. 1098–1105 Ye P, Kumar J, Kang L, Doermann D (2012) Unsupervised feature learning framework for no-reference image quality assessment. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2012), pp. 1098–1105
49.
Zurück zum Zitat Xu J, Ye P, Li Q, Du H, Liu Y, Doermann D (2016) Blind image quality assessment based on high order statistics aggregation. IEEE Trans Image Process 25(9):4444–4457MathSciNetCrossRef Xu J, Ye P, Li Q, Du H, Liu Y, Doermann D (2016) Blind image quality assessment based on high order statistics aggregation. IEEE Trans Image Process 25(9):4444–4457MathSciNetCrossRef
50.
Zurück zum Zitat Bosse S, Maniry D, Müller K-R, Wiegand T, Samek W (2018) Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans Image Process 27(1):206–219MathSciNetCrossRef Bosse S, Maniry D, Müller K-R, Wiegand T, Samek W (2018) Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans Image Process 27(1):206–219MathSciNetCrossRef
51.
Zurück zum Zitat Zhang W, Ma K, Yan J, Deng D, Wang Z (2020) Blind image quality assessment using a deep bilinear convolutional neural network. IEEE Trans Circuits Syst Video Technol 30(1):36–47CrossRef Zhang W, Ma K, Yan J, Deng D, Wang Z (2020) Blind image quality assessment using a deep bilinear convolutional neural network. IEEE Trans Circuits Syst Video Technol 30(1):36–47CrossRef
52.
Zurück zum Zitat Zhu H, Li L, Wu J, Dong W, Shi G (2020) Metaiqa: Deep meta-learning for no-reference image quality assessment. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 14131–14140 Zhu H, Li L, Wu J, Dong W, Shi G (2020) Metaiqa: Deep meta-learning for no-reference image quality assessment. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 14131–14140
53.
Zurück zum Zitat Ying Z, Niu H, Gupta P, Mahajan D, Ghadiyaram D, Bovik A (2020) From patches to pictures (paq-2-piq): Mapping the perceptual space of picture quality. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 3572–3582 Ying Z, Niu H, Gupta P, Mahajan D, Ghadiyaram D, Bovik A (2020) From patches to pictures (paq-2-piq): Mapping the perceptual space of picture quality. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), pp. 3572–3582
54.
Zurück zum Zitat Otroshi-Shahreza H, Amini A, Behroozi H (2018) No-reference image quality assessment using transfer learning. In: Proceedings of the 9th International Symposium on Telecommunications (IST 2018), pp. 637–640 Otroshi-Shahreza H, Amini A, Behroozi H (2018) No-reference image quality assessment using transfer learning. In: Proceedings of the 9th International Symposium on Telecommunications (IST 2018), pp. 637–640
55.
Zurück zum Zitat Dendi SVR, Dev C, Kothari N, Channappayya SS (2019) Generating image distortion maps using convolutional autoencoders with application to no reference image quality assessment. IEEE Signal Processing Letters 26(1):89–93CrossRef Dendi SVR, Dev C, Kothari N, Channappayya SS (2019) Generating image distortion maps using convolutional autoencoders with application to no reference image quality assessment. IEEE Signal Processing Letters 26(1):89–93CrossRef
56.
Zurück zum Zitat Zhou Z, Li J, Xu Y, Quan Y (2020) Full-reference image quality metric for blurry images and compressed images using hybrid dictionary learning. Neural Comput Appl 32:12403–12415CrossRef Zhou Z, Li J, Xu Y, Quan Y (2020) Full-reference image quality metric for blurry images and compressed images using hybrid dictionary learning. Neural Comput Appl 32:12403–12415CrossRef
57.
Zurück zum Zitat Tang Z, Chen Z, Li Z, Zhong B, Zhang X, Zhang X (2023) Unifying dual-attention and siamese transformer network for full-reference image quality assessment. ACM Trans Multim Comput Commun Appl 19(6):1–24CrossRef Tang Z, Chen Z, Li Z, Zhong B, Zhang X, Zhang X (2023) Unifying dual-attention and siamese transformer network for full-reference image quality assessment. ACM Trans Multim Comput Commun Appl 19(6):1–24CrossRef
Metadaten
Titel
Lightweight transformer and multi-head prediction network for no-reference image quality assessment
verfasst von
Zhenjun Tang
Yihua Chen
Zhiyuan Chen
Xiaoping Liang
Xianquan Zhang
Publikationsdatum
18.11.2023
Verlag
Springer London
Erschienen in
Neural Computing and Applications / Ausgabe 4/2024
Print ISSN: 0941-0643
Elektronische ISSN: 1433-3058
DOI
https://doi.org/10.1007/s00521-023-09188-3

Weitere Artikel der Ausgabe 4/2024

Neural Computing and Applications 4/2024 Zur Ausgabe

Premium Partner