skip to main content
research-article
Open Access

Unifying Dual-Attention and Siamese Transformer Network for Full-Reference Image Quality Assessment

Authors Info & Claims
Published:12 July 2023Publication History

Skip Abstract Section

Abstract

Image Quality Assessment (IQA) is a critical task of computer vision. Most Full-Reference (FR) IQA methods have limitation in the accurate prediction of perceptual qualities of the traditional distorted images and the Generative Adversarial Networks (GANs) based distorted images. To address this issue, we propose a novel method by Unifying Dual-Attention and Siamese Transformer Network (UniDASTN) for FR-IQA. An important contribution is the spatial attention module composed of a Siamese Transformer Network and a feature fusion block. It can focus on significant regions and effectively maps the perceptual differences between the reference and distorted images to a latent distance for distortion evaluation. Another contribution is the dual-attention strategy that exploits channel attention and spatial attention to aggregate features for enhancing distortion sensitivity. In addition, a novel loss function is designed by jointly exploiting Mean Square Error (MSE), bidirectional Kullback–Leibler divergence, and rank order of quality scores. The designed loss function can offer stable training and thus enables the proposed UniDASTN to effectively learn visual perceptual image quality. Extensive experiments on standard IQA databases are conducted to validate the effectiveness of the proposed UniDASTN. The IQA results demonstrate that the proposed UniDASTN outperforms some state-of-the-art FR-IQA methods on the LIVE, CSIQ, TID2013, and PIPAL databases.

Skip 1INTRODUCTION Section

1 INTRODUCTION

With the popularity of smartphones and the rapid development of image processing and communication technology, digital images are easily acquired and have been widely used in many systems [1]. For an image communication system, visual quality distortion may be introduced in every stage of the system, such as acquisition, compression, transmission, and display. As a result, efficient Image Quality Assessment (IQA) is in demand. In fact, IQA is a critical task of computer vision [2] and has attracted much attention in recent years. There are two types of IQA: subjective IQA and objective IQA. Since human observers are the final receiver of visual signals in most visual communication systems [3], subjective assessment by humans is a valid and effective method to gauge perceptual image quality, which is often expressed by the Mean Opinion Score (MOS) of the gathered subjective assessments. In general, subjective assessment is expensive, time-consuming, and labor-intensive. As a result, there is an urgent need to conceive easy and reliable objective IQA methods, which can be extensively employed in the optimization of parameters of various image processing methods [4, 5].

According to the degree of availability of the reference image information, objective IQA methods can be divided into three categories: Full-Reference (FR) IQA methods [6, 7, 8], Reduced-Reference (RR) IQA methods [9, 10, 11], and No-Reference (NR) IQA methods [12, 13, 14]. In the FR-IQA development, the Mean Squared Error (MSE) and Peak Signal-to-Noise Ratio (PSNR) are two popular FR-IQA methods widely used. However, the MSE and PSNR are not well connected with the perception of Human Visual System (HVS) [15].

To design efficient IQA methods with high prediction accuracy, many researchers conducted extensive explorations to collect image features from the spatial domain or the transform domain, and then performed distance calculations on image features for distortion evaluation. Since the HVS characteristics and structural distortion are the two significant factors affecting the quality of visual perception, they are well considered in developing FR-IQA methods based on traditional hand-crafted features [16, 17, 18, 19, 20, 21, 22, 23].

With the development of deep learning, Convolutional Neural Networks (CNNs) have shown excellent performance in many image applications, such as segmentation and classification. Motivated by the success of CNNs, some researchers have used deep learning techniques to design various IQA methods. An early learning-based method is presented in [24]. This method combines feature learning and regression as a complete optimization process, and can well predict the quality of images. Nowadays, most learning-based methods extract visual sensitivity maps from IQA databases to predict image quality scores [25, 26, 27]. To consider the characteristics of HVS, some researchers designed a visual sensitivity map extraction method that matches human perception and achieved good results [28, 29]. Some other learning-based methods have achieved excellent performance in FR-IQA by focusing on the perceptual error of inter-image information [30, 31, 32]. These research results also show that the “perceptual loss” between images is related to human visual perception and plays a crucial role in quality assessment.

Recently, there is a rapid development in image restoration based on Generative Adversarial Networks (GAN) [33]. The GAN-based image restoration methods can generate images with realistic-looking but unrealistic texture-like noises [34]. However, the existing objective FR-IQA methods, such as the PSNR, Structural SIMilarity (SSIM) [35], have limitation in predicting quality of the GAN-based distorted images. Therefore, effective IQA methods for measuring visual qualities of images produced by GAN-based image restoration methods are needed.

In the past several years, the success of Transformer network [36] in natural language processing motivates and inspires various computer vision tasks [37]. The research results demonstrate that the Transformer network can significantly improve the performance of identification tasks [38, 39, 40] and low-level vision [41]. For the IQA task, several studies based on Transformer [42, 43, 44, 45] have been also investigated and shown advanced performance. However, the Transformer based IQA methods do not reach ideal performance yet. Therefore, more efforts in developing IQA methods based on the Transformer network are required, especially concerning the various structural distortion of images and the characteristics of human visual attention. Furthermore, most deep learning-based IQA methods solely use MSE as a loss function for quality score regression. It is difficult to produce an appropriate quality assessment using this regression approach since it does not take into account the relative ranking relationship of distorted image quality scores.

To address the issue of accurate quality prediction, we propose a novel method by Unifying Dual-Attention and Siamese Transformer Network (UniDASTN) for FR-IQA. Since the proposed UniDASTN incorporates HVS characteristics and exploits the distortion information between the reference image and the distorted image, it can provide an accurate quality assessment of significant regions of human visual attention and obtain better results correlated with human visual perception. The main contributions of our proposed UniDASTN are highlighted below.

We propose a novel spatial attention module. This spatial attention module is composed of a novel Siamese Transformer network and a new feature fusion block. It can focus on significant regions in line with human visual perception and thus effectively maps the perceptual differences between the reference and distorted images to a latent distance for distortion evaluation.

We propose a novel dual-attention strategy that exploits channel attention and spatial attention to adaptively extract much more significant features for accurate IQA. Specifically, we use a squeeze-and-excitation network to model channel attention and exploit the proposed spatial attention module to learn spatial attention for enhancing the distortion sensitivity of the extracted ViT feature representations.

We design a new joint loss function. The designed loss function combines MSE, Bidirectional Kullback–Leibler (KL) divergence, and rank order of quality scores. It can offer stable training and thus enables the proposed UniDASTN to learn visual perceptual image quality effectively.

The rest of this article is organized as follows. Section 2 reviews the related work. Section 3 explains the proposed UniDASTN in detail. Section 4 presents the experimental results of the proposed UniDASTN and the comparison with some state-of-the-art FR-IQA methods. Section 5 concludes our work.

Skip 2RELATED WORK Section

2 RELATED WORK

In general, a reliable FR-IQA method should make an accurate quality prediction for all types of distortion. In the literature, many scholars have tried their best to design some meaningful FR-IQA methods for pursuing high prediction performance. According to the techniques of image feature extraction, the existing FR-IQA methods can be divided into the below three categories.

2.1 Traditional IQA Methods

In the previous few decades, a lot of research has been done on FR-IQA [2]. The aim of the FR-IQA is to predict the quality of a target image with full access to its reference image. Many FR-IQA methods follow a two-stage framework, which firstly calculates the distance between the distorted and reference image features and then translates the distance to the quality ratings. The most commonly used FR-IQA methods are MSE and PSNR [46], which have been successfully used in many visual applications due to their simplicity and easy use [47]. However, the two methods measure the differences between images by pixel-level differences, which differ from the perceptual consistency of the HVS [15].

The HVS can extract structural information from the scene in a highly adaptive manner. To conduct IQA from structural similarity, Wang et al. [35] proposed a useful technique called SSIM to evaluate quality of the distorted image in terms of brightness, structure, and contrast. Then, an improved method called Multi-Scale SSIM (MS-SSIM) [8] was proposed by extending SSIM to multiple scales. Subsequently, many FR-IQA methods gradually extract the structural information of images in the spatial domain, such as PSIM [16] and GMSD [21]. They both extract gradient features in the spatial domain to describe structural changes. However, the FR-IQA methods designed from the perspective of image structure are inefficient in predicting the quality of specific distortion types, such as blurring and compression. Based on these observations, researchers have started to investigate IQA methods based on low-level features, such as color, edge, and texture. For example, Zhang et al. [23] designed an efficient method called Feature SIMilarity (FSIM) to understand image scenes by some low-level features. In another work, Zhang et al. [22] proposed a visual evaluation model to calculate Visual Saliency-induced Index (VSI). Subsequently, some other useful IQA methods are proposed, such as Most Apparent Distortion (MAD) [18] and Normalized Laplacian PyramiD (NLPD) [17].

Furthermore, some IQA methods try to extract features from transform domain because image quality degradation can be also shown in the transform domain. For example, Sheikh et al. investigated the HVS with the Information Fidelity Criterion (IFC) index [20] and proposed the Visual Information Fidelity (VIF) [19] method for quantifying the information errors between the distorted image and the reference image.

2.2 CNN-Based IQA Methods

There have been some attempts to address FR-IQA tasks with CNN techniques due to the success of CNN in many vision applications. For example, Kang et al. [24] were the first to apply CNN to the IQA task without the use of any existing features. Similar to other computer vision tasks, this method combines feature learning and regression as complete optimization schemes. Specifically, several patches are cropped out of the image, and each patch is given a label corresponding to the visual quality score of the image for training. Later, visual sensitivity information of images is combined with CNN to predict image quality scores. In another work, Kim and Lee [27] exploited CNN to design a new FR-IQA method that learns human visual sensitivity to predict subjective scores by feeding distorted images and their objective error map into the network.

Recently, Zhang et al. [32] designed a new FR-IQA method called LPIPS (Learned Perceptual Image Patch Similarity). The trained deep features of LPIPS optimized by the Euclidean distance between the reference image and the distorted image are effective in FR-IQA. In another work, Prashnani et al. [31] presented a new FR-IQA framework called PieAPP, which is based on pairwise learning. The results of PieAPP show that pairwise learning helps to improve image quality prediction. Subsequently, Ding et al. [30] introduced a novel FR-IQA method that unifies structure and texture similarity. This method shows good performance in predicting visual qualities of textural images and natural images.

2.3 Transformer-Based IQA Methods

The Vision Transformer (ViT) [38] is the famous work about the Transformer in computer vision. This work has demonstrated that the Transformer may “fully” replace the convolutions regularly used in deep neural networks. In general, a pure Transformer architecture includes three important components, i.e., Multi-Head Attention (MHA), Multi-Layer Perceptron (MLP), and Layer Normalization (LN). The ViT exploits a pure Transformer to conduct image classification via taking an image as a sequence of patches. The ViT can learn with minimum inductive bias. Compared with CNNs, the relational representation learned by the Transformer is more general and robust than the local patterns of convolutional modules [48].

The ViT is an efficient technique based on self-attention. Since the self-attention layer collects global information from the whole input sequence, it can measure “perceptual loss” between the distorted image and the reference image when predicting image quality. This design might be helpful in the FR-IQA task because recent studies [42, 43, 44, 45] have proved the usefulness of deep features in the IQA task.

Inspired by ViT, You et al. [45] proposed the first Transformer-based IQA model named TRIQ. The TRIQ uses CNN to extract feature maps which are sent to a shallow Transformer encoder to solve blind IQA tasks. Taking the advantage of the inductive capability of CNN architecture and Transformer encoder for aggregated representation of attention mechanism, the TRIQ achieves outstanding performance. Subsequently, Cheon et al. [42] designed a new Transformer model named IQT for FR-IQA. In the IQT, the encoder-decoder architecture is employed to measure similarity. In another work, Ke et al. [44] used the Transformer to build a new method called MUSIQ (MUlti-Scale Image Quality). The MUSIQ can handle images with different resolutions by mapping input image to a multi-scale representation.

Skip 3PROPOSED METHOD Section

3 PROPOSED METHOD

In this section, we introduce the overall of the proposed UniDASTN. As shown in Figure 1, the proposed UniDASTN takes pairs of reference images and distorted images as input, and it consists of five key components: a feature extraction module, a channel attention module, a feature embedding module, a spatial attention module, and a prediction head. Sections 3.13.5 present the five components in detail. Section 3.6 explains our loss function.

Fig. 1.

Fig. 1. Overview of the proposed UniDASTN. The proposed UniDASTN takes pairs of distorted images and reference images as input and then produces feature maps using a pre-trained ViT. The feature maps are input to a Squeeze-and-Excitation network to focus on the most informative channel features before being converted to fixed-size vectors, and then fed into a spatial attention module for training. The final output is a score for assessing image quality.

3.1 Feature Extraction Module

As depicted in Figure 1, the front part of the feature extraction module is a two-branch pre-trained ViT, which acts as a feature extraction backbone that mainly focuses on extracting low-level features of color, texture, shape information, and high-level features of semantic information. We crop out M patches for each input reference and distorted images separately. Let \(p_{i}^{ref}\) and \(p_{i}^{dist}\) be the \({i}\)th patches of the input reference and distorted images, where \(i\in \lbrace 1,2,\ldots , M \rbrace\). The \({i}\)th input patch \(p_{i}^{ref}\) of the reference image and the \({i}\)th input patch \(p_{i}^{dist}\) of the distorted image are input to the feature extractor to obtain the corresponding feature representations which are extracted from four intermediate layers of ViT, i.e., \(layer \in \lbrace 3,5,8,11\rbrace\). For these feature representations, each \(f_{layer} \in \mathbb {R}^{ c\times w\times h}\) has the same shape. Finally, we concatenate these four feature maps by channel dimension to get a total feature map \(f_{ref} \in \mathbb {R}^{C\times w\times h }\) and \(f_{dist} \in \mathbb {R}^{C\times w\times h }\), where \(C=c\times 4\).

To construct perceptual differences between the reference and distorted images, a simple subtraction operation between the feature maps \(f_{ref}\) and \(f_{dist}\) is conducted to get the difference map \(f_{diff}\) as follows: (1) \(\begin{align} f_{diff} = f_{ref} - f_{dist}. \end{align}\)

To illustrate the effectiveness of the difference map for FR-IQA, some visual examples are presented in Figure 2. In this figure, the first row shows a reference image and its distorted images with different distortions introduced by additive white Gaussian noise. The second row shows their corresponding feature maps, and the third row shows the difference maps. It can be seen that as more distortions are introduced to the image, there are significant changes in the difference map. This means that the difference map can be used for FR-IQA.

Fig. 2.

Fig. 2. A reference image, its distorted images, their feature maps, and difference maps.

After that, these feature maps are sent to a 1 × 1 convolutional layer to aggregate the information of each channel and downsampling.

3.2 Channel Attention Module

The channel attention module is made up of a Squeeze-and-Excite network. Firstly, the feature maps generated by the feature extraction module are squeezed to acquire global features at the channel level. The global features are then subjected to an excitation operation to learn the channel relationship and acquire the channel weights. Finally, the channel weights are multiplied by the input feature maps to generate the final features.

In summary, the channel attention module performs the channel rating action. This attention mechanism enables the network to prioritize relevant channel features while suppressing unimportant channel features. Note that the maps \(f_{ref}\), \(f_{dist}\), and \(f_{diff}\) are separately fed to the channel attention module. Take \(f_{diff}\) as an example for illustrating the formal descriptions of the whole process as follows: (2) \(\begin{align} z_k &= F_{sq}(f_{diff})=\frac{1}{h\times w} \sum _{i=1}^{h} \sum _{j=1}^{w} f_{diff}(k,i,j), (k=1, 2, \dots , C), \end{align}\) (3) \(\begin{align} z = [z_1, z_2, \dots , z_C], \end{align}\) (4) \(\begin{align} s = F_{ex}(z)=Sigmod(FC(Relu(FC(z)))), \end{align}\) (5) \(\begin{align} \tilde{f_{diff}} = (f_{diff}\cdot s)+f_{diff.} \end{align}\)

3.3 Feature Embedding Module

To feed these three feature maps into the spatial attention module for training, we need to perform feature embedding operations on them.

Since our siamese transformer network utilizes constant latent vectors sized D in all layers, the feature maps are reshaped by spatial dimension to convert to N without destroying the pixel information. Thus, the intermediate embedding features can be determined by the below equation, where \(N=h \times w\). (6) \( \begin{equation} z^{diff} = \left[ \tilde{f_{diff}^{1}}; \tilde{f_{diff}^{2}}; \cdots ; \tilde{f_{diff}^{N}} \right] \in \mathbb {R}^{N \times D }. \end{equation} \)

Similar to typical ViT [38] models, we prepare a learnable embedding called “quality token” \(f_{qua}\) to the sequence of \(z^{diff}\), whose state at the output of the network is viewed as the predicted quality. Here the function of “quality token” is similar to the “classification token” in the classical ViT. In addition, the position embedding \(f_{pos}^{diff} \in \mathbb {R}^{\left(N+1 \right)\times D }\) is also added to \(z^{diff}\) for maintaining positional information. Consequently, an embedding feature with positional information is available as follows: (7) \(\begin{equation} z_{0}^{diff} = \left[ f_{qua }; \tilde{f_{diff}^{1}}; \cdots ; \tilde{f_{diff}^{N}} \right] + f_{pos}^{diff}. \end{equation}\)

Similarly, the embedding features of \(f_{ref}\) and \(f_{dist}\) can be determined by the Equations (8) and (9), respectively. (8) \(\begin{align} &z_{0}^{ref} =\left[ f_{qua}; \tilde{f_{ref}^{1}};\cdots ; \tilde{f_{ref}^{N}} \right] + f_{pos}^{ref}, \end{align}\) (9) \(\begin{align} &z_{0}^{dist} =\left[ f_{qua}; \tilde{f_{dist}^{1}};\cdots ; \tilde{f_{dist}^{N}} \right] + f_{pos}^{dist}. \end{align}\)

A visual example of feature embedding is provided in Figure 3 to facilitate comprehension. The feature embedding module’s inputs are feature maps of size \(28 \times 28 \times 256\) output from the channel attention module. We perform a reshape operation on the feature map by the spatial dimension that produces the intermediate embedding features sized \(784 \times 256\). Finally, the final embedding features sized \(785 \times 256\) are obtained by embedding position and quality. Next, these embedded features are sent to a spatial attention module for training.

Fig. 3.

Fig. 3. Detailed process of feature embedding module.

3.4 Spatial Attention Module

The spatial attention module consists of a novel Siamese Transformer network and a new feature fusion block. It has three inputs, i.e., embedding features of the distorted image, embedding features of the reference image, and embedding features of the perceptual differences between distorted and reference images. Unlike conventional FR-IQA methods that only perform distance calculation between the reference and distorted images, the spatial attention module employs a Siamese Transformer network to learn information from perceptual differences between the distorted image features and the reference image features. This spatial attention module exploits the latent distance to calculate the perceptual differences between the reference and distorted images for quality prediction.

As shown in Figure 3, we connect each pixel in the feature map by channel dimension and use the concatenated vector as the token input to the siamese transformer network. As a result, the MHA employed in the transformer encoder and the transformer decoder will act as special spatial attention. Its role is to assign various weights to different pixels and thus allow the network to focus on significant features while suppressing irrelevant ones. This strategy can better learn perceptual image features.

In summary, the calculations of the spatial attention module can be formulated as Equations (10)–(23), where L denotes the number of the module layers. (10) \(\begin{align} &q_{l}= k_{l}= v_{l}=z_{l-1}^{diff}, \end{align}\) (11) \(\begin{align} &z_{l}^{\ast } = \mathrm{LN}\left(\mathrm{MSA}\left(q_{l},k_{l},v_{l} \right) + z_{l-1}^{diff} \right), \end{align}\) (12) \(\begin{align} &z_{l}^{diff} = \mathrm{LN}\left(\mathrm{MLP}\left(z_{l}^{\ast } \right) + z_{l}^{\ast } \right) ,l=1, 2, \ldots , L , \end{align}\) (13) \(\begin{align} &z_{L}^{diff}=\left[ z_{E_{qua}},z_{E_{0}},z_{E_{1}},\ldots ,z_{E_{N}} \right], \end{align}\) where MSA(\(\cdot\)) is the operation of Multi-head Self-Attention (MSA).

The inputs of one transformer decoder are \(z_{0}^{dist}\), and \(z_{L}^{diff}\) which is the output of the transformer encoder. Similarly, the inputs of the other transformer decoder are \(z_{0}^{ref}\) and \(z_{L}^{diff}\). The calculations of one single branch of the process of spatial attention module can be formulated as Equations (14)–(18). (14) \(\begin{align} &q_{l}= k_{l}= v_{l}=z_{l-1}^{ref}, \end{align}\) (15) \(\begin{align} &z_{l}^{\ast } = \mathrm{LN}\left(\mathrm{MSA}\left(q_{l},k_{l},v_{l} \right) + z_{l-1}^{ref} \right), \end{align}\) (16) \(\begin{align} &\hat{q_{l} } = z_{l}^{\ast },\hat{k_{l} }=\hat{v_{l} }=z_{L}^{diff}, \end{align}\) (17) \(\begin{align} &z_{l}^{\ast \ast }=\mathrm{LN}\left(\mathrm{MCA}\left(\hat{q_{l} }, \hat{k_{l} },\hat{v_{l} }\right) + z_{l}^{\ast } \right), \end{align}\) (18) \(\begin{align} &z_{l}^{ref }=\mathrm{LN}\left(\mathrm{MLP}\left(z_{l}^{\ast \ast } \right) +z_{l}^{\ast \ast } \right) ,l=1, 2, \ldots , L, \end{align}\) where MCA(\(\cdot\)) is the operation of Multi-head Cross-Attention (MCA).

Note that the outputs of the siamese transformer network are \(z_{L}^{ref\sim diff}\) and \(z_{L}^{dist\sim diff}\), which are defined by Equations (19)–(20). (19) \(\begin{align} &z_{L}^{ref\sim diff} = [ z_{R_{qua}},z_{R_{0}},z_{R_{1}},\ldots ,z_{R_{N}}], \end{align}\) (20) \(\begin{align} &z_{L}^{dist\sim diff} = [ z_{D_{qua}},z_{D_{0}},z_{D_{1}},\ldots ,z_{D_{N}}]. \end{align}\)

Obviously, the output features \(z_{L}^{ref\sim diff}\) and \(z_{L}^{dist\sim diff}\) are complementary. Therefore, it is necessary to aggregate the features of both branches. To do so, an additional feature fusion block is designed. Figure 4 depicts the diagram of our proposed feature fusion block. The proposed feature fusion block can efficiently fuse information from both branches of the siamese transformer network. It makes the proposed UniDASTN more effective in predicting image quality.

Fig. 4.

Fig. 4. Diagram of the proposed feature fusion block.

Specifically, a sequence of tokens called \(\tilde{z_{i}}\) is constructed by concatenating all tokens in \(z_{L}^{ref\sim diff}\) and \(z_{L}^{dist\sim diff}\) except for the “quality token”. And then a new “quality token” \(\tilde{z_{qua}}\) is determined by the summation of \(z_{R_{qua}}\) and \(z_{D_{qua}}\). Consequently, a new sequence of tokens called \(\tilde{z}\) can be obtained by using \(\tilde{z_{qua}}\) and \(\tilde{z_{i}}\). Formal descriptions of these calculations are defined in Equations (21)–(23). Next, \(\tilde{z}\) is delivered to a transformer encoder for learning information from every token to help to improve the prediction performance. (21) \(\begin{align} &\tilde{z_{i} } = \mathrm{concatenate}\left(z_{R_{0} },\ldots ,z_{R_{N} },z_{D_{0} },\ldots ,z_{D_{N} } \right) , \end{align}\) (22) \(\begin{align} &\tilde{z_{qua}} = z_{R_{qua} } + z_{D_{qua} } , \end{align}\) (23) \(\begin{align} &\tilde{z} = \left[ \tilde{z_{qua}},\tilde{z_{i} } \right]. \end{align}\)

3.5 Prediction Head

The quality score is finally determined in the prediction head. The “quality token” \(\tilde{z_{qua}} \in \mathbb {R}^{1\times D }\) output by the spatial attention module contains the quality information. It is input to the prediction head. The prediction head is achieved by MLP which is composed of two fully connected layers. In the MLP, the first layer is activated by the ReLU, and the second layer only outputs one channel to predict a score, i.e., the final predicted score.

3.6 Loss Function

The novel loss function is designed to evaluate the predicted image scores and optimize the proposed model. Note that the MSE loss is sensitive to outliers. If just the MSE loss is utilized as the loss function for training, the training process is subject to substantial gradient changes caused by outliers, making loss convergence unstable.

To obtain more accurate image prediction scores, we design a novel joint loss function to optimize the proposed UniDASTN for stable training. Our loss function consists of three components, including the MSE loss \(L_{MSE}\), the bidirectional KL divergence loss \(L_{BKL}\), and the rank loss \(L_{RANK}\). The MSE loss \(L_{MSE}\) is defined as (24) \(\begin{equation} L_{MSE} = \sum _{i=1}^{K} MSE(p_i,\hat{q_i}) , \end{equation}\) where \(p_i\) and \(\hat{q_i}\) are the predicted score and the MOS of the ith image, \(MSE(\cdot ,\cdot)\) is the standard MSE, and K is the image number.

We select KL divergence as one component of the loss function because we want to use KL divergence to calculate the distance between the two distributions of prediction scores and MOS values in space. However, since KL divergence assesses the loss of information from one distribution compared to the other, it does not measure the proximity of two distributions in space. As a result, we improve the original KL divergence and create a bidirectional KL divergence for the IQA task to measure the differences between predicted scores and MOS values.

The bidirectional KL divergence loss \(L_{BKL}\) is defined as (25) \(\begin{equation} \begin{aligned}L_{BKL} &= KL\left(Q||\frac{1}{2}(Q+\hat{Q})\right)+KL\left(\hat{Q}||\frac{1}{2}(Q+\hat{Q})\right)\\ &=\sum _{i=1}^{K}Q\times log\frac{Q}{\frac{1}{2}(Q+\hat{Q})} +\sum _{i=1}^{K}\hat{Q}\times log\frac{Q}{\frac{1}{2}(Q+\hat{Q})} \end{aligned} \end{equation}\) where \(\hat{Q}\) and Q are the outputs of softmax regression defined as follows: (26) \(\begin{equation} \hat{Q} = \frac{\hat{e^{q_i}}}{\sum _{i=1}^{K}\hat{e^{q_i}} } , \end{equation}\) (27) \(\begin{equation} Q = \frac{e^{p_i} }{\sum _{i=1}^{K}e^{p_i} } . \end{equation}\)

Furthermore, only considering the regression ability of the network on the quality scores does not consistent with the perception of human vision. That is because, in terms of regression, the same degree of regression values is considered to achieve the same performance results. However, in terms of rank order, regression values of the same degree do not guarantee that the results predicted by the network for distorted images are of the same order as the MOS values.

Motivated by the concept of “rank loss”, we consider that the rank order of the prediction scores should be consistent with the ground truth for reflecting the rank order among the degraded images in the design of the loss function of the network. In a previous study [31], the network is trained by considering the probability of preference for each image pair with different distortions. In another study [29], the authors think that when the rank of the output scores does not agree with the rank of the ground truth scores regardless of their distortion types, the network should be penalized. Inspired by [29], we devise the following pairwise rank loss \(L_{r}(I_i,I_j)\) to determine the rank order of a pair of distorted images \(I_i\) and \(I_j\) as follows: (28) \(\begin{align} L_{r}(I_i,I_j) &= max\left(0,\frac{-(\hat{M}_i -\hat{M}_j)(M_i -M_j)}{|\hat{M}_i -\hat{M}_j |+ \varepsilon } \right)\times \mu , \end{align}\) (29) \(\begin{align} \mu &= (M_i -\hat{M}_i)^{2}+(M_j -\hat{M}_j)^{2}, \end{align}\) where \(\varepsilon\) is a small stability term and \(\mu\) is a penalty term. Note that \(L_{r}(I_i,I_j)\) is always 0 if the rank order of the predicted scores exactly matches that of MOS values. Otherwise, it is a positive number that penalizes the network. Therefore, we can obtain the total rank loss \(L_{RANK}\) by considering all pair rank losses as follows: (30) \(\begin{equation} \begin{aligned}L_{RANK} &= L_{r}(I_1,I_2) + L_{r}(I_1,I_3) + \dots + L_{r}(I_1,I_K)\\ &\quad +L_{r}(I_2,I_3) + L_{r}(I_2,I_4) + \dots + L_{r}(I_2,I_K)\\ &\quad +\dots \\ &\quad +L_{r}(I_i,I_{i+1}) + L_{r}(I_i,I_{i+2}) + \dots + L_{r}(I_i,I_K)\\ &\quad +\dots \\ &\quad +L_{r}(I_{K-1},I_K) \end{aligned} \end{equation}\)

Consequently, the total loss can be assessed as a weighted sum of these three loss items as follows: (31) \(\begin{equation} L_{total} = \alpha L_{MSE}(M,\hat{M}) + \beta L_{BKL}(Q,\hat{Q}) + \gamma L_{RANK} , \end{equation}\) where \(\alpha\), \(\beta ,\) and \(\gamma\) are the hyper-parameters for adjusting the weights.

Skip 4EXPERIMENTAL RESULTS Section

4 EXPERIMENTAL RESULTS

4.1 Implementation Details of Our UniDASTN

The implementation details of the proposed UniDASTN are provided here. In the training phase, the hyper-parameters of the loss function are \(\alpha =1.0\), \(\beta =0.5,\) and \(\gamma =0.1\). Our used data augmentation strategies include horizontal flip, random rotation, and random cropping. Specifically, we do a horizontal flip with a probability of 0.5. With a probability of 0.25, we do rotations of 90, 180, or 270 degrees, as well as no rotation. Next, the images are randomly cropped to provide \(224\times 224\times 3\) image patches for the feature extraction backbone. The training is carried out with a batch size of 8 and employs an ADAM optimizer for 300 epochs using the designed joint loss function. An initial learning rate is set to 1e-4, and a weight decay of 1e-5 is used. We set the learning rate of each parameter by using a cosine annealing schedule, where \(\eta _{max}\) is set to the initial learning rate and \(\eta _{min}\) is set to 0 and the number of cosine learning rate period \(T_{cur}\) is 50. The value of \(\varepsilon\) used in the loss function is set to 1e-8.

In the testing phase, we perform test-time augmentation on the images by cropping into M overlapping patches of size \(224\times 224\times 3\) and predict the quality score by averaging the scores of all crops, where M is set to 20.

The hyper-parameters of the proposed UniDASTN are as follows. (1) The number of layers in the spatial attention module is 2, i.e., \(L=2\). (2) The head numbers of MSA and MCA are both 8. (3) The dimension of per head is 32. (4) The dimension of Transformer is 256, i.e., \(D=256\). (5) The dimension of the MLP in Transformer is 1024. (6) The dimension of the first layer in prediction head is 512, i.e., \(Dhead=512\).

4.2 Evaluation Criteria

In the following experiments, three popular IQA metrics are employed to quantify performance. These three metrics are: the Pearson Linear Correlation Coefficient (PLCC) [49], the Spearman Rank-order Correlation Coefficient (SRCC) [50], and the Kendall Rank-order Correlation Coefficient (KRCC) [51]. The PLCC is used to assess the ability of predicting subjective scores with low error. The SRCC and KRCC are used to measure the prediction monotonicity of an FR-IQA method. The ranges of PLCC, SRCC, and KRCC are all \([-1, 1]\). A positive value indicates a positive correlation between the two sets of data, and a higher value indicates better performance.

4.3 Datasets

We use five open databases to conduct extensive experiments. Specifically, the used databases include the TID2013 [52], the Categorical Subjective Image Quality (CSIQ) database [18], the LIVE IQA database [53], KADID-10k [54], and PIPAL [34]. These five databases are commonly used in IQA research. Table 1 summarizes the basic information of these databases. Traditional IQA databases are built for general IQA tasks, and they include some common types of distortions, such as JPEG compression, JPEG2000 compression, white noise injection and blurring.

Table 1.
DatabaseRef ImgsDist ImgsDistored TypeNum of TypesScore TypeEnv
KADID-10k8110.1ktraditional25DMOScrowdsourcing
LIVE29779traditional5DMOSlab
CSIQ30866traditional6DMOSlab
TID2013253000traditional24MOSlab
PIPAL200232ktradi + alg. outputs40MOScrowdsourcing

Table 1. IQA Databases for Model Training and Performance Evaluation

4.4 Performance Comparison

To demonstrate the advantage, we conduct two types of experiments to compare our UniDASTN and some state-of-the-art FR-IQA methods. In the comparisons, conventional FR-IQA methods and deep learning-based FR-IQA methods are both selected. The first type is to validate the quality prediction performance on the same dataset. This evaluation experiment is to test the accuracy of the quality prediction of each objective IQA method. The other type is to validate each FR-IQA method across datasets. Such evaluation experiment is to verify the generalization ability of each objective FR-IQA method.

Evaluation on the same dataset. According to the suggestion [25, 29, 59], we randomly divide the dataset into training set (80% images) and test set (20% images) based on the reference image. During the network training, only the data from the training set are visible to the FR-IQA methods. We determine the weights loaded by model based on the training and use the test set to evaluate the final performance. We repeat the experiment five times with different seeds, and the results are averaged for a fair comparison.

We compare our UniDASTN method with some state-of-the-art FR-IQA methods on the traditional distorted types of datasets, i.e., LIVE, CSIQ, and TID2013. In this experiment, our UniDASTN is compared with a total of 17 FR-IQA methods, including 10 non-deep-learning-based methods and 7 deep-learning-based methods. The 10 non-deep-learning-based methods are PSNR [46], SSIM [35], MS-SSIM [8], FSIMc [23], VSI [22], MAD [18], VIF [19], NLPD [17], GMSD [21], and SCQI [55]. The seven deep-learning-based methods are DOG-SSIMc [65], DeepQA [27], DualCNN [66], WaDiQaM-FR [25], PieAPP [31], JND-SalCAR [29], and AHIQ [59].

The PLCC and SRCC results of these methods are shown in Table 2. Note that the results of the compared methods are taken from the corresponding original papers and [30]. In Table 2, the best results are in bold and the missing results are marked with “–”. It can be seen that our UniDASTN consistently outperforms all state-of-the-art FR-IQA methods in terms of PLCC and SRCC on three IQA datasets. Furthermore, Table 2 shows that our UniDASTN made significant progress in CSIQ and TID2013 when compared to other state-of-the-art FR-IQA methods. Although some existing methods have achieved high levels of visual quality score assessment for the LIVE dataset, our UniDASTN still makes some progress on the SRCC metric. These results illustrate that our UniDASTN achieves solid improvements over the previous methods and can cope well with traditional distortion types of images.

Table 2.
MethodLIVECSIQTID2013
PLCCSRCCPLCCSRCCPLCCSRCC
PSNR [46]0.8650.8730.8190.8100.6770.687
SSIM [35]0.9370.9480.8520.8650.7770.727
MS-SSIM [8]0.9400.9510.8890.9060.8300.786
FSIMc [23]0.9610.9650.9190.9310.8770.851
VSI [22]0.9480.9520.9280.9420.9000.897
MAD [18]0.9680.9670.9500.9470.8270.781
VIF [19]0.9600.9640.9130.9110.7710.677
NLPD [17]0.9320.9370.9230.9320.8390.800
GMSD [21]0.9570.9600.9450.9500.8550.804
SCQI [55]0.9370.9480.9270.9430.9070.905
DOG-SSIMc [65]0.9660.9630.9430.9540.9340.926
DeepQA [27]0.9820.9810.9650.9610.9470.939
DualCNN [66]----0.9240.926
WaDiQaM-FR [25]0.9800.970--0.9460.940
PieAPP [31]0.9860.9770.9750.9730.9460.945
JND-SalCAR [29]0.9870.9840.9770.9760.9560.949
AHIQ [59]0.9890.9840.9780.9750.9680.962
Our UniDASTN0.9890.9870.9870.9830.9770.974

Table 2. Performance Comparison on Same Dataset Experiment

Furthermore, to validate the effectiveness of the proposed UniDASTN in predicting the quality of images generated by GAN-based image restoration methods, we also compare our UniDASTN method with some state-of-the-art FR-IQA methods on the GAN-based synthetic image dataset, i.e., the PIPAL dataset. The PIPAL dataset contains many traditional distortion types of images and some distorted images generated by GAN-based image restoration methods. In this experiment, our UniDASTN is compared with a total of 19 FR-IQA methods, including 14 non-deep-learning-based methods and 5 deep-learning-based methods. The 14 non-deep-learning-based FR-IQA methods are PSNR [46], NQM [57], UQI [69], SSIM [35], MS-SSIM [8], RFSIM [68], GSM [61], SRSIM [60], FSIM [23], VSI [22], NIQE [63], MA [62], PI [56], and Brisque [64]. The five deep-learning-based methods are LPIPS-Alex [32], LPIPS-VGG [32], DISTS [30], IQT [42], and AHIQ [59]. All deep-learning-based methods are trained on the PIPAL training set, and the images from the validation set of PIPAL are used for the test.

As can be seen from Table 3, the best results are in bold, and our UniDASTN achieves the second-best result in the SRCC performance comparison and the third-best result in the PLCC performance comparison. Although our UniDASTN shows a slightly lower performance on the PIPAL dataset, our UniDASTN still outperforms most compared methods. This experiment demonstrates that our UniDASTN is effective in assessing the quality of GAN-based synthetic images.

Table 3.
MethodPIPAL Validation
PLCCSRCC
PSNR [46]0.2690.234
NQM [57]0.3640.302
UQI [69]0.5050.461
SSIM [35]0.3770.319
MS-SSIM [8]0.1190.338
RFSIM [68]0.2850.254
GSM [61]0.4500.379
SRSIM [60]0.6260.529
FSIM [23]0.5530.452
VSI [22]0.4930.411
NIQE [63]0.1290.012
MA [62]0.0970.099
PI [56]0.1340.064
Brisque [64]0.0520.008
LPIPS-Alex [32]0.6060.569
LPIPS-VGG [32]0.6110.551
DISTS [30]0.6340.608
IQT [42]0.8400.820
AHIQ [59]0.8450.835
Our UniDASTN0.8310.829

Table 3. Performance Comparison on PIPAL Validation Dataset

Evaluation on the cross-dataset. We conduct a cross-dataset experiment to evaluate the generality of our UniDASTN. In this experiment, the FR-IQA methods are trained on a single dataset and thus tested on several other databases to validate their generalization ability in IQA.

As recommended by [30, 31, 32, 42], we use the KADID-10k dataset for training the deep-learning-based FR-IQA methods, and evaluate their performances on the full set of the CSIQ, the LIVE, and the TID2013. For non-deep-learning-based FR-IQA methods, we directly assess their performance.

In this experiment, our UniDASTN is compared with a total of 14 FR-IQA methods, including 9 non-deep-learning-based methods and 5 deep-learning-based methods. The nine non-deep-learning-based FR-IQA methods are PSNR [46], SSIM [35], MS-SSIM [8], VSI [22], MAD [18], VIF [19], FSIMc [23], NLPD [17], and GMSD [21]. The five deep-learning-based methods are DISTS [30], IQT [42], PieAPP [31], LPIPS-Alex [32], and LPIPS-VGG [32].

The PLCC, SRCC, and KRCC results of these methods are shown in Table 4, where the best results are in bold. In this cross-dataset experiment, it can be seen that our UniDASTN consistently outperforms all compared FR-IQA methods in terms of PLCC, SRCC, and KRCC. In particular, for TID2013, our UniDASTN shows a significant performance improvement compared to some state-of-the-art FR-IQA methods. This experiment illustrates that our UniDASTN achieves satisfactory generalization ability.

Table 4.
MethodCSIQLIVETID2013
PLCCSRCCKRCCPLCCSRCCKRCCPLCCSRCCKRCC
PSNR [46]0.7960.8060.6080.8720.8760.6870.7020.6390.470
SSIM [35]0.8610.8760.6900.9450.9480.7960.7900.7420.559
MS-SSIM [8]0.8990.9130.7390.9490.9510.8040.8320.7860.605
VSI [22]0.9280.9420.7860.9480.9520.8050.9000.8970.718
MAD [18]0.9500.9510.8020.9600.9610.8240.7960.7730.587
VIF [19]0.9280.9200.7540.9600.9640.8280.7720.6770.515
FSIMc [23]0.9190.9310.7690.9610.9650.8360.8770.8510.667
NLPD [17]0.9230.9320.7690.9320.9370.7780.8390.8000.625
GMSD [21]0.9540.9570.8120.9600.9600.8270.8590.8040.634
DISTS [30]0.9360.9390.7800.9510.9540.8110.8680.8300.639
IQT [42]0.9430.9390.7990.9700.9680.8490.8990.8910.717
PieAPP [31]0.8900.8970.7120.9100.9180.7490.8310.8440.657
LPIPS-Alex [32]0.8930.9280.7570.9000.9220.7510.7870.7690.568
LPIPS-VGG [32]0.9010.8840.6980.9380.9410.7800.7430.6790.505
Our UniDASTN0.9690.9670.8270.9760.9730.8610.9170.9050.724

Table 4. Performance Comparison on Cross-Dataset Experiment

In addition, it can be seen that the performances of some compared methods are not stable among different databases. For example, the VSI [22] achieves good results in all metrics on the TID2013 database, but its performances on the CSIQ and LIVE databases degrade significantly. Similarly, the deep-learning-based method called IQT [42] reaches good results for all metrics on the LIVE database. However, its performances on the other two databases also degrade slightly. For our UniDASTN, it can make consistent performance among the three databases. This also illustrates the effectiveness of our UniDASTN in evaluating the qualities of various distorted images.

Furthermore, Figure 5 displays scatter plots of subjective MOS values versus objective prediction scores for our UniDASTN and the 14 FR-IQA methods on the TID2013 database, as well as the fitted curves. In this figure, the x-axis is the objective prediction score generated by a specific FR-IQA method, the y-axis is the MOS value, and each point on the plot represents one image in the database. The straight line is fitted by simple linear regression. If the correlation is strong, the points in the graph are tightly clustered around the curve. If the correlation is weak, the points are dispersed around the curve. As can be seen from the Figure 5, the points of our UniDASTN are much closer to the fitted curve than those of the compared FR-IQA methods, implying a stronger consistency with the subjective scores.

Fig. 5.

Fig. 5. Scatter plots of different FR-IQA methods on the TID2013 database.

To analyze the statistical significance of our UniDASTN relative to the compared methods, we employ the left-tailed F-test [21, 53] to decide whether or not an FR-IQA method is statistically superior to another FR-IQA method. We perform a series of hypothesis tests on the residuals between the subjective quality scores and the objective quality scores obtained from each FR-IQA method after nonlinear regression. Figures 6(a)–(c) show the statistically significant results on the LIVE, CSIQ, and TID2013 databases, respectively. In this figure, the value of H = 1 for the left-tailed F-test represents that the FR-IQA method in that row is significantly superior to the method in the corresponding column with a confidence level of more than 95% at the 0.05 significance level. The value of H = 0 means that the FR-IQA method in that row is not significantly superior to the FR-IQA method in the corresponding column. The symbol “-” means that the two FR-IQA methods in the corresponding row and column are statistically indistinguishable. From the statistical significance results of the LIVE, CSIQ, and TID2013 databases, it can be seen that our UniDASTN significantly outperforms some state-of-the-art FR-IQA methods and achieves the best performances on the three databases.

Fig. 6.

Fig. 6. Statistical significance tests on different databases.

4.5 Ablation Study

To assess the effectiveness of the network components used in the UniDASTN, we perform a detailed performance analysis of networks with various network component combinations.

Effectiveness of our channel attention module. To assess the effectiveness of the proposed channel attention module, we perform an ablation study on the CSIQ and TID2013 datasets. The results are shown in Table 5, where the best results in bold, and N1–N2 represent our networks with different usage strategies of channel attention module. Specifically, the N1 network means that the channel attention module is not used, and the N2 network means that the channel attention module is used. From Table 5, we can see that the N2 network produces the best PLCC and SRCC outcomes. These experimental results demonstrate that the channel attention module plays a significant role in our UniDASTN. The use of the channel attention module can result in more accurate quality prediction results.

Table 5.
NetworkStrategyCSIQTID2013
PLCC/SRCCPLCC/SRCC
N1×0.983/0.9790.975/0.971
N2\(\surd\)0.987/0.9830.977/0.974

Table 5. Comparison of the Use of Channel Attention Module

Effectiveness of our loss function. To assess the effectiveness of the proposed joint loss function, we perform an ablation study on the CSIQ and TID2013 datasets. The results are shown in Table 6, where the best results in bold, and M1–M5 represent our networks with different loss functions. Specifically, in the experiment, the M5 network with MSE loss, bidirectional KL divergence loss, and rank loss produces the best PLCC and SRCC outcomes. These experimental results demonstrate that both the bidirectional KL divergence loss and the rank loss contribute to the performance improvement of our UniDASTN.

Table 6.
NetworkLoss functionCSIQTID2013
MSEBKLRankPLCC/SRCCPLCC/SRCC
M1\(\surd\)0.979/0.9720.971/0.967
M2\(\surd\)0.973/0.9740.968/0.967
M3\(\surd\)\(\surd\)0.981/0.9750.972/0.968
M4\(\surd\)\(\surd\)0.975/0.9770.970/0.969
M5\(\surd\)\(\surd\)\(\surd\)0.987/0.9830.977/0.974

Table 6. Comparison of Different Loss Functions

In addition, the typical visual results of PSNR, SSIM, DISTS, and our UniDASTN are demonstrated in Figure 7. It can be seen that the quality rank of the distorted image predicted by our UniDASTN is consistent with the MOS result. This verifies that our UniDASTN can distinguish distortion types with similar scores.

Fig. 7.

Fig. 7. For each distorted image, MOS represents the subjective rating, and the scores corresponding to each FR-IQA method represent its predicted results. The number in parentheses indicates the quality rank of the distorted image.

Effectiveness of our feature extraction backbone. To assess the effectiveness of our feature extraction backbones, we perform an ablation study on the CSIQ and TID2013 datasets. The results are shown in Table 7, where the best results are in bold, and M6–M8 represent our networks using different backbones. Specifically, the M6 network uses the ResNet50 as the feature extraction backbone, the M7 network uses the Inception-Resnet-V2 as the feature extraction backbone, and the M8 network uses the ViT-B\(/\)8 as the feature extraction backbone. It can be seen that the M8 network makes the best performances on the two datasets. These experimental results demonstrate that the selection of the ViT-B\(/\)8 is beneficial and effective in FR-IQA.

Table 7.
NetworkBackboneCSIQTID2013
PLCC/SRCCPLCC/SRCC
M6Resnet500.844/0.8390.772/0.769
M7Inception-Resnet-V20.966/0.9610.958/0.950
M8ViT-B/80.987/0.9830.977/0.974

Table 7. Comparison of Different Feature Extraction Backbones

Effectiveness of our feature fusion technique. To assess the effectiveness of our feature fusion technique, we perform an ablation study on the CSIQ and TID2013 datasets using five different feature fusion techniques. Figure 8 shows the five techniques used for feature fusion, where Scheme (a) is the fusion technique reported in [43, 67], Scheme (b) is the fusion technique used in [58], Scheme (c) is the proposed feature fusion block, Scheme (d) is derived from Scheme (c) by adding the average pooling operation, and Scheme (e) is derived from Scheme (c) by replacing the operation of “concatenate” with “add”. Note that for the fusion techniques presented in Scheme (a) and Scheme (b), their outputs are the quality scores. These experimental results are provided in Table 8.

Fig. 8.

Fig. 8. Different fusion techniques, where (a) and (b) are the reported techniques, (c) is the proposed feature fusion module, (d) is derived from (c) by adding the pooling operation, and (e) is derived from (c) by replacing the operation of “concatenate” with “add”.

Table 8.
SchemeCSIQTID2013
PLCC/SRCCPLCC/SRCC
scheme (a)0.984/0.9800.973/0.971
scheme (b)0.980/0.9750.970/0.966
scheme (c)0.987/0.9830.977/0.974
scheme (d)0.983/0.9800.972/0.969
scheme (e)0.968/0.9660.949/0.953

Table 8. Comparison of Different Fusion Techniques

Table 8 shows that the fusion technique of Scheme (c) makes the best PLCC and SRCC results. In a comparison of Scheme (a) and Scheme (c), an additional Transformer encoder used to fuse the extracted features can improve the prediction performance. This is because the additional transformer encoder can act as an aggregator to aggregate the information of each token. Moreover, in a comparison of Scheme (c) and Scheme (d) as well as Scheme (c) and Scheme (e), if the pooling operation is added or the operation of “concatenate” is replaced with “add”, the information of these tokens is inevitably lost and thus the FR-IQA performance decreases. These experimental results demonstrate that our used feature fusion technique can result in more accurate quality prediction.

Effectiveness of our siamese transformer network. To assess the effectiveness of our Siamese transformer network, we perform an ablation study on the CSIQ and TID2013 datasets. The results are shown in Table 9, where the best results are in bold, and the schemes 1–3 represent different input combinations used in the spatial attention module. There are three types of inputs used in our siamese transformer network, namely \(z_{0}^{ref}, z_{0}^{dist},\) and \(z_{0}^{diff}.\) If the number of inputs decreases to 2 (i.e., only \(z_{0}^{ref}\) and \(z_{0}^{diff}\) or \(z_{0}^{dist}\) and \(z_{0}^{diff}\) is input to the spatial attention module), there are two technical schemes in total. Specifically, the scheme 1 selects \(z_{0}^{ref}\) and \(z_{0}^{diff}\) as the inputs of the spatial attention module, and the scheme 2 uses \(z_{0}^{dist}\) and \(z_{0}^{diff}\) as the inputs of the spatial attention module. This means that, for the schemes 1–2, only one branch of the siamese transformer network is used in the spatial attention module. For comparison, the proposed scheme is denoted as the scheme 3 in which \(z_{0}^{ref}, z_{0}^{dist},\) and \(z_{0}^{diff}\) are all selected as the inputs of the spatial attention module. From Table 9, it can be seen that the scheme 3 makes the best results. These experimental results demonstrate that the features extracted by the two branches of our siamese transformer network are highly complementary. The scheme 1 and the scheme 2 only use the features of one branch. For the scheme 3, the features of two branches are both used and help to make good IQA whole performance.

Table 9.
SchemeInputs of SAMCSIQTID2013
\(z_{0}^{ref}\)\(z_{0}^{dist}\)\(z_{0}^{diff}\)PLCC/SRCCPLCC/SRCC
1\(\surd\)\(\surd\)0.983/0.9780.963/0.965
2\(\surd\)\(\surd\)0.986/0.9810.964/0.963
3\(\surd\)\(\surd\)\(\surd\)0.987/0.9830.977/0.974

Table 9. Comparison of Different Input Numbers of Spatial Attention Module

To make easy understanding of the attention mechanism of Transformer [70], some visual examples of our Siamese transformer network are shown in Figure 9, where three distorted images and their corresponding attention maps of different heads are presented. In this experiment, the attention map of an image is obtained by averaging all attention weights in the self-attention module and resizing to the original image size. It can be found that different heads have different preferences in their attention maps. Through the collaboration of these heads, our UniDASTN can simulate the HVS to make an accurate quality prediction. Furthermore, visual examples of the two branches \(dist\sim diff\) (up branch) and \(ref\sim diff\) (down branch) of our Siamese transformer network are also illustrated. Figure 10 presents the attention maps of the Head 5, where the attention map of the up branch, the attention map of the down branch, and the attention map of two branches are all shown. It can be observed that the attentions of the two branches are spatially placed in different regions. This illustrates that the use of both branches can help in perceiving image quality.

Fig. 9.

Fig. 9. Visualization of attention maps.

Fig. 10.

Fig. 10. Attention maps of different branches.

Effectiveness of the inputs of siamese transformer network. To assess the effectiveness of our strategies for the input types, we perform an ablation study on the CSIQ and TID2013 datasets. The results are shown in Table 10, where the best results are in bold. Note that there are three embedding features feed into the siamese transformer network, namely \(z_{0}^{ref}, z_{0}^{dist},\) and \(z_{0}^{diff}\). Therefore, there are three schemes input of siamese transformer network in total. Specifically, for the Scheme 4, \(z_{0}^{diff}\) is input to the transformer encoder, and \(z_{0}^{dist}\) and \(z_{0}^{ref}\) are input to the transformer decoder. For the Scheme 5, \(z_{0}^{ref}\) is input to the transformer encoder, and \(z_{0}^{dist}\) and \(z_{0}^{diff}\) are input to the transformer decoder. For the Scheme 6, \(z_{0}^{dist}\) is input to the transformer encoder, and \(z_{0}^{ref}\) and \(z_{0}^{diff}\) are input to the transformer decoder. It can be seen that the Scheme 4 makes three best results and one second-best result. Therefore, the Scheme 4 is better than the Scheme 5 and the Scheme 6. These experimental results demonstrate that, as to the input of transformer encoder, the embedding features of “perceptual differences” are preferable to the reference or distortion embedding features. The aggregation of the embedding features of “perceptual differences” with the transformer encoder helps to improve IQA performance.

Table 10.
SchemeEncoderDecoderCSIQTID2013
\(z_{0}^{dist}\)\(z_{0}^{ref}\)\(z_{0}^{diff}\)\(z_{0}^{dist}\)\(z_{0}^{ref}\)\(z_{0}^{diff}\)PLCC/SRCCPLCC/SRCC
4\(\surd\)\(\surd\)\(\surd\)0.987/0.9830.977/0.974
5\(\surd\)\(\surd\)\(\surd\)0.987/0.9840.970/0.960
6\(\surd\)\(\surd\)\(\surd\)0.986/0.9810.972/0.969

Table 10. Comparison of Different Input Types of Spatial Attention Module

Skip 5CONCLUSIONS Section

5 CONCLUSIONS

We have proposed a novel method for FR-IQA by UniDASTN. A key contribution is the spatial attention module composed of a Siamese Transformer network and a feature fusion block. It can focus on significant regions and thus effectively maps the perceptual differences between the reference and distorted images to a latent distance for distortion evaluation. The second contribution is the dual-attention strategy which exploits a squeeze-and-excitation network and the proposed spatial attention module to extract much more significant features for accurate IQA. The third contribution is the new joint loss function that combines the MSE, Bidirectional KL divergence, and rank order of quality scores. It can provide stable training and thus enables the proposed UniDASTN to learn visual perceptual image quality effectively. Many experiments on some standard IQA databases have been conducted to verify the IQA performance of the proposed UniDASTN. The comparison results have demonstrated that the proposed UniDASTN outperforms some state-of-the-art FR-IQA methods on the image databases of LIVE, CSIQ, TID2013, and PIPAL according to three well-known IQA metrics.

Skip ACKNOWLEDGMENTS Section

ACKNOWLEDGMENTS

Many thanks to the reviewers for their helpful suggestions.

REFERENCES

  1. [1] Feng Ping and Tang Zhenjun. 2023. A survey of visual neural networks: Current trends, challenges and opportunities. Multimedia Systems 29, 2 (2023), 693724.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. [2] Zhai Guangtao and Min Xiongkuo. 2020. Perceptual image quality assessment: A survey. Science China Information Sciences 63, 11 (2020), 152.Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Henderson John M.. 2003. Human gaze control during real-world scene perception. Trends in Cognitive Sciences 7, 11 (2003), 498504.Google ScholarGoogle ScholarCross RefCross Ref
  4. [4] Zhai Guangtao, Sun Wei, Min Xiongkuo, and Zhou Jiantao. 2021. Perceptual quality assessment of low-light image enhancement. ACM Transactions on Multimedia Computing, Communications, and Applications 17, 4 (2021), Article 130, 1–24.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Shi Ran, Ma Jing, Ngan King Ngi, Xiong Jian, and Qiao Tong. 2022. Objective object segmentation visual quality evaluation: Quality measure and pooling method. ACM Transactions on Multimedia Computing, Communications, and Applications 18, 3 (2022), Article 73, 1–19.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Chen Chenglizhao, Zhao Hongmeng, Yang Huan, Yu Teng, Peng Chong, and Qin Hong. 2021. Full-reference screen content image quality assessment by fusing multilevel structure similarity. ACM Transactions on Multimedia Computing, Communications, and Applications 17, 3 (2021), Article 94, 1–21.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Liang Zhiyuan Chen, Yihua Chen, Xiaoping, and Tang Zhenjun. 2022. Multi-level feature aggregation network for full-reference image quality assessment. In Proceedings of the IEEE 34th International Conference on Tools with Artificial Intelligence. 861867.Google ScholarGoogle Scholar
  8. [8] Wang Zhou, Simoncelli Eero P., and Bovik Alan Conrad. 2003. Multiscale structural similarity for image quality assessment. In Proceedings of the 37th Asilomar Conference on Signals, Systems & Computers, Vol. 2. 13981402.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Huang Ziqing and Liu Shiguang. 2021. Perceptual hashing with visual content understanding for reduced-reference screen content image quality assessment. IEEE Transactions on Circuits and Systems for Video Technology 31, 7 (2021), 28082823.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Tang Zhenjun, Huang Ziqing, Yao Heng, Zhang Xianquan, Chen Lv, and Yu Chunqiang. 2018. Perceptual image hashing with weighted DWT features for reduced-reference image quality assessment. Computer Journal 61, 11 (2018), 16951709.Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Yu Mengzhu, Tang Zhenjun, Zhang Xianquan, Zhong Bineng, and Zhang Xinpeng. 2022. Perceptual hashing with complementary color wavelet transform and compressed sensing for reduced-reference image quality assessment. IEEE Transactions on Circuits and Systems for Video Technology 32, 11 (2022), 75597574.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. [12] Liu Yutao, Gu Ke, Li Xiu, and Zhang Yongbing. 2020. Blind image quality assessment by natural scene statistics and perceptual characteristics. ACM Transactions on Multimedia Computing, Communications, and Applications 16, 3 (2020), Article 91, 1–20.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Yan Chenggang, Teng Tong, Liu Yutao, Zhang Yongbing, Wang Haoqian, and Ji Xiangyang. 2021. Precise no-reference image quality evaluation based on distortion identification. ACM Transactions on Multimedia Computing, Communications, and Applications 17, 3 (2021), Article 110, 1–21.Google ScholarGoogle Scholar
  14. [14] Zhang Chaofan, Huang Ziqing, Liu Shiguang, and Xiao Jian. 2022. Dual-channel multi-task CNN for no-reference screen content image quality assessment. IEEE Transactions on Circuits and Systems for Video Technology 32, 8 (2022), 50115025.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Wang Zhou and Bovik Alan Conrad. 2009. Mean squared error: Love it or leave it? A new look at signal fidelity measures. IEEE Signal Processing Magazine 26, 1 (2009), 98117.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Gu Ke, Li Leida, Lu Hong, Min Xiongkuo, and Lin Weisi. 2017. A fast reliable image quality predictor by fusing micro-and macro-structures. IEEE Transactions on Industrial Electronics 64, 5 (2017), 39033912.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Laparra Valero, Balle Johannes, Berardino Alexander, and Simoncelii Eero P.. 2016. Perceptual image quality assessment using a normalized Laplacian pyramid. In Proceedings of the Human Vision and Electronic Imaging. 4348.Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Larson Eric Cooper and Chandler Damon Michael. 2010. Most apparent distortion: Full-reference image quality assessment and the role of strategy. Journal of Electronic Imaging 19, 1 (2010), Article ID. 011006.Google ScholarGoogle Scholar
  19. [19] Sheikh Hamid Rahim and Bovik Alan Conrad. 2006. Image information and visual quality. IEEE Transactions on Image Processing 15, 2 (2006), 430444.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Sheikh Hamid Rahim, Bovik Alan Conrad, and Veciana Gustavo De. 2005. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Transactions on Image Processing 14, 12 (2005), 21172128.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. [21] Xue Wufeng, Zhang Lei, Mou Xuanqin, and Bovik Alan Conrad. 2013. Gradient magnitude similarity deviation: A highly efficient perceptual image quality index. IEEE Transactions on Image Processing 23, 2 (2013), 684695.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] Zhang Lin, Shen Ying, and Li Hongyu. 2014. VSI: A visual saliency-induced index for perceptual image quality assessment. IEEE Transactions on Image Processing 23, 10 (2014), 42704281.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Zhang Lin, Zhang Lei, Mou Xuanqin, and Zhang David. 2011. FSIM: A feature similarity index for image quality assessment. IEEE Transactions on Image Processing 20, 8 (2011), 23782386.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. [24] Kang Le, Ye Peng, Li Yi, and Doermann David. 2014. Convolutional neural networks for no-reference image quality assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 17331740.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Bosse Sebastian, Maniry Dominique, Müller Klaus-Robert, Wiegand Thomas, and Samek Wojciech. 2017. Deep neural networks for no-reference and full-reference image quality assessment. IEEE Transactions on Image Processing 27, 1 (2017), 206219.Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Bosse Sebastian, Maniry Dominique, Wiegand Thomas, and Samek Wojciech. 2016. A deep neural network for image quality assessment. In Proceedings of the IEEE International Conference on Image Processing. 37733777.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Kim Jongyoo and Lee Sanghoon. 2017. Deep learning of human visual sensitivity in image quality assessment framework. In Proceedings of the IEEE/CVF Conference on Computer Vision & Pattern Recognition. 16761684.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Ahn Sewoong, Choi Yeji, and Yoon Kwangjin. 2021. Deep learning-based distortion sensitivity prediction for full-reference image quality assessment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 344353.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Seo Soomin, Ki Sehwan, and Kim Munchurl. 2021. A novel just-noticeable-difference-based saliency-channel attention residual network for full-reference image quality predictions. IEEE Transactions on Circuits and Systems for Video Technology 31, 7 (2021), 26022616.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Ding Keyan, Ma Kede, Wang Shiqi, and Simoncelli Eero P.. 2022. Image quality assessment: Unifying structure and texture similarity. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 5 (2022), 25672581.Google ScholarGoogle Scholar
  31. [31] Prashnani Ekta, Cai Hong, Mostofi Yasamin, and Sen Pradeep. 2018. PieAPP: Perceptual image-error assessment through pairwise preference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 18081817.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Zhang Richard, Isola Phillip, Efros Alexei A., Shechtman Eli, and Wang Oliver. 2018. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 586595.Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Goodfellow Ian J., Pouget-Abadie Jean, Mirza Mehdi, Xu Bing, Warde-Farley David, Ozair Sherjil, Courville Aaron, and Bengio Yoshua. 2014. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Vol. 2. 26722680.Google ScholarGoogle Scholar
  34. [34] Gu Jinjin, Cai Haoming, Chen Haoyu, Ye Xiaoxing, Ren Jimmy S., and Dong Chao. 2020. PIPAL: A large-scale image quality assessment dataset for perceptual image restoration. In Proceedings of the European Conference on Computer Vision. 633651.Google ScholarGoogle Scholar
  35. [35] Wang Zhou, Bovik Alan Conrad, Sheikh Hamid Rahim, and Simoncelli Eero P.. 2004. Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing 13, 4 (2004), 600612.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. [36] Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N., Kaiser Łukasz, and Polosukhin Illia. 2017. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems. 60006010.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Khan Salman, Naseer Muzammal, Hayat Munawar, Zamir Syed Waqas, Khan Fahad Shahbaz, and Shah Mubarak. 2022. Transformers in vision: A survey. ACM Computing Surveys 54, 10 (2022), Article 200, 1–41.Google ScholarGoogle Scholar
  38. [38] Dosovitskiy Alexey, Beyer Lucas, Kolesnikov Alexander, Weissenborn Dirk, Zhai Xiaohua, Unterthiner Thomas, Dehghani Mostafa, Minderer Matthias, Heigold Georg, Gelly Sylvain, Uszkoreit Jakob, and Houlsby Neil. 2021. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  39. [39] He Kaiming, Chen Xinlei, Xie Saining, Li Yanghao, Dollár Piotr, and Girshick Ross. 2022. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1600016009.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Ze Liu, Han Hu, Yutong Lin, Zhuliang Yao, Zhenda Xie, Yixuan Wei, Jia Ning, Yue Cao, Zheng Zhang, Li Dong, Furu Wei, and Baining Guo. 2022. Swin transformer V2: Scaling up capacity and resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 12009–12019.Google ScholarGoogle Scholar
  41. [41] Wang Zhendong, Cun Xiaodong, Bao Jianmin, Zhou Wengang, Liu Jianzhuang, and Li Houqiang. 2022. Uformer: A general U-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1768317693.Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Cheon Manri, Yoon Sung-Jun, Kang Byungyeon, and Lee Junwoo. 2021. Perceptual image quality assessment with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 433442.Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Golestaneh S. Alireza, Dadsetan Saba, and Kitani Kris M.. 2022. No-reference image quality assessment via transformers, relative ranking, and self-consistency. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 39893999.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Ke Junjie, Wang Qifei, Wang Yilin, Milanfar Peyman, and Yang Feng. 2021. MUSIQ: Multi-scale image quality transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 51485157.Google ScholarGoogle ScholarCross RefCross Ref
  45. [45] You Junyong and Korhonen Jari. 2021. Transformer for image quality assessment. In Proceedings of the IEEE International Conference on Image Processing. 13891393.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Avcibas Ismail, Sankur Bulent, and Sayood Khalid. 2002. Statistical evaluation of image quality measures. Journal of Electronic Imaging 11, 2 (2002), 206223.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Ma Siwei, Gao Wen, and Lu Yan. 2005. Rate-distortion analysis for H. 264/AVC video coding and its application to rate control. IEEE Transactions on Circuits and Systems for Video Technology 15, 12 (2005), 15331544.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. [48] Jian Liang, Dapeng Hu, Jiashi Feng, and Ran He. 2022. DINE: Domain adaptation from single and multiple black-box predictors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 7993–8003.Google ScholarGoogle Scholar
  49. [49] Benesty Jacob, Chen Jingdong, Huang Yiteng, and Cohen Israel. 2009. Pearson Correlation Coefficient. Springer, 14.Google ScholarGoogle Scholar
  50. [50] Zar Jerrold H.. 2005. Spearman rank correlation. In Encyclopedia of Biostatistics, Peter Armitage and Theodore Colton (Eds.). Vol. 7, Wiley, Hoboken, NJ.Google ScholarGoogle ScholarCross RefCross Ref
  51. [51] Abdi Hervé. 2007. The Kendall rank correlation coefficient. In Encyclopedia of Measurement and Statistics, Neil J. Salkind (Ed.). Sage, 508510.Google ScholarGoogle Scholar
  52. [52] Nikolay Ponomarenko, Lina Jin, Oleg Ieremeiev, Vladimir Lukin, Karen Egiazarian, Jaakko Astola, Benoit Vozel, Kacem Chehdi, Marco Carli, Federica Battisti, and Chung Chieh Jay Kuo. 2015. Image database TID2013: Peculiarities, results and perspectives. Signal Processing: Image Communication 30 (2015), 57–77.Google ScholarGoogle Scholar
  53. [53] Sheikh Hamid Rahim, Sabir Muhammad F., and Bovik Alan Conrad. 2006. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Transactions on Image Processing 15, 11 (2006), 34403451.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. [54] Lin Hanhe, Hosu Vlad, and Saupe Dietmar. 2019. KADID-10k: A large-scale artificially distorted IQA database. In Proceedings of the 11th International Conference on Quality of Multimedia Experience. 13.Google ScholarGoogle ScholarCross RefCross Ref
  55. [55] Bae Sung Ho and Kim Munchurl. 2016. A novel image quality assessment with globally and locally consilient visual quality perception. IEEE Transactions on Image Processing 25, 5 (2016), 23922406.Google ScholarGoogle ScholarDigital LibraryDigital Library
  56. [56] Blau Yochai and Michaeli Tomer. 2018. The perception-distortion tradeoff. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 62286237.Google ScholarGoogle ScholarCross RefCross Ref
  57. [57] Damera-Venkata Niranjan, Kite Thomas D., Geisler Wilson S., Evans Brian L., and Bovik Alan Conrad. 2000. Image quality assessment based on a degradation model. IEEE Transactions on Image Processing 9, 4 (2000), 636650.Google ScholarGoogle ScholarCross RefCross Ref
  58. [58] Guo Haiyang, Bin Yi, Hou Yuqing, Zhang Qing, and Luo Hengliang. 2021. IQMA network: Image quality multi-scale assessment network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 443452.Google ScholarGoogle ScholarCross RefCross Ref
  59. [59] Lao Shanshan, Gong Yuan, Shi Shuwei, Yang Sidi, Wu Tianhe, Wang Jiahao, Xia Weihao, and Yang Yujiu. 2022. Attentions help CNNs see better: Attention-based hybrid image quality assessment network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 11401149.Google ScholarGoogle ScholarCross RefCross Ref
  60. [60] Zhang Lin and Li Hongyu. 2012. SR-SIM: A fast and high performance IQA index based on spectral residual. In Proceedings of the 19th IEEE International Conference on Image Processing. 14731476.Google ScholarGoogle ScholarCross RefCross Ref
  61. [61] Liu Anmin, Lin Weisi, and Narwaria Manish. 2012. Image quality assessment based on gradient similarity. IEEE Transactions on Image Processing 21, 4 (2012), 15001512.Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. [62] Ma Chao, Yang Chih-Yuan, Yang Xiaokang, and Yang Ming-Hsuan. 2017. Learning a no-reference quality metric for single-image super-resolution. Computer Vision and Image Understanding 158 (2017), 116. https://www.sciencedirect.com/science/article/abs/pii/S107731421630203X.Google ScholarGoogle ScholarDigital LibraryDigital Library
  63. [63] Mittal Anish, Soundararajan Rajiv, and Bovik Alan Conrad. 2013. Making a “Completely Blind” image quality analyzer. IEEE Signal Processing Letters 20, 3 (2013), 209212.Google ScholarGoogle ScholarCross RefCross Ref
  64. [64] Mittal Anish, Moorthy Anush K., and Bovik Alan Conrad. 2011. Blind/referenceless image spatial quality evaluator. In Proceedings of the 45th Asilomar Conference on Signals, Systems and Computers. 723727.Google ScholarGoogle ScholarCross RefCross Ref
  65. [65] Pei Soo-Chang and Chen Li-Heng. 2015. Image quality assessment using human visual DOG model fused with random forest. IEEE Transactions on Image Processing 24, 11 (2015), 32823292.Google ScholarGoogle ScholarDigital LibraryDigital Library
  66. [66] Varga Domonkos. 2020. Composition-preserving deep approach to full-reference image quality assessment. Signal, Image and Video Processing 14, 6 (2020), 12651272.Google ScholarGoogle ScholarCross RefCross Ref
  67. [67] Zhang Hainan, Meng Fang, and Han Yawen. No-reference image quality assessment based on a multi-feature extraction network. In Proceedings of the 2nd International Conference on Image, Video and Signal Processing. 8185.Google ScholarGoogle Scholar
  68. [68] Zhang Lin, Zhang Lei, and Mou Xuanqin. 2010. RFSIM: A feature based image quality assessment metric using Riesz transforms. In Proceedings of IEEE International Conference on Image Processing. 321324.Google ScholarGoogle ScholarCross RefCross Ref
  69. [69] Wang Zhou and Bovik Alan Conrad. 2002. A universal image quality index. IEEE Signal Processing Letters 9, 3 (2002), 8184.Google ScholarGoogle ScholarCross RefCross Ref
  70. [70] He Yangji, Liang Weihan, Zhao Dongyang, Zhou Hong-Yu, Ge Weifeng, Yu Yizhou, and Zhang Wenqiang. 2022. Attribute surrogates learning and spectral tokens pooling in transformers for few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 91199129.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Unifying Dual-Attention and Siamese Transformer Network for Full-Reference Image Quality Assessment

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Multimedia Computing, Communications, and Applications
        ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 19, Issue 6
        November 2023
        858 pages
        ISSN:1551-6857
        EISSN:1551-6865
        DOI:10.1145/3599695
        • Editor:
        • Abdulmotaleb El Saddik
        Issue’s Table of Contents

        Copyright © 2023 Copyright held by the owner/author(s).

        This work is licensed under a Creative Commons Attribution-NonCommercial International 4.0 License.

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 12 July 2023
        • Online AM: 18 May 2023
        • Accepted: 9 May 2023
        • Revised: 25 March 2023
        • Received: 21 December 2022
        Published in tomm Volume 19, Issue 6

        Check for updates

        Qualifiers

        • research-article

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader