Elsevier

Neurocomputing

Volume 139, 2 September 2014, Pages 310-320
Neurocomputing

Single image super-resolution via subspace projection and neighbor embedding

https://doi.org/10.1016/j.neucom.2014.02.026Get rights and content

Abstract

In this paper, we present a novel learning-based single image super-resolution algorithm to address the problems of inefficient learning and improper estimation in coping with nonlinear high-dimensional feature data. Our method named as subspace projection and neighbor embedding (SPNE) first projects the high-dimensional data into two different subspaces respectively, i.e., kernel principal component analysis (KPCA) subspace and modified locality preserving projection (MLPP) subspace to obtain the global and local structures of data. In an optimal low-dimensional feature space, the k-nearest neighbors of each input low-resolution (LR) image patch can be found for efficient learning. Then within similarity measures and proportional factors, the k embedding weights are used to estimate high-frequency information from a training dataset. Finally, we apply iterative back projection (IBP) to further enhance the super-resolution results. Experiments on simulative and actual LR images demonstrate that the proposed approach outperforms the existing NE-based super-resolution methods in terms of visual quality and some selected objective metrics.

Introduction

.In many practical applications, such as computer vision, medical imaging, safety surveillance, and preprocessing of hyper-spectral fusion, high resolution (HR) images are usually desired to deal with. However, images from the existing digital acquisition system do not meet people׳s requirement. In order to alleviate the contradiction, image super-resolution (SR) reconstruction technique becomes more and more popular to estimate a target HR image from one or more low resolution (LR) images [1], [2].

In recent years, image super-resolution algorithms have achieved great success. Generally speaking, those methods can be separated into three categories: interpolation-based [3], [4], multi-frame-based [5], [6], [7], and learning-based [8]. Although interpolation-based methods can enlarge the size of image in any magnification factor, the results often have blurring artifacts. The multi-frame-based methods use a simplified imaging degradation model with some terms (e.g., motion warping, optical dimming, low sampling, and random noise). However, these methods often depend on registration precision among LR images and have difficulty in estimating a HR image when magnification factor is large. Relatively, learning-based methods do not face the above problems. The goal of these methods is to estimate the missing high frequency information from a training set consisting of many LR–HR image patch pairs. Freeman et al. [9] first proposed a learning-based SR algorithm using a Markov random field model to learn the relationship between LR and corresponding HR image patches. But this method is very sensitive to examples in the training set. Chang et al. [10] developed a neighbor embedding (NE) based method, which introduced locally linear embedding (LLE) to estimate high frequency details linearly combining the k-nearest neighbors in the training set. Then, many researchers extended the NE-based method [11], [12], [13], [14], [15], [16], [17]. Chan et al. [13] emphasized on edge detection and feature selection using first-order gradient feature and normalized luminance feature together to find the k-nearest neighbors. In order to represent complicated texture structures, Zhang et al. [14] proposed a novel NE-based SR algorithm through a partially supervised distance measurement with class information. Recently, Gao et al. [15] applied a joint learning technique to map LR and HR feature spaces onto a unified feature subspace by grouping patch pairs. Yang et al. [18] presented a sparse coding based SR approach to generate HR output from an over-complete dictionary. Then, they proposed a coupled dictionary training method [29] for reconstructing well the underlying HR image patches within the coupled dictionary. Jia et al. [30] introduced a local parametric regression approach over the learned coupled dictionaries across source and target image spaces using sparse representation. Gao et al. [31] presented a sparse neighbor selection scheme simultaneously combining neighbor search process and reconstruction weights estimation process.

In order to represent LR image patches, we need to extract their texture features. Due to high dimensionality [19], the extracted features are often nonlinear. Simply applying linear transformation for high-dimensional data may lead to an over-fitting problem [20] during the training process, as well as provide poor results. Kernel techniques [21] are capable of dealing with nonlinear problem. However, after applying kernel techniques, the feature data has higher dimensionality and greater redundancy as well. To further address this problem, dimensionality reduction technique is a feasible way. In general, principal component analysis (PCA) is one of representative global dimensionality reduction methods, while locality preserving projection (LPP) is a local dimensionality reduction method [22]. Global and local structures are both significant for learning high-dimensional data. Therefore, we exploit kernel PCA (KPCA) and modified LPP (MLPP) to reserve both global and local structures of high-dimensional feature data. Within the proposed approach, we can adaptively find the 0k-nearest neighbors with higher matching precision for each test LR patch.

Our approach is based on NE, representing a HR image patch as a linear combination of its neighbors. Here we propose a novel subspace projection and neighbor embedding (SPNE) for single image super-resolution reconstruction. The neighbor search is implemented by KPCA transformation, followed by MLPP transformation. Then, embedding weights are estimated by a new metric, combining similarity measures and proportional factors. Using the learnt high-frequency details, the target HR image patches can be obtained corresponding to the test LR image patches. In order to achieve better results in SR reconstruction, we apply iterative back projection (IBP) for enhancement. The proposed approach preserves global and local structures of feature data using KPCA and MLPP transformations. For each test LR image patch, its k-nearest neighbors can be searched with high precision. The novel metric for embedding weights is crucial in estimating target HR image patches. Experimental results on simulative and actual test images demonstrate that the proposed SPNE algorithm achieves much higher SR reconstruction performance than the existing SR methods [13], [14], [15].

The remainder of this paper is organized as follows. Section 2 describes kernel principal component analysis (KPCA). The SPNE-based image super-resolution algorithm is proposed in Section 3. Sections 4 and 5 present experimental results and conclusion, respectively.

Section snippets

Kernel principal component analysis

To overcome nonlinear problem in high-dimensional data, we apply kernel techniques mapping the high-dimensional data to higher-dimensional data. Let X=[x1,,xn] be high-dimensional data where xid, and n be the number of selected feature vectors. Consider a nonlinear mapping Φ that relates the input space d to another space F, i.e., Φ:dF. The mapping data becomes linearly related in the feature space F. The KPCA transformation [24] is briefly described as follows.

First, the mapping data Φ(xi

Proposed algorithm

To handle with the problems of nonlinearity and redundancy, the original high-dimensional data is projected using KPCA transformation followed by the MLPP transformation, in order to find the k-nearest neighbors with higher matching precision in the final subspace. The framework of the proposed SPNE algorithm is shown in Fig. 1. SPNE outperforms the conventional NE-based algorithms, such as NeedFS [13], CSNE [14], and JLNE [15] with the following two advantages: (1) For each test

Experimental results

To validate the effectiveness of the proposed algorithm for single image super-resolution reconstruction, experiments with a magnification factor of 3 are conducted to compare with several existing SR approaches including bicubic interpolation, NeedFS [13], CSNE [14], and JLNE [15].

Conclusion

In this paper, we propose a learning-based single image super-resolution algorithm via subspace projection and neighbor embedding, which can achieve better performance in the neighbor search process and embedding weights estimation. Using KPCA and MLPP transformations, the global and local structures in the original high-dimensional data can be preserved. The k-nearest neighbors with higher matching precision are selected in the final low-dimensional subspace rather than the original

Acknowledgment

This work is supported in part by the National Natural Science Foundation of China (Grant no. 61373180), the Fundamental Research Funds for the Central Universities (Grant no. SWJTU09CX039, and no. SWJTU10CX09), the 2014 Doctoral Innovation Funds of Southwest Jiaotong University, and the Fundamental Research Funds for the Central Universities.

Xiaoyan Li received the B.S. degree in Communication Engineering from Southwest Jiaotong University, Chengdu, China, in 2009. Since September 2009, she has been directly working towards the Ph.D. degree in Communication and Information Systems at Southwest Jiaotong University. Her research interests include image super-resolution, signal and image processing, and computer vision.

References (31)

  • X. Ma et al.

    Kernel principal component analysis for stochastic input model generation

    J. Comput. Phys.

    (2011)
  • W.K. Wong et al.

    Supervised optimal locality preserving projection

    Pattern Recognit.

    (2012)
  • M. Irani et al.

    Motion analysis for image enhancement: resolution, occlusion and transparency

    J. Vis. Commun. Image Represent.

    (1993)
  • M.B. A. Haghighat et al.

    A non-reference image fusion metric based on mutual information of image features

    Comput. Electr. Eng.

    (2011)
  • S. Park et al.

    Super-resolution image reconstruction: a technical overview

    IEEE Signal Process. Mag.

    (2003)
  • Cited by (18)

    • Hypergraph-regularized sparse representation for single color image super resolution

      2021, Journal of Visual Communication and Image Representation
      Citation Excerpt :

      Reference [11–13] solve SR problem by using k-nearest neighbor learning method among a huge database at the cost of massive time. References [14,15] pay attention to reducing the computation, but fixed neighbor numbers may lead to over-fitting or under-fitting cases. Sparse representation based approaches [16–24] have been extensively utilized and overcomes the former two learning methods.

    • Deep learning-based image super-resolution considering quantitative and perceptual quality

      2020, Neurocomputing
      Citation Excerpt :

      Nevertheless, it is still highly challenging due to the lack of information in the low-resolution images, i.e., an ill-posed problem [1]. Until the mid-2010s, feature extraction-based methods have been proposed, including sparse coding [2], neighbor embedding [3], and Bayes forest [4]. After that, the emergence of deep learning for visual representation [5], which is triggered by an image classification challenge (i.e., ImageNet) [6], has also flowed into the field of super-resolution [7].

    • High-frequency details enhancing DenseNet for super-resolution

      2018, Neurocomputing
      Citation Excerpt :

      Recently, learning-based SR methods [11–14] achieve the state-of-the-art performance by learning mappings between pairs of LR and HR images patches in large external databases. Among them, algorithms based on Convolution Neural Network (CNN) [15–17,21–23] have shown excellent performance. Dong et al. [15] introduced CNN to the SISR task by using bicubic interpolation to up-sample the LR images first and then training one three layers convolutional network (SRCNN) via back-propagation to learn a mapping from the up-sampled LR input to the ground truth HR.

    • Interactive intuitionistic fuzzy methods for multilevel programming problems

      2017, Expert Systems with Applications
      Citation Excerpt :

      Multilevel programming problems (MLPPs), which are the crucial mathematical optimization problems for describing large decentralized decision problems with multiple interacting decision makers in a hierarchical organization, have been widely applied in industry (Nicholls, 1996), agriculture (Candler, Fortuny-Amat, & McCarl, 1981), transportation (Suh & Kim, 1992), government policy (Amouzegar & Moshirvaziri, 1999), finance (Bard, Plummer, & Sourie, 2000), warfare (Bracken, Falk, & Miercort, 1977), planning (Hobbs & Nelson, 1992), cloud computing resource allocation (Yeh, Yeh, & Lin, 2012), single image super-resolution (Li, He, Yin, Chen, & Cheng, 2014), label propagation (Zoidi, Tefas, Nikolaidis, & Pitas, 2014), municipal waste system (Saranwong & Likasiri, 2016) and supply chain management (Gao, Zhang, Lu, & Wee, 2011; Saranwong & Likasiri, 2016).

    • Modified sparse representation based image super-resolution reconstruction method

      2017, Neurocomputing
      Citation Excerpt :

      In the image reconstruction stage, for input LR image patches, those HR image patches highly similar to LR image patches are found in dictionaries and used as high components of LR image patches, and then the HR image patches can be obtained by adding these high components and LR image patches [28–30]. Common learning based methods include Markov random field (MRF) based ones [30,31], locally linear embedding (LLE) based ones [32,33], sparse representation based ones [34,35] and so on. Although MRF based methods can improve the edge effect of reconstructed images, they need millions of HR and LR image patches in order to get better reconstruction effect.

    View all citing articles on Scopus

    Xiaoyan Li received the B.S. degree in Communication Engineering from Southwest Jiaotong University, Chengdu, China, in 2009. Since September 2009, she has been directly working towards the Ph.D. degree in Communication and Information Systems at Southwest Jiaotong University. Her research interests include image super-resolution, signal and image processing, and computer vision.

    Hongjie He received the Ph.D. degree in Signal and Information Processing from Southwest Jiaotong University, Chengdu, China, in 2009. Currently, she is a professor of Southwest Jiaotong University. Her research interests are in the areas of digital forensics and image processing.

    Zhongke Yin received the Ph.D. degree in Signal and Information Processing from University of Electronic Science and Technology of China, in 1997. Currently, he is a professor of Southwest Jiaotong University. His research interests include signal and image sparse decomposition, and digital image processing.

    Fan Chen received the Ph.D. degree in Information Security from Southwest Jiaotong University, Chengdu, China, in 2012. Currently, he is an associate professor of Southwest Jiaotong University, Chengdu, China. His research interests include multimedia security, and digital watermarking.

    Jun Cheng received the B.Eng., B.Fin., and M.Eng. degrees from the University of Science and Technology of China, Hefei, China, in 1999 and 2002, respectively, and the Ph.D. degree from the Chinese University of Hong Kong, Hong Kong, in 2006. He is currently with the Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China, as a Professor and the Director of the Laboratory for Human Machine Control. His current research interests include computer visions, robotics, machine intelligences, and control.

    View full text