Elsevier

Digital Signal Processing

Volume 48, January 2016, Pages 298-309
Digital Signal Processing

Dictionary learning with the cosparse analysis model based on summation of blocked determinants as the sparseness measure

https://doi.org/10.1016/j.dsp.2015.09.011Get rights and content

Abstract

Dictionary learning is crucially important for sparse representation of signals. Most existing methods are based on the so called synthesis model, in which the dictionary is column redundant. This paper addresses the dictionary learning and sparse representation with the so-called analysis model. In this model, the analysis dictionary multiplying the signal can lead to a sparse outcome. Though it has been studied in the literature, there is still not an investigation in the context of dictionary learning for nonnegative signal representation, while the algorithms designed for general signal are found not sufficient when applied to the nonnegative signals. In this paper, for a more efficient dictionary learning, we propose a novel cost function that is termed as the summation of blocked determinants measure of sparseness (SBDMS). Based on this measure, a new analysis sparse model is derived, and an iterative sparseness maximization scheme is proposed to solve this model. In the scheme, the analysis sparse representation problem can be cast into row-to-row optimizations with respect to the analysis dictionary, and then the quadratic programming (QP) technique is used to optimize each row. Therefore, we present an algorithm for the dictionary learning and sparse representation for nonnegative signals. Numerical experiments on recovery of analysis dictionary show the effectiveness of the proposed method.

Introduction

Many natural signals are distributed in a high dimensional space. They can be modeled using low-dimensional structures and be represented using just a few parameters in an appropriate model [1]. Signal models are the cores of various signal and image processing tasks, such as compression, denoising, sampling, inpainting and so on [2]. Essentially, a signal model is a set of mathematical properties that the data is believed to satisfy [3]. A good model should be simple while matching the signals. In the past decade, the sparse and redundant representation model has been proved to be important and useful [4], [5]. Most existing methods for these are based on a so called synthesis model for describing signals. Here we describe this model briefly.

Assume that we are to model the signal Xm×N. The synthesis model suggests that the signal could be described asX=DH, orXDH,s.t.XDHFε, where Dm×n is a possible dictionary [6]. Since (mn), where it contains n atoms of size m×1, the dictionary is column overcomplete, i.e., redundant. Here Hn×N is the representation coefficient matrix. If the purpose is for a sparse representation, it is assumed that the representation H is sparse. That is, most elements are 0-valued, which means each signal element can be represented as a linear combination of few columns of the dictionary D. The name “synthesis” comes from the relation X=DH, with the interpretation that the model describes a way to synthesize signals X [3].

There have been many publications that are based on the synthesis model; a long series of applications in signal and image processing with this model demonstrate state-of-the-art results [8]. It can be safely said that the synthesis model is a mature and stable field [3]. And this model has been heavily investigated in the last decade and many remarkable results achieved in almost all signal processing applications, the examples in [8], [9]. However, in this model, the error XDHF as the cost function is usually non-convex so that the optimization problem with the mode is hard to solve mathematically, and the deduced algorithm is complicated in computation.

Interestingly, the synthesis model has a “twin” model that takes an analysis point of view [3]. Assume that there is Ωn×m so that Ω can be used to find sparse H so that H=ΩX. This analysis model can be converted as a minimization problem of the error function HΩXF. Remarkable, this error function is convex, contrasting the nonconvexity of the synthesis model, and the optimization will become easier so that it can be solved by some simpler optimization methods. To emphasize the difference with that in the synthesis model, the term “co-sparsity” has been introduced in literature [3], [7], which simply counts the number of zero elements of ΩX [10]. The analysis model relies on a matrix Ωn×m, which we refer to as the analysis dictionary. Here “analysis” means the dictionary analyzes the signal to produce a sparser result [11]. Different from the synthesis model, the rows of analysis dictionary represent analysis atoms.

Now we look carefully at the analysis model mentioned above. It seems to be similar to the synthesis model. First, let Ωn×m be a signal transformation or an analysis operator. Its rows are the row vectors which will be applied to the signal. To capture various aspects of the information in the signal by shorten signal, we typically have mn. Since there is a relationship between the synthesis sparse dictionary and the analysis sparse dictionary by the equation D=Ω. Then we can use the result obtained from the analysis model for obtaining a result for the synthesis model, in some special cases (m=n or Ω existing). The analysis model plays important roles in applications, such as low-dimensional signal model [41], image denoising or inpainting, bimodal super-resolution and image registration [2], [21], [37], [38].

In the past decade, a great deal of effort has been dedicated to learn the dictionary for the synthesis model; however, the analysis dictionary learning problem has received less attention with only a few algorithms proposed recently [2], [11], [12], [13], [14], [15]. Exploiting the fact that a row of the analysis dictionary is orthogonal to a subset of signals X, a sequential minimal eigenvalue based analysis dictionary learning algorithm is proposed in [12]. In [13], [16], [17], [36], an 1 norm penalty function is imposed to the sparse coefficient and a projected subgradient-based algorithm is proposed for the analysis dictionary learning. The authors of [36] formulate the learning task as a bilevel-programming optimization problem, which in turn is handled using the gradient descent optimization. These works employ a uniformly normalized tight frame as a constraint on the dictionary to avoid the trivial solution. In [2], the Analysis K-SVD (AK-SVD) algorithm is proposed for analysis dictionary learning. By keeping Ω fixed, the optimal backward greedy algorithm (OBG) is employed to estimate a submatrix of Ω whose rows are then used to determine a submatrix and the eigenvector associated with the smallest eigenvalue of this submatrix is then used to update Ω. Greedy Analysis Pursuit is proposed in [18]. This paper presented preliminary and simplified versions of analysis IHT (AIHT), analysis HTP (AHTP), analysis CoSaMP (ACoSaMP) and Analysis SP (ASP) in [19], [20] as analysis versions of the synthesis counterpart methods. The article [21] proposed analysis thresholding algorithm for analysis sparse representation, the analysis reveals significant properties of the dictionary, which govern the pursuit performance. The first is the degree of linear dependencies between sets of rows in the dictionary, depicted by the co-sparsity level. The second property, termed the Restricted Orthogonal Projection Property, is the level of independence between such dependent sets and other rows in Ω. The authors of [39] take a very different route towards the dictionary learning problem, employing a constraint combining the unit row norm and the full rank constraint, with an additional consideration that the analysis operator has no trivially linear dependent rows. As these approaches have shown promising empirical results, we expect that this will open the door to explore the analysis sparse representation more broadly.

As mentioned above, studies on analysis model have led to several dictionary learning algorithms, such as the analysis K-SVD [2], Greedy Analysis Pursuit (GAP) [18], the analysis thresholding algorithm [21]. However, there are lesser investigations in the context of analysis dictionary learning for nonnegative signal representation. In practice some particular signals such as spectral data have the limitations of nonnegativity, including the factorization factors obtained by some specific data analysis [22], [23]. The purpose of this paper is just to address the problem on nonnegative analysis sparse representation.

In this paper, we presented a novel algorithm of analysis dictionary learning for nonnegative signal representation, which is parallel to the synthesis model in its rationale and structure. The algorithm utilizes the summation of blocked determinants of sparseness measure as the sparse constraints. The optimization of the error function can be cast into a series of QP problems, which can be easily solved. Finally, an efficient QP algorithm was developed to perform nonnegative analysis sparse representation. We will describe details of the proposed method in the following sections.

Notations used in this paper are as follows. A bold uppercase letter, such as H, denotes a matrix and a bold lowercase letter h denotes a vector. HT is the transpose of the matrix H, H is the pseudo-inverse of matrix H and H1 denotes the inverse of matrix H. det(H) denotes the determinant value of matrix H. h is the column vector in H. The notation ()ij, has been used to specify the element locating in the i-th row and j-th column of the operand matrix. Also, we define the 0 norm of a vector as h0=i|hi|0, and the 1 norm of a matrix as H1=ij|Hij|1.

The remainder of this paper is organized as follows. Section 2, we firstly introduce the analysis model with the sparsity constraints. In Section 3, we describe the problem formulation of the analysis sparse model. The detailed algorithm is described in Section 4. Then we describe the computation complexity of our algorithm in Section 5. In Section 6 we present experimental results for the proposed algorithm and compare these results. We use synthesis data for numerical experiments. Finally, conclusions are drawn in Section 7.

Section snippets

The analysis sparse model

The analysis model for a signal xm, which is a column in the signals X, uses the possibly redundant analysis dictionary Ωn×m, where redundant here implies (m<n) [21]. Throughout this paper the i-th row in Ω will be denoted by Ωi. Assume that the analysis representation vector h=Ωx is sparse. This strategy is to obtain h as the sparsest possible representation based on the sparsity measure M(h),h=ArgminhM(h)s.t.h=Ωx. Of specific interest is the 0 norm case, where M(h)=h0 counts the

Formulation

Generally, the analysis sparse problem for nonnegative signals is formulated as follows: given a signal data matrix X, find a matrice Ω so thatΩX=H, becomes as sparse as possible, where the analysis dictionary Ωn×m (nm) is row redundant and H+n×N is a nonnegative matrix. The problem can also described as to find the most sparsity matrix H subject to the equation above. In this paper, we adopt the determinant measure as the sparsity constraint. Therefore, based on the determinant measure,

Algorithm

In this section, we describe the details of the proposed algorithms.

Computational complexity

As far as the computational complexity of the algorithm is concerned, by using a reflective Newton method in [40], each QP problem in (12) can be solved with an approximate computational complexity of O(mN) in the best case and O(m2N) in the worst case. Practically, by using the subroutine improve method, the complexity of the QP problem is approximately O(kmN), where 1k<m is often small [40]. Therefore, the computational complexity of the proposed QP is approximately O(kmnN) in each iteration.

Numerical experiments

In this section, we present the results of the experiments for evaluating the proposed algorithms. From the experiments we tested whether this algorithm can recover the grand true nonnegative analysis dictionary. In the experiments, all problems were coded in Matlab, and were run in the environment of Matlab 7.8 (R2009a) on a computer with a 3.0 GHz Inter Core i5 CPU and 12 G of memory, under Microsoft Windows 7 operating system.

Conclusions

In this paper, we presented a novel algorithm for learning the analysis dictionary, which is parallel to the synthesis model in its rationale and structure. The algorithm utilizes the summation of blocked determinant-type of sparseness measure as the sparse constraint. Results of numerical experiments demonstrated the ability of our method to correctly and effectively learn the nonnegative analysis dictionary. Furthermore, we can find that the summation of blocked determinants sparseness

Yujie Li received the B.Sc. degree in physics from School of Physical Science and Electronics, Henan University, Kaifeng, China, in 2010 and the M.Sc. degree in plasma physics from School of Physics and Optoelectronic, Dalian University of Technology, Dalian, China, in 2013. She is currently a Ph.D. student in Graduate Department of computer and Information System, the University of Aizu, Japan. Her current research interests include Signal Processing, Sparse representation, Blind source

References (41)

  • S. Nam et al.

    Cosparse analysis modeling—uniqueness and algorithms

  • M. Elad

    Sparse and Redundant Representations from Theory to Applications in Signal and Image Processing

    (2010)
  • J. Starck et al.

    Sparse Image and Signal Processing—Wavelets, Curvelets, Morphological Diversity

    (2010)
  • S. Hawe et al.

    Analysis operator learning and its application to image reconstruction

    IEEE Trans. Image Process.

    (2013)
  • S. Ravishankar et al.

    Learning sparsifying transforms

    IEEE Trans. Signal Process.

    (2013)
  • B. Ophir et al.

    Sequential minimal eigenvalues: an approach to analysis dictionary learning

  • M. Yaghoobi et al.

    Constrained overcomplete analysis operator learning for cosparse signal modelling

    IEEE Trans. Signal Process.

    (2013)
  • M. Yaghoobi et al.

    Relaxed analysis operator learning

  • M. Yaghoobi et al.

    Noise aware analysis operator learning for approximately cosparse signals

  • R. Giryesa et al.

    Greedy-like algorithms for the cosparse analysis model

    Linear Alg. Appl.

    (2014)
  • Cited by (12)

    View all citing articles on Scopus

    Yujie Li received the B.Sc. degree in physics from School of Physical Science and Electronics, Henan University, Kaifeng, China, in 2010 and the M.Sc. degree in plasma physics from School of Physics and Optoelectronic, Dalian University of Technology, Dalian, China, in 2013. She is currently a Ph.D. student in Graduate Department of computer and Information System, the University of Aizu, Japan. Her current research interests include Signal Processing, Sparse representation, Blind source separation and Compressed sensing.

    Shuxue Ding received the M.Sc. degree in physics from Dalian University of Technology, P. R. China, in 1988, and the Ph.D. degree in physics from Tokyo Institute of Technology, Japan, in 1996. From 1989 to 1991 and from 1991 to 1992, respectively, he was an Assistant Professor and Associate Professor in the School of Physics and Optoelectronic Energy at Dalian University of Technology. From 1996 to 1998, he was with Fujisoft-ABC Inc., Japan, where he was involved in algorithm design for telecommunication systems. From 1998 to 2003, he was with Clarion Co., Ltd, Japan, where he engaged in researches in communication and signal processing, especially in speech recognition. From 2003 to 2005, he was a visiting faculty at the University of Aizu, Japan. From 2005 to 2010, he was an Associate Professor in the School of Computer Science and Engineering at the University of Aizu. He is currently a Professor in the School of Computer Science and Engineering, the University of Aizu. He is a member of IEEE, ACM, and IEICE. Dr. Ding has been engaged in research in a wide range of areas of mathematical and physical engineering, such as statistical signal processing, optimization, neural computation, bioelectromagnetism, and information sciences. In particular, he has devoted himself to blind source separation and independent component analysis, and their applications in acoustic signals and vital signs. Recently, he is also conducting research in brain-style information processing, pattern recognition, compressive sensing and sparse representation. He is also interested in speech and image processing, quantum computation and optimization, quantum information, and other physical theories of information.

    Zhenni Li received the bachelor's degree in 2009 from School of Physical Science and Electronics, Shanxi Datong University, China. She received the master's degree in 2012 from School of Physics and Optoelectronic, Dalian University of Technology, China. Now she is a Ph.D. student in School of Computer Science and Engineering, University of Aizu, Japan. Her research interests include Machine Learning, Signal Processing, Sparse Representation and Quantum computing.

    View full text