Next Article in Journal / Special Issue
Fourth Order Diffusion Equations with Increasing Entropy
Previous Article in Journal
Case Studies and Benchmark Examples for the Use of Grading Entropy in Geotechnics
Previous Article in Special Issue
Deterministic Thermal Reservoirs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Minimum Mutual Information and Non-Gaussianity Through the Maximum Entropy Method: Theory and Properties

by
Carlos A. L. Pires
* and
Rui A. P. Perdigão
Instituto Dom Luiz, Faculdade de Ciências, University of Lisbon, DEGGE, Ed. C8, Campo-Grande, 1749-016 Lisbon, Portugal
*
Author to whom correspondence should be addressed.
Entropy 2012, 14(6), 1103-1126; https://doi.org/10.3390/e14061103
Submission received: 20 May 2012 / Revised: 8 June 2012 / Accepted: 15 June 2012 / Published: 19 June 2012
(This article belongs to the Special Issue Concepts of Entropy and Their Applications)

Abstract

:
The application of the Maximum Entropy (ME) principle leads to a minimum of the Mutual Information (MI), I(X,Y), between random variables X,Y, which is compatible with prescribed joint expectations and given ME marginal distributions. A sequence of sets of joint constraints leads to a hierarchy of lower MI bounds increasingly approaching the true MI. In particular, using standard bivariate Gaussian marginal distributions, it allows for the MI decomposition into two positive terms: the Gaussian MI (Ig), depending upon the Gaussian correlation or the correlation between ‘Gaussianized variables’, and a non‑Gaussian MI (Ing), coinciding with joint negentropy and depending upon nonlinear correlations. Joint moments of a prescribed total order p are bounded within a compact set defined by Schwarz-like inequalities, where Ing grows from zero at the ‘Gaussian manifold’ where moments are those of Gaussian distributions, towards infinity at the set’s boundary where a deterministic relationship holds. Sources of joint non-Gaussianity have been systematized by estimating Ing between the input and output from a nonlinear synthetic channel contaminated by multiplicative and non-Gaussian additive noises for a full range of signal-to-noise ratio (snr) variances. We have studied the effect of varying snr on Ig and Ing under several signal/noise scenarios.

1. Introduction

One of the most commonly used information theoretic measures is the mutual information (MI) [1], measuring the total amount of probabilistic dependence among random variables (RVs)—see [2] for a unifying perspective and axiomatic review. MI is a positive quantity vanishing iff RVs are independent.
MI is an effective tool for multiple purposes, namely and among others: (a) Blind signal separation [3] and Independent Component Analysis (ICA) [4], both of which look for transformed and/or lagged stochastic time series [5] which minimize MI; (b) Predictability studies, Predictable Component Analysis [6] and Forecast Utility [7], all of which are focused on the analysis and decomposition of the MI between probabilistic forecasts and observed states.
Analytical expressions of MI are known for a few number of parametric joint distributions [8,9]. Alternatively, it can be numerically estimated by different methods, such as Maximum Likelihood estimators, Edgeworth expansion, Bayesian methods, equiprobable and equidistant histograms, kernel‑based probability distribution functions (PDFs), K-nearest neighbors technique—see [10,11] and references therein for a survey of estimation methods and scoring comparison studies.
In the bivariate case, treated here, the MI I ( X , Y ) , between RVs X , Y is the Kullback–Leibler (KL) divergence: I ( X , Y ) D ( p X Y | | p X p Y ) = E p X Y ( log ( p X Y / ( p X p Y ) ) ) 0 , between the joint probability function p X Y and the product of marginal probability distributions p X p Y where E p X Y is the expectation operator over the measure p X Y . The MI is invariant for smooth invertible transformations of X , Y .
The goal is the determination of theoretical lower MI bounds under certain conditions or, in other words, the minimum mutual information (MinMI) [12] between two RVs X , Y , consistent, both with imposed marginal distributions and cross-expectations assessing their linear and nonlinear covariability. Those lower bounds can be obtained due to the application of the Maximum Entropy (ME) method to distributions [13] and to the inequality: D ( p M E | | q ) D ( p | | q ) . Here, p M E is a Maximum Entropy probability distribution (MEPD) with respect to q, say p M E = arg min p Ω ( D ( p | | q ) ) , where Ω is a PDF class verifying a given set of constraints [1]. Therefore, by using the joint MEPD p = p X Y M E and q = p X p Y , the lower MI bound is obtained: D ( p X Y M E | | p X p Y ) D ( p X Y Ω | | p X p Y ) I ( X , Y ) . Finding of the bivariate p X Y M E is straightforward if the marginal distributions are themselves well defined ME distributions. We solve that by transforming the single variables X , Y into others with imposed ME probability mass distributions through the so called ME-anamorphoses [14].
The joint ME probability distribution p X Y M E is derived from the minimum of a functional in terms of a certain number of Lagrange parameters. The properties of multivariate ME distributions have been studied for various ME constraints, namely: (a) imposed marginals and covariance matrix [15]; (b) generic joint moments [16]. Abramov [17,18,19] has developed efficient and stable numerical iterative algorithms for computing ME distributions forced by sets of polynomial expectations. Here we use a bivariate version of the algorithm of [20], already tested in [21].
By taking a sequence of encapsulated sets of joint ME constraints we obtain an increasing hierarchy of lower MI bounds converging towards the total MI.
We particularize this methodology to the case where X , Y are standard Gaussians, issued from single homeomorphisms of an original pair of variables X ^ , Y ^ by Gaussian anamorphosis [14]. Then we get the MI I ( X , Y ) = I ( X ^ , Y ^ ) , which is decomposed into two generic positive quantities, a Gaussian MI Ig and a non-Gaussian MI Ing [21], vanishing under bivariate Gaussianty. The Gaussian MI is given by I g ( c g ) 1 / 2 log ( 1 c g 2 ) , where c g ( X ^ , Y ^ ) c o r ( X , Y ) is the Pearson linear correlation between the Gaussianized variables. As for the non-Gaussian MI term Ing it relies upon imposed nonlinear correlations between the single Gaussian variables. The MI reduces to I g ( c g ) when only moments of order one and two are imposed as ME constraints. We will note that, for certain extreme non-Gaussian marginals, I g ( c = c o r ( X ^ , Y ^ ) ) > I ( X ^ , Y ^ ) , thus showing that I g ( c ) [22] is not a proper MI lower bound in general.
The correlation c g , hereafter called Gaussian correlation [21] is a nonlinear concordance measure like the Spearman rank correlation [23] and the Kendall τ, which, by definition are invariant for monotonically growing smooth marginal transformations. These measures have the good property of being expressed as functionals of the copula density functions [24], uniquely dependent on the cross dependency between variables.
The non-Gaussian MI ng holds some interesting characteristics. It coincides with the joint negentropy (deficit of entropy with respect to that of the Gaussian PDF with the same mean, variance and covariance) in the space of ‘Gaussianized’ variables, which is invariant for any orthogonal or oblique rotation of them. In particular for uncorrelated rotated variables, it coincides with the ‘compactness’, which measures the concentration of the joint distribution around a lower-dimensional manifold [25] which is given by D ( p X Y | | p S G ) , i.e., the KL divergence with respect to the spherical Gaussian p S G (Gaussian with an isotropic covariance matrix with the same trace as that of p X Y , say the total variance).
We also show that Ing comprises a series of positive terms associated to a p-sequence of imposed monomial expectations of total even order p (2,4,6,8…). The higher the number of independent constraints, the higher the order of terms that are retained in that series and the more information is extracted from the joint PDF.
We have shown that the possible values of the cross-expectations lie within a bounded set obtained by Schwarz-like inequalities. We illustrate the range of Ing values within those sets as function of third and fourth-order cross moments. Near the set’s boundary, Ing tends to infinity where a deterministic relationship holds and the ME problem functional is ill-conditioned.
In order to better understand the possible sources of joint non-Gaussianity and non-Gaussian MI, we have used the preceding method for computing Ig and Ing between the input and the output of a nonlinear channel contaminated by multiplicative and non-Gaussian noise for a full range of the signal-to-noise (snr) variance ratio. We put in evidence that sources of non-Gaussian MI arise from the nonlinearity of the transfer function, multiplicative noise and additive non-Gaussian noise [26].
Many of the results of the paper are straightforwardly generalized to the multivariate case with three or more random variables.
The paper is then organized as follows: Section 2 formalizes the Minimum Mutual Information (MinMI) principle from maximum entropy distributions, while Section 3 particularizes that principle to the MI decomposition into Gaussian and non-Gaussian MI parts. Section 4 addresses the non‑Gaussianity in a nonlinear non-Gaussian channel. The paper ends with conclusions and appendices with theoretical proofs and the numerical algorithm for solving the ME problem. This paper is followed by a companion one [27] on the estimation of non-Gaussian MI from finite samples with practical applications.

2. MI Estimation from Maximum Entropy PDFs

In this section we present the basis of the MI estimation between bivariate RVs, through the use of joint PDFs inferred by the maximum entropy (ME) method (ME-PDFs for short) in the space of transformed (anamorphed) marginals into specified ME distributions. We start with preliminary general concepts and definitions.

2.1. General Properties of Bivariate Mutual Information

Let ( X , Y ) be continuous RVs with support given by the Cartesian product S = S X S Y 2 , and let the joint PDF ρ X , Y be taken absolutely continuous with respect to the product of the marginal PDFs ρ X and ρ Y . The mutual information (in nat) between X and Y is a non-negative real functional of ρ X , Y expressed in different equivalent forms as:
I ( X , Y ) S ρ X , Y ( x , y ) log ( ρ X , Y ( x , y ) ρ X ( x ) ρ Y ( y ) ) d x d y = D ( ρ X Y | | ρ X ρ Y ) = H ρ X + H ρ Y H ρ X Y 0 ,
in terms both of a KL divergence and Shannon entropies. The MI equals zero iff the RVs X and Y are statistically independent, i.e., iff ρ X , Y ( x , y ) = ρ X ( x ) ρ Y ( y ) , x S X , y S Y , except possibly in a zero‑measure set in S . Under quite general regularity conditions of PDFs, statistical independence of X and Y is equivalent to the vanishing of the Pearson correlation between any pair of linear and/or nonlinear mapping functions X ˜ ( X ) and Y ˜ ( Y ) . Then, linearly uncorrelated non-independent variables have at least a non-zero nonlinear correlation. The MI between transformed variables X ˜ ( X ) and Y ˜ ( Y ) , differentiable almost everywhere, satisfies the ‘data processing inequality’ I ( X ˜ ( X ) , Y ˜ ( Y ) ) I ( X , Y ) [1] with equality occurring if both X ˜ and Y ˜ are smooth homeomorphisms of X and Y respectively.

2.2. Congruency between Information Moment Sets

  • Definition 1: Following the notation of [28], we define the moment class Ω T , θ of bivariate PDFs ρ X Y of ( X , Y ) , as:
    Ω T , θ = { ρ X Y : E ρ [ T ] = θ }
    where E ρ is the ρ-expectation operator, T = ( T 1 , ... , T J ) T is a J-dimensional vector composed of J absolutely ( X , Y ) integrable functions with respect to ρXY, and θ = ( θ 1 , ... , θ J ) T is the vector of function expectations of T . Here and henceforth, ( T , θ ) is denoted as an information moment set.
The PDF ρ Ω T , θ that maximizes the Shannon entropy or the ME-PDF verifying the constraints associated to ( T , θ ) , hereby represented by ρ T , θ * , must exist since H ρ is a concave function in the non-empty convex set Ω T , θ . The form and estimation of the ME-PDF is described in Appendix 1. The ME-PDF satisfies the following Lemma [1]:
  • Lemma 1: Given two information encapsulated information moment sets: ( T , θ ) ( T 1 , θ 1 ) i.e., with T 1 including more constraining functions than T , the respective ME-PDFs ρ Τ , θ * , ρ T 1 , θ 1 * , if they exist, satisfy the following conditions:
    E ρ [ log ρ T , θ * ] = E ρ T 1 , θ 1 * [ log ρ T , θ * ] = E ρ T , θ * [ log ρ T , θ * ] = H ρ T , θ *
    E ρ [ log ρ T 1 , θ 1 * ] = E ρ T 1 , θ 1 * [ log ρ T 1 , θ 1 * ] = H ρ T 1 , θ 1 *
This means that any ME-PDF log-mean is unchanged when it is computed with respect to a more constraining ME-PDF. As a corollary we have D ( ρ T 1 , θ 1 * | | ρ Τ , θ * ) = H ρ T , θ * H ρ T 1 , θ 1 * 0 , i.e., the ME decreases with the increasing number of constraints. This can be understood bearing in mind that the entropy maximization is performed in a more reduced PDF class, since Ω T 1 , θ 1 Ω T , θ .
  • Definition 2: If two information moment sets ( T 1 , θ 1 ) , ( T 2 , θ 2 ) are related by linear affine relationships, then the sets are referred to as “congruent”, a property hereby denoted as ( T 1 , θ 1 ) P D F ( T 2 , θ 2 ) , and consequently both PDF sets are equal, i.e., Ω T 1 , θ 1 = Ω T 2 , θ 2 [15]. A stronger condition than ‘congruency’ is the ME-congruency, denoted as ( T 1 , θ 1 ) M E ( T 2 , θ 2 ) , holding when both the associated ME-PDFs are equal. For example, both the univariate constraint sets ( T 1 = X 2 , θ 1 = 1 ) and ( T 2 = ( X 2 , X 4 ) T , θ 2 = ( 1 , 3 ) T ) for X lead to the same ME-PDF, the standard Gaussian N(0,1). Consequently both information moment sets are ME-congruent but not congruent since Ω T 2 , θ 2 Ω T 1 , θ 1 . This is because the Lagrange multiplier of the ME functional (see Appendix 1) corresponding to the fourth moment is set to zero without any constraining effect. The congruency implies ME-congruency but not the converse.

2.3. MI Estimation from Maximum Entropy Anamorphoses

We are looking for a method of obtaining lower bound MI estimates from ME-PDFs. For that purpose we will decompose the information moment set as: ( T , θ ) = ( T i n d , θ i n d ) ( T c r , θ c r ) , say into a marginal or independent part ( T i n d , θ i n d ) = ( T X , θ X ) ( T Y , θ Y ) of single X and Y independent moments on S X , S Y , and a cross part ( T c r , θ c r ) on S X S Y , made by moments of joint ( X , Y ) functions not expressible as sums of single moments. For example, by taking ( T X , θ X ) = ( ( X , X 2 ) , ( 0 , 1 ) ) , ( T Y , θ Y ) = ( ( Y , Y 2 ) , ( 0 , 1 ) ) and ( T c r , θ c r ) = ( X Y , c ) for ( X , Y ) 2 leads to a ME-PDF which is the bivariate Gaussian with correlation c and standard Gaussians as marginal distributions.
The joint ME-PDF associated to the independent part ( T i n d , θ i n d ) is the product of two independent ME-PDFs related to ( T X , θ X ) and ( T Y , θ Y ) , with the joint entropy being the sum of marginal maximum entropies, as in the case of independent random variables [15]. The KL divergence between the ME-PDFs associated to ( T , θ ) and those associated to ( T i n d , θ i n d ) is a proper MI lower bound or, expressed in other terms, a constrained mutual information. We denote it as:
I ( X , Y : ( T , θ ) , ( T i n d , θ i n d ) ) D ( ρ T , θ * | | ρ Τ i n d , θ i n d * ) = H ρ T i n d , θ i n d * H ρ T , θ * = H ρ T X , θ X * + H ρ T Y , θ Y * H ρ T , θ * 0
Its difference with respect to I ( X , Y ) is given by:
I ( X , Y ) I ( X , Y : ( T , θ ) , ( T i n d , θ i n d ) ) = D ( ρ X Y | | ρ T , θ * ) [ D ( ρ X | | ρ Τ X , θ X * ) + D ( ρ Y | | ρ Τ Y , θ Y * ) ]
which can be negative. However, the positiveness is ensured when marginal distributions are set to the ME-PDFs constrained by ρ X ( x ) = ρ Τ X , θ X * ( x ) , x and ρ Y ( y ) = ρ Τ Y , θ Y * ( y ) , y , which results in KL divergences vanishing with respect to the marginal PDFs in Equation (5). This procedure is quite general because ( X , Y ) can appropriately be obtained through an injective smooth maps ( X = X ( X ^ ) , Y = Y ( Y ^ ) ) from an original pair of RVs ( X ^ , Y ^ ) preserving the MI, i.e., I ( X ^ , Y ^ ) = I ( X , Y ) . Those maps are monotonically growing homeomorphisms, the hereby called ME-anamorphoses, which are obtained by equaling mass probability functions of the original variable ρ X ^ to those of the transformed variable ρ X (equally for ρ Y ^ and ρ Y ) as:
X ( X ^ ) ρ Τ X , θ X * ( u ) d u = X ^ ρ X ^ ( u ) d u ; Y ( Y ^ ) ρ Τ Y , θ Y * ( u ) d u = Y ^ ρ Y ^ ( u ) d u
The moments in ( T i n d , θ i n d ) are invariant under the transformation, ( X ^ , Y ^ ) ( X , Y ) , i.e., they are the same for both original and transformed variables: E ρ X ^ [ T X ] = E ρ X [ T X ] = θ X (idem for Y ). Therefore, in the space of ME-anamorphed variables the ME-based MI (4) gives the minimum MI (MinMI), compatible with the cross moments ( T c r , θ c r ) .
Thanks to Lemma 1, a hierarchy of ME-based MI bounds is obtainable by considering successive supersets of the ME constraints on the ME-anamorphed variables, which is justified by the theorem below.
  • Theorem 1: Let ( X , Y ) be a pair of single random variables (RVs), distributed as the ME-PDF associated to the independent constraints ( T i n d = ( T X , T Y ) , θ i n d = ( θ X , θ Y ) ) . Both variables can be obtained from previous ME-anamophosis. Let ( T 1 , θ 1 ) = ( T c r 1 T i n d 1 , θ c r 1 θ i n d 1 ) be a subset of ( T 2 , θ 2 ) = ( T c r 2 T i n d 2 , θ c r 2 θ i n d 2 ) , i.e., T c r 1 T c r 2 and T i n d T i n d 1 T i n d 2 , such that all independent moment sets are ME-congruent (see Definition 2), i.e., ( T i n d , θ i n d ) M E ( T i n d 1 , θ i n d 1 ) M E ( T i n d 2 , θ i n d 2 ) , i.e., such that the independent extra moments in ( T 2 , θ 2 ) are not further constraining the ME-PDF. Each marginal moment set is decomposed as ( T i n d j = ( T X j , T Y j ) , θ i n d j = ( θ X j , θ Y j ) ) , ( j = 1 , 2 ) . For simplicity of notation, let us denote I ( X , Y ) j I ( X ; Y : ( T j , θ j ) , ( T i n d j , θ i n d j ) ) ( j = 1 , 2 ) . Then, the following inequalities between constrained mutual informations hold:
    I ( X , Y ) j = I ( X , Y ) D ( ρ X Y | | ρ T j , θ j * ) I ( X , Y ) ( j = 1 , 2 )
    I ( X , Y ) 2 I ( X , Y ) 1 = D ( ρ T 2 , θ 2 * | | ρ T 1 , θ 1 * ) = H ρ T 1 , θ 1 * H ρ T 2 , θ 2 * 0
The proof is given in Appendix 2. The Theorem 1 justifies the possibility of building a monotonically growing sequence of lower bounds of I ( X , Y ) from encapsulated sequences of cross-constraints T c r 1 ... T c r j T c r j + 1 ... and independent constraints T i n d T i n d 1 ... T i n d j T i n d j + 1 ... . In the sequence, the entropy associated to independent constraint sets is always constant due to the ME-congruency, while the entropy of the joint ME-PDF decreases, thus allowing the MinMIs to grow. Therefore, I ( X , Y ) j is the part of MI due to cross moments in T c r j and the positive difference I ( X , Y ) j + 1 I ( X , Y ) j is the increment of MI due to the additional cross moments in T i n d j + 1 / T i n d j , while marginals are kept as preset ME-PDFs (e.g., Gaussian, Gamma, Weibull).

3. MI Decomposition under Gaussian Marginals

3.1. Gaussian Anamorphosis and Gaussian Correlation

In this section we explain how to implement the sequence of MI estimators detailed in Section 2.3 for the particular case where X and Y are standard Gaussian RVs. Our aim is to estimate the MI between two original variables ( X ^ , Y ^ ) of null mean and unit variance with real support. Those variables are then transformed through a homeomorphism, the Gaussian anamorphosis [14], into standard Gaussian RVs, respectively X ~ N ( 0 , 1 ) and Y ~ N ( 0 , 1 ) , given by:
X ( x ^ ) = Φ 1 ( x ^ ρ X ^ ( u ) d u ) = G X ^ ( x ^ ) ; Y ( y ^ ) = Φ 1 ( y ^ ρ Y ^ ( u ) d u ) = G Y ^ ( y ^ )
where Φ is the mass distribution function for the standard Gaussian. If ( X ^ , Y ^ ) are marginally non‑Gaussian, then the Gaussian anamorphoses are nonlinear transformations. In practice, Gaussian anamorphoses can be approximated empirically from finite data sets by equaling cumulated histograms. However, for certain cases, it is analytically possible to construct bivariate distributions with specific marginal distributions and the knowledge of the joint cumulative distribution function [29].
In the case of Gaussian anamorphosis, the information moment set ( T i n d , θ i n d ) of Theorem 1 includes the first and second independent moments of each variable: E [ X ] = E [ Y ] = 0 ; E [ X 2 ] = E [ Y 2 ] = 1 . Then, following the proposed procedure of Section 2.3, we will consider a sequence of cross-constraint sets for determining the hierarchy of lower MI bounds.
The most obvious cross moment to be considered is the X Y expectation, equal to the Gaussian correlation c g c o r ( G X ^ , G Y ^ ) = c o r ( X , Y ) between the ‘Gaussianized’ variables ( X , Y ) . The difference between c g and the linear correlation c = cor ( X ^ , Y ^ ) is easily expressed as:
c g c = E [ ( G X ^ X ^ ) ( G Y ^ Y ^ ) ]
The signal of the factor G X ^ X ^ in Equation (9) roughly depends on the skewness s k ( X ^ ) = E [ X ^ 3 ] and excess of kurtosis k u r ( X ^ ) = E [ X ^ 4 ] 3 through the rule of thumb, stating that sgn ( G X ^ X ^ ) is approximated by sgn ( s k ( X ^ ) ) and sgn ( k u r ( X ^ ) ) sgn ( X ^ ) , respectively for a skewed X ^ PDF and a symmetric X ^ PDF (idem for Y ^ ). Therefore, c g can result in an enhancement of correlation c or in the opposite effect, as shown in [21] for the RV pair of meteorological variables ( X ^ = North Atlantic Oscillation Index, Y ^ = monthly precipitation). The Gaussian correlation is a concordance measure like the rank correlation and Kendall τ, being thus invariant for a monotonically growing smooth homeomorphism of both X ^ and Y ^ . Those measures are expressed as functionals of the bivariate copula-function c [ u = X 1 ρ X 1 ( x ) d x , v = X 2 ρ X 2 ( y ) d y ] = ρ X 1 X 2 ( X 1 , X 2 ) / ( ρ X 1 ( X 1 ) ρ X 2 ( X 2 ) ) , which is uniquely dependent on the cumulated marginal probabilities and equal to the density ratio, independently from the specific forms of marginal PDFs [24]. In particular, the Gaussian correlation is given by:
c g = 0 1 0 1 c [ u , v ] Φ 1 ( u ) Φ 1 ( v ) d u d v

3.2. Gaussian and Non-Gaussian MI

The purpose of this sub-section is to express which part of MI comes from joint non-Gaussianity. If the ‘Gaussianized’ variables ( X , Y ) are jointly non-Gaussian, then the original standardized variables ( X ^ , Y ^ ) , obtained from ( X , Y ) by invertible smooth monotonic transformations, are jointly non‑Gaussian as well. However, the converse is not true. In fact, any nonlinear transformation ( X ^ , Y ^ ) of jointly Gaussian RVs ( X , Y ) leads to non-Gaussian ( X ^ , Y ^ ) with the joint non-Gaussianity arising in a trivial way. Therefore, the ‘genuine’ joint non-Gaussianity can only be diagnosed in the space of Gaussianized variables ( X , Y ) . When the marginal PDFs of ( X ^ , Y ^ ) are non-Gaussian then c g c in general. In particular, if the correlation c is null as it occurs when ( X ^ , Y ^ ) are principal components (PCs), then c g can be non-null, leading to statistically dependent PCs since they are non-linearly correlated.
The MI between Gaussianized variables ( X , Y ) or ( X ^ , Y ^ ) is expressed as I ( X ^ , Y ^ ) = I ( X , Y ) = 2 H g H ρ X Y , where H g 1 / 2 log ( 2 π e ) is the Shannon entropy of the standard Gaussians X , Y and H ρ X Y is the ( X , Y ) joint entropy, associated to the joint PDF ρ X Y . Given the above equality, the MI is decomposed into two positive terms: I ( X ^ , Y ^ ) = I ( X ^ , Y ^ ) g + I ( X ^ , Y ^ ) n g . The first term is the Gaussian MI [21] given by I ( X ^ , Y ^ ) g = I ( X , Y ) g = 1 / 2 log ( 1 c g 2 ) I g ( c g ) 0 , as a function of the Gaussian correlation c g (see its graphic in Figure 1), and the second one is the non-Gaussian MI I ( X ^ , Y ^ ) n g = I ( X , Y ) n g , which is due to joint non-Gaussianity and nonlinear statistical relationships among variables.
The MI I ( X ^ , Y ^ ) is related to the negentropy J ( X ^ , Y ^ ) , i.e., to the KL divergence between the PDF and the Gaussian PDF with the same moments of order one and two. That is shown by:
  • Theorem 2: Given ( X ^ r , Y ^ r ) T = A ( X ^ , Y ^ ) T , a pair of rotated standardized variables (A being an invertible 2 × 2 matrix), one has the following result with proof in Appendix 2:
    J ( X ^ , Y ^ ) = J ( X ^ ) + J ( Y ^ ) + I ( X ^ , Y ^ ) I g ( c o r ( X ^ , Y ^ ) ) = = J ( X ^ r , Y ^ r ) = J ( X ^ r ) + J ( Y ^ r ) + I ( X ^ r , Y ^ r ) I g ( c o r ( X ^ r , Y ^ r ) )
A simple consequence is that in the space of uncorrelated variables (i.e., I g ( c o r ( X ^ , Y ^ ) ) = 0 ), the joint negentropy is the sum of marginal negentropies with the MI, thus showing that there are intrinsic and joint sources of non-Gaussianity. One interesting corollary is derived from that.
  • Corollary 1: For standard Gaussian variables ( X , Y ) and standardized rotated ones ( X r , Y r ) T = A ( X , Y ) T , we have
    I n g ( X , Y ) = J ( X , Y ) = J ( X r , Y r ) = J ( X r ) + J ( Y r ) + I ( X r , Y r ) I g ( c o r ( X r , Y r ) )
For the proof it suffices to consider Gaussian variables ( X ^ , Y ^ ) T = ( X , Y ) T in (11). Their self-negentropy vanishes by definition and the correlation term is the Gaussian MI.
Negentropy has the property of being invariant for any orthogonal or oblique rotations of the Gaussianized variables ( X , Y ) . However, this invariance does not extend to I ( X , Y ) n g . From (12), in particular when ( X r , Y r ) are uncorrelated (e.g., standardized principal components of ( X , Y ) or uncorrelated standardized linear regression residuals), the negentropy equals the KL divergence between the joint PDF and that of an isotropic Gaussian with the same total variance. That KL divergence is the compactness (level of concentration around a lower-dimensional manifold), as defined in [25]. This measure is invariant under orthogonal rotations. The last term of (12) vanishes due the fact that the null correlation allows one to decompose of I ( X , Y ) n g into positive contributions: the self-negentropies of uncorrelated variables ( X r , Y r ) and their MI I ( X r , Y r ) = I ( X r , Y r ) g + I ( X r , Y r ) n g . These variables can be ‘Gaussianized’ and rotated, leading to further decomposition of I ( X r , Y r ) n g until the possible “emptying”/depletion of the initial joint non-Gaussianity into Gaussian MIs and univariate negentropies. The PDF of the new rotated variables will be closer to an isotropic spherical Gaussian PDF. Since it is algorithmically easier to compute univariate rather than multivariate entropies, the above method can be used for an efficient estimation of MIs.
The search for rotated variables maximizing the sum of individual negentropies J ( X r ) + J ( Y r ) in (12) with minimization of I ( X r , Y r ) or their statistical dependency is the goal of Independent Component Analysis (ICA) [4].
A natural generalization of the MI decomposition is possible when ( X , Y ) is obtained from a generic ME-anamorphosis by decomposing the MI into a term associated to correlation, under the constraint that marginals are set to given ME-PDFs (the equivalent to Ig), and to a term not explained by correlation (the equivalent to Ing). There is however no guarantee that this decomposition is unique as in the case of non-Gaussians, since there is no natural bivariate extension of univariate prescribed PDFs with a given correlation [30].
By looking again at Equation (12), we notice that when original variables are correlated ( c = c o r ( X r , Y r ) 0 ) , the sum of marginal negentropies is not necessarily a lower bound of the joint negentropy J ( X ^ , Y ^ ) , because in some cases I ( X ^ , Y ^ ) I g ( c ) can be negative. This means that I g ( c ) is not generally a proper lower bound of the MI. An example of that is given by the following discrete distribution with support on four points: ( X ^ , Y ^ ) = ( 1 , 1 ) , ( 1 , 1 ) , ( 1 , 1 ) , ( 1 , 1 ) , with mass probabilities P X ^ Y ^ , respectively of (1 + c)/4, (1 + c)/4, (1 − c)/4 and (1 − c)/4. This 4-point distribution has X ^ and Y ^ zero means, unit variances and Pearson correlation c . The PDF is made of four Dirac-Deltas. In this case, the mutual information I ( X ^ , Y ^ ) is easily computed from the 4-point discrete mean of log ( P X ^ Y ^ / ( P X ^ P Y ^ ) ) while the marginal mass probabilities are P X ^ ( 1 ) = P Y ^ ( 1 ) = P X ^ ( 1 ) = P Y ^ ( 1 ) = 1 / 2 . After some lines of algebra the MI becomes:
I ( X , Y ) = log [ 2 ( 1 + | c | 2 ) ( 1 + | c | 2 ) ( 1 | c | 2 ) ( 1 | c | 2 ) ] I d 4 ( c )
The MI is bounded for any value of c and I d 4 ( c ) log ( 2 ) < as ( | c | 1 ) . The graphic of I d 4 ( c ) is included in Figure 1, showing that I g ( c ) > I d 4 ( c ) for about | c | > 0.45 . This behavior is also reproduced by finite discontinuous PDFs (e.g., replacing the Dirac-Deltas by cylinders of probability) as well as continuous PDFs. We test it by approximating the discrete PDF by the weighted superposition of 4 spherical bivariate Gaussian PDFs, all of which have a sharp isotropic standard deviation σ = 0.001 and are centered at the 4 referred centroids. Figure 1 depicts successively growing MI lower bounds for each value of c, using the maximum entropy method (labels, I g , I g + I n g , 4 , I g + I n g , 6 , I g + I n g , 8 , see Section 3.3 for details), showing the convergence to the asymptotic value log(2).
Figure 1. Semi-logarithmic graphs of I g ( c ) (black thick line), I 4 ( c ) , (grey thick line) and of the successive growing estimates of I 4 ( c ) : I g , I g + I n g , 4 , I g + I n g , 6 and I g + I n g , 8 (grey thin lines). See text for details.
Figure 1. Semi-logarithmic graphs of I g ( c ) (black thick line), I 4 ( c ) , (grey thick line) and of the successive growing estimates of I 4 ( c ) : I g , I g + I n g , 4 , I g + I n g , 6 and I g + I n g , 8 (grey thin lines). See text for details.
Entropy 14 01103 g001

3.3. The Sequence of Non-Gaussian MI Lower Bounds from Cross-Constraints

In order to build a monotonically increasing sequence of lower bounds for I ( X , Y ) (cf. Section 2.3), we have considered a sequence of encapsulated sets of functions whose moments will constrain the ME-PDFs. Those functions consist of single (univariate) and combined (bivariate) monomials of the standard Gaussians X and Y. The numerical implementation of joint ME-PDFs constrained by polynomials in dimensions d = 2, 3 and 4 was studied by Abramov [17,18,19], with particular emphasis on the efficiency and convergence of iterative algorithms. Here, we use the algorithm proposed in [21] and explained in the Appendix 1. Let us define the information moment set T p as the set of bivariate monomials with total order up to p:
T p ( X i Y j : 1 i + j p , ( i , j ) N 0 2 ) , p N
This set is decomposed into marginal (independent) and cross monomials as T p ( T p i n d , T p c r ) with the corresponding vector of expectations θ p ( θ p i n d , θ p c r ) = ( E [ T p i n d ] , E [ T p c r ] ) referring respectively to 2p and p(p − 1)/2 independent and cross moments. The components of θ p are of the form m i , j E [ X i Y j ] , 1 i + j p , where the independent part is set to the moment values of the standard Gaussian: E g [ X p ] = 0 for an odd p and E g [ X p ] = ( p 1 ) ! ! = ( p 1 ) ( p 3 ) ... for an even p (idem for Y), where E g denotes expectation with respect to the standard Gaussian. The cross moments are bounded by a simple application of the Schwartz inequality as ( m i , j ) 2 E g [ X 2 i ] E g [ Y 2 j ] . The monomials of T p are linearly independent and space-generating by linear combinations, thus forming a basis in the sense of vector spaces.
For an even integer p and constraint set T p there exists an integrable ME-PDF with support in , of the form ρ * ( X , Y ) = C exp ( P p ( X , Y ) ) , where P p ( X , Y ) is a pth order polynomial given by a linear combination of monomials in T p with weights given by Lagrange multipliers and C being a normalizing constant. That happens because | ( X i Y j ) | O ( X p + Y p ) , X , Y , 1 i + j p , thus allowing for P p ( X , Y ) as | X | , | Y | , leading to convergence of ME-integrals [18]. For odd p, those integrals diverge because the dominant monomials X p , Y p change sign from positive to negative real and therefore the ME-PDF is not well defined.
In order to build a sequence of MI lower bounds and use the procedure of Section 2.3, we have considered the sequence of information moment sets of even order ( T 2 , θ 2 ) , ( T 4 , θ 4 ) , ( T 6 , θ 6 ) , ... with any pair of consecutive sets satisfying the premises of Theorem 1, i.e., all independent moment sets are ME-congruent. This will lead to the corresponding monotonically growing sequence of lower bounds of MI, denoted as I ( X , Y ) g , 2 I ( X , Y ) g , 4 I ( X , Y ) g , 6 ... , where the subscript g means that variables are marginally standard Gaussian. The first term of the sequence is the Gaussian MI I g , 2 ( X , Y ) I ( X , Y ) g , dependent upon the Gaussian correlation. The difference between the subsequent terms and the first one leads to the non-Gaussian MI of order p defined as I ( X , Y ) n g , p I ( X , Y ) g , p I ( X , Y ) g , 2 I ( X , Y ) n g , which increases with p and converges to I ( X , Y ) n g as p , under quite general conditions [15].
In the same manner as stated in (12), the lower bound I ( X , Y ) n g , p of I ( X , Y ) n g is also a lower bound for the joint negentropy, which is invariant for any affine transformations ( X r , Y r ) T = A ( X , Y ) T , i.e.,
I ( X , Y ) n g , p = H ( X , Y ) 2 H ( X , Y ) p = H ( X r , Y r ) 2 H ( X r , Y r ) p
where H ( X , Y ) p is the bivariate ME associated to ( T p , θ p ) and H ( X , Y ) 2 = 2 H g I ( X , Y ) g . The successive differences I ( X , Y ) n g , p + 2 I ( X , Y ) n g , p are non-negative, with the extra MI being explained by moments of orders p + 1 and p + 2, while not explained by lower-order moments.
There is no analytical closed formula for the dependence of non-Gaussian MI on cross moments. However, under the scenario of low joint non-Gaussianity (small KL divergence to the joint Gaussian), the ME-PDF can be approximated by the Edgeworth expansion [31], based on orthogonal Hermite polynomials and ng approximated as a polynomial of joint bivariate cumulants: k [ i , j ] E [ X r i Y r j ] E g [ X r i ] E g [ Y r j ] , i + j > 2, of any pair of uncorrelated standardized variables ( X r , Y r ) T = A ( X , Y ) T . Cross-cumulants are nonlinear correlations measuring joint non-Gaussianity [32], vanishing when ( X , Y ) are jointly Gaussian. For example I ( X , Y ) n g , p = 4 is approximated as in [13] by the sum of squares:
I ( X , Y ) n g ( E d , p = 4 ) 1 12 [ 3 ( k [ 2 , 1 ] ) 2 + 3 ( k [ 1 , 2 ] ) 2 + ( k [ 3 , 0 ] ) 2 + ( k [ 0 , 3 ] ) 2 ] + 1 48 [ ( k [ 4 , 0 ] ) 2 + 4 ( k [ 3 , 1 ] ) 2 + 6 ( k [ 2 , 2 ] ) 2 + 4 ( k [ 1 , 3 ] ) 2 + ( k [ 0 , 4 ] ) 2 ] + 1 72 [ 5 ( k [ 3 , 0 ] ) 4 + 30 ( k [ 3 , 0 ] k [ 2 , 1 ] ) 2 + 2 3 ( 9 k [ 3 , 0 ] k [ 1 , 2 ] + 6 k [ 2 , 1 ] k [ 2 , 1 ] ) 2 + 2 3 ( 9 k [ 0 , 3 ] k [ 2 , 1 ] + 6 k [ 1 , 2 ] k [ 1 , 2 ] ) 2 + 1 2 ( 2 k [ 3 , 0 ] k [ 0 , 3 ] + 18 k [ 2 , 1 ] k [ 1 , 2 ] ) 2 + 30 ( k [ 1 , 2 ] k [ 0 , 3 ] ) 2 + 5 ( k [ 0 , 3 ] ) 4 ] = I ( X , Y ) n g + O ( n e q 3 / 2 )
where ( X , Y ) is assumed to be the arithmetic average of an equivalent number neq of independent and identically distributed (iid) bivariate RVs. Therefore, from the multidimensional Central Limit Theorem [33], the larger neq is, the closer the distribution is to joint Gaussianity, and the smaller the absolute value of cumulants become.

3.4. Non-Gaussian MI across the Polytope of Cross Moments

The ( X , Y ) cross moments in the expectation vector θ p (p even) are not completely free. Rather, they satisfy to Schwarz-like inequalities defining a compact set D p within which cross-moments lie. That set portrays all the possible non-Gaussian ME-PDFs with p-order independent moments equal to those of the standard Gaussian. Under these conditions I ( X , Y ) n g = I ( X , Y ) n g , p . In order to have a better feel on how ng behaves, we have numerically evaluated I n g , p = 4 along the allowed set of cross-moments.
In order to determine that set, let us begin by invoking some generalities about polynomials. Any bivariate polynomial P p ( x , y ) : 2 of total order p is expressed as a linear combination of linearly independent monomials from the basis T p 1 T p { 1 } , obtained from T p including unity. Then:
P p ( x , y ) = i a i T i , p ( x , y ) , T i , p T p 1
If the condition P p ( x , y ) 0 , ( x , y ) 2 holds i.e., P p ( x , y ) is positive semi-definite (PSD), then its expectation is non-negative: E [ P p ( x , y ) ] = i a i θ i , p 0 , where θ i , p E [ T i , p ] , thus imposing a constraint on the components of θ p . A sufficient condition for the positiveness is that P p ( x , y ) is a sum of squares (SOS) or, without loss of generality, the square of a certain polynomial Q p / 2 ( x , y ) = b T T p / 2 1 of total order p/2, where b is a column vector of coefficients multiplying monomials of T p / 2 1 . Then, P p ( x , y ) is written as a quadratic form P p ( x , y ) = ( Q p / 2 ( x , y ) ) 2 = b T ( T p / 2 1 T p / 2 1 T ) b 0 . By taking the expectation operator we have b T E [ ( T p / 2 1 T p / 2 1 T ) ] b 0 , b , which implies the positiveness of the matrix of moments E [ ( T p / 2 1 T p / 2 1 T ) ] C p , which is given in terms of components of θ p .
When p = 4 and d = 2, the case of bivariate quartics, any PSD polynomial is a SOS [34] and vice versa. However, for p ≥ 6 there are PSD-non-SOS polynomials (e.g., those coming from the inequality between arithmetic and geometric means [35]). Therefore, a necessary and sufficient condition among fourth-order moments is that E [ ( T 2 1 T 2 1 T ) ] C 4 be a PSD matrix. Let us study the conditions for that.
By ordering the set T 2 1 ( 1 , x , y , x 2 , y 2 , x y ) T , one has the 6 × 6 matrix C 4 , written in the simplified form in terms of moments:
C 4 = [ ( 1 , 0 , 0 , 1 , 1 , c g ) ( 0 , 1 , c g , 0 , m 1 , 2 , m 2 , 1 ) ( 0 , c g 1 , m 2 , 1 , 0 , m 1 , 2 ) ( 1 , 0 , m 2 , 1 3 , m 2 , 2 , m 3 , 1 ) ( 1 , m 1 , 2 0 , m 2 , 2 3 , m 1 , 3 ) ( c g , m 2 , 1 , m 1 , 2 , m 3 , 1 , m 1 , 3 , m 2 , 2 ) ]
A necessary and sufficient condition for the positiveness of C 4 is given by the application of the Sylvester criterion, stating that the determinants d1, d2, d3, d4, d5 and d6 of the 6 upper sub-matrices of C 4 are positive. From these, only those of orders 4, 5 and 6 lead to nontrivial relationships, given with help of Mathematica® [36] as:
d 4 2 2 c g 2 m 2 , 1 2   0 ; d 4 , d u a l 2 2 c g 2 m 1 , 2 2   0
d 5 2   ( m 2 , 1 2 + m 2 , 1 2 ) + m 1 , 2 2   m 2 , 1 2 2   m 1 , 2 m 2 , 1   ( m 2 , 2 1 )   c g + ( m 2 , 2 3 )   ( m 2 , 2 + 1 )   ( 1 + c g 2 ) 0
d 6 ( 9 m 2 , 2 2 ) c g 4 + 2 [ ( m 1 , 3 + m 3 , 1 ) ( m 2 , 2 3 ) + m 1 , 2   m 2 , 1   m 2 , 2 ] c g 3 + [ ( m 2 , 1 2 + m 1 , 2 2 ) ( 9 2 m 2 , 2 ) m 1 , 2   m 2 , 1 ( 2 ( m 1 , 3 + m 3 , 1 ) + m 1 , 2   m 2 , 1 ) + 2 ( m 3 , 1 2 + m 1 , 3 2 + m 1 , 3   m 3 , 1 ( 1 m 2 , 2 ) ) + ( m 2 , 2 3 ) ( 3 + m 2 , 2 ( 2 + m 2 , 2 ) ) ] c g 2 + 2 [ ( 3 2 m 2 , 1 2 2 m 1 , 2 2 m 2 , 2 ) ( m 1 , 3 + m 3 , 1 ) + m 2 , 2 ( m 1 , 2 2 m 3 , 1 + m 2 , 1 2 m 1 , 3 ) + m 1 , 2   m 2 , 1 ( m 1 , 3   m 3 , 1 + m 2 , 1 2 + m 1 , 2 2 3 + m 2 , 2 ( 5 2 m 2 , 2 ) ) ] c g + [ m 2 , 2 3 + ( m 2 , 1 2 + m 1 , 2 2 + 2 ) m 2 , 2 2 + ( 3 4 ( m 2 , 1 2 + m 1 , 2 2 ) + m 1 , 2   m 2 , 1 ( 3 m 1 , 2   m 2 , 1 2 ( m 1 , 3 + m 3 , 1 ) ) ) m 2 , 2 + m 1 , 2   m 2 , 1 ( 2 ( m 1 , 2 2 m 3 , 1 + m 2 , 1 2 m 1 , 3 + m 1 , 2   m 2 , 1 ) + 6 ( m 1 , 3 + m 3 , 1 ) ) + 2 ( m 2 , 1 4 + m 1 , 2 4 ) + 3 ( m 2 , 1 2 + m 1 , 2 2 ) 2 ( m 1 , 3 2 + m 3 , 1 2 ) 2 m 1 , 3   m 3 , 1 + ( m 1 , 3   m 2 , 1 ) 2 + ( m 3 , 1   m 1 , 2 ) 2 ] 0
In Equation (22), the inequality for d 4 has a dual relationship (d4,dual), its sign being reversed by swapping the two indices in mi,j, whereas d 5 and d 6 are symmetric with respect to indicial swap. The term d6 of Equation (21) is a fourth-order polynomial of cg. The inequalities for d 4 , d 4 d u a l , d 5 and d 6 hold inside a compact domain denoted D4, with the shape of what resembles a rounded polytope in the space of cross moments ( m 1 , 2 ,   m 2 , 1 ,   m 1 , 3 ,   m 3 , 1 ,   m 2 , 2 ), for each value of the Gaussian correlation cg. The case of Gaussianity lies within the interior of D4, corresponding to m 1 , 2 = m 2 , 1 = 0 ;   m 1 , 3 = m 3 , 1 = 3 c g m 2 , 2 = 2 c g 2 + 1 , thus defining the hereafter called one-dimensional ‘Gaussian manifold’.
In order to illustrate how non-Gaussianity depends on moments of third and fourth order, we have computed the non-Gaussian MI of order 4 ( I n g , p = 4 ), along a set of 2-dimensional cross-sections of D4 crossing the Gaussian manifold and extending up the boundary of D4. For Gaussian moments, I n g , 4 vanishes, being approximated by the Edgeworth expansion (16) near the Gaussian manifold.
In order to get a picture of I n g , p = 4 , we have chosen six particular cross-sections of D4 by varying two moments and setting the remaining to their ‘Gaussian values’. The six pairs of varying parameters are: A (cg, m2,1), B (cg, m3,1), C (cg, m2,2), D (m2,1, m3,1) at cg=0, E (m2,1, m2,2) at cg=0 and F (m3,1, m2,2) at cg = 0, with the contours of the corresponding Ing field shown in Figure 2a–f. The fields are retrieved from a discrete mesh of 100 × 100 in moment space. The Gaussian state lies at: (cg, 0), (cg, 3cg), (cg, 2cg2 + 1), (0, 0), (0, 1) and (0, 1), respectively for cases A up to F. The moment domains are obtained by solving inequalities for d4, d5 and d6 and applying the restrictions imposed by the crossing of the Gaussian manifold (e.g., m1,2 = 0, m1,3 = m3,1 = 3cg, m2,2 = 2cg2 + 1 for case A). We obtain the following restrictions for cases A–F:
Case   A : m 2 , 1 2 1 2 [ 3 + 3 c g 2 ( 1 2 c g 2 ) ( 1 c g 2 ) ( 1 + 52 c g 2 + 28 c g 4 ) 1 / 2 ]
Case   B : c g + 2 c g 3 2 ( 1 c g 2 c g 4 + c g 6 ) m 3 , 1 c g + 2 c g 3 + 2 ( 1 c g 2 c g 4 + c g 6 )
Case   C : 1 2 [ 1 + c g 2 + ( 1 + 34 c g 2 + c g 4 ) 1 / 2 ] m 2 , 2 3
Case   D : m 3.1 2 2 3 m 2 , 1 2 + 4 m 2 , 1 4 ; | m 21 | 1 ; | m 21 | 2
Case   E : m 2.1 2 m 2 , 2 1 + ( 4 2 m 2 , 1 2 ) 1 / 2 ; m 2 , 1 2 3
Case   F :   m 2 , 2 3 2 m 2 , 2 2 3 m 2 , 2 + 2 m 3 , 1 2 0
The analytical boundary of allowed domains is emphasized with thick solid lines in all Figure 2a–f. There are some common aspects among the figures. As expected, Ing vanishes at the Gaussian states, the Gaussian manifold, marked with G in Figure 2. Ing grows monotonically towards the boundary of the moment domains D4. There, det ( C 4 ) = 0 , meaning that C 4 is singular and there is a vector ( b 0 ) K e r ( C 4 ) . This holds if one gets the deterministic relationship Q p / 2 ( x , y ) = b T T 2 1 = 0 , leading to a Dirac-Delta-like ME-PDF along a one-dimensional curve. This in turn leads to Ing = ∞, except possibly in a set of singular points of D4 on which Ing is not well defined. In practice, infinity is not reached due to stopping criteria for convergence of the iterative method used for obtaining the ME‑PDF.
Figure 2. Field of the non-Gaussian MI I n g , 4 along 6 bivariate cross-sections of the set of allowed moments. Two varying moments are featured in each cross-section: A (cg, m2,1), B (cg, m3,1), C (cg, m2,2), D (m2,1, m3,1) at cg=0, E (m2,1, m2,2) at cg = 0 and F (m3,1, m2,2) at cg = 0. The letter G indicates points or curves where Gaussianity holds.
Figure 2. Field of the non-Gaussian MI I n g , 4 along 6 bivariate cross-sections of the set of allowed moments. Two varying moments are featured in each cross-section: A (cg, m2,1), B (cg, m3,1), C (cg, m2,2), D (m2,1, m3,1) at cg=0, E (m2,1, m2,2) at cg = 0 and F (m3,1, m2,2) at cg = 0. The letter G indicates points or curves where Gaussianity holds.
Entropy 14 01103 g002
At states where |cg| = 1, Ig = ∞ and Ing has a second-kind singularity discontinuity where the contours merge together without a well-defined limit for Ing. In the neighborhood of the Gaussian state with cg = 0 in Figure 2d–f, Ing is approximated by the quadratic form (16) as is confirmed by the elliptic shape of Ing contours. The value of Ing can surpass Ig, thus emphasizing the fact that in some cases much of the MI may come from nonlinear (X,Y) correlations.
The joint entropy is invariant for a mirror symmetry in one or both variables: X X or Y Y , because the absolute value of the determinant of that transformation equals 1. As a consequence, the dependence of the Gaussian and non-Gaussian MI on moments also reflects these intrinsic mirror symmetries. For instance, in Figure 2d, the symmetry X X leads to the dependency relations Ing(m2,1, m3,1) = Ing(m2,1, −m3,1), where arguments are the varying moments, while symmetry Y Y leads to Ing(m2,1, m3,1) = Ing(−m2,1, −m3,1). The symmetries corresponding to the remaining figures are: Ing(cg, m2,1) = Ing(−cg, m2,1) = Ing(cg, −m2,1); Ing(cg, m3,1) = Ing(−cg, −m3,1); Ing(cg, m2,2) = Ing(−cg, m22); Ing(m2,1, m2,2) = Ing(−m2,1, m2,2); Ing(m3,1, m2,2) = Ing(−m3,1, m2,2), respectively for Figure 2a–c, e and f.
Near the boundary of D 4 , the ME problem is very ill-conditioned. The Hessian matrix of the corresponding ME functional is the inverse M 4 1 , where M 4 E [ T 4 T 4 T ] E [ T 4 ] E [ T 4 T ] is the covariance matrix of T 4 , made by 8 independent plus 6 cross moments. The condition number CN (ratio between the largest and smallest eigenvalue) of M 4 (and of M 4 1 ) tends to ∞ at the boundary of D 4 where M 4 is singular, due to the above deterministic relationship. Therefore, the closer that boundary is, the closer the ME-PDF is to a deterministic relationship, the more ill conditioned the ME problem is and the slower the numerical convergence of the optimization algorithm becomes.

4. The Effect of Noise and Nonlinearity on Non-Gaussian MI

The aim of this section is an exploratory analysis of the possible sources of non-Gaussianity in a bivariate statistical relationship. Towards that aim, we explore the qualitative behavior of Ig and Ing between a standardized signal X ^ (with null mean and unit variance) and an X ^ -dependent standardized response variable Y ^ contaminated by noise. For this purpose, a full range of signal-to-noise variance ratios (snr) shall be considered, from pure signal to pure noise. The statistics are evaluated from one‑million-long synthetic ( X ^ , Y ^ ) iid independent realizations produced by a numeric Gaussian random generator. Many interpretations are possible for the output variable: (i) Y ^ taken as the observable outcome emerging from a noisy transmission channel fed by X ^ (ii) given by the direct or indirect observation affected by measurement and representativeness errors corresponding to a certain value X ^ of the model state vector [37] (iii) the outcome from a stochastic or deterministic dynamical system [38].
In order to estimate I ( X ^ , Y ^ ) , the working variables ( X ^ , Y ^ ) are transformed by anamorphosis into standard Gaussian variables ( X , Y ) .
We consider, without loss of generality, X = X ^ . The variable Y is given by Gaussian anamorphosis Y = G Y ^ ( Y ^ ) ~ N ( 0 , 1 ) as in Equation (8), with:
Y ^ = s F ( X ) + ( 1 s 2 ) 1 / 2 n ( X , W ) ; s [ 0 , 1 ]
where F ( X ) is a purely deterministic transfer function and n ( X , W ) is a scalar noise uncorrelated with F ( X ) , depending in general on X (e.g., multiplicative noise) and from a vector W of independent Gaussians contaminating the signal. Both F ( X ) and n ( X , W ) have unit variance with n ( X , 0 ) = 0 . The signal-to-noise variance ratio is s n r = s 2 / ( 1 s 2 ) .
Then, the Gaussian MI Ig is computed for each value of s∈[0,1] and compared among several scenarios of F ( X ) and n ( X , W ) . A similar comparison is done for the non-Gaussian MI, approximated here by Ing,p=8. Six case studies have been considered (A, B, C, D, E and F); their signal and noise terms are summarized in Table 1, along with the colors with which they are represented in Figure 3 further below.
Table 1. Types of signal and noise in Equation (35) and corresponding colors used in Figure 3.
Table 1. Types of signal and noise in Equation (35) and corresponding colors used in Figure 3.
Case study F ( X ) n ( X , W ) Color
A—Gaussian noise (reference) X W Black
B—Additive non-Gaussian independent noise X W 3 / 15 Red
C—Multiplicative noise W X Blue
D—Smooth nonlinear homeomorphism X 3 / 15 W Magenta
E—Smooth non-injective transformation ( X 3 X ) / 10 Green
F—Combined non-Gaussianity ( X 3 X ) / 10 X W 1 / 2 + W 2 3 / 30 Cyan
In Table 1, W , W 1 , W 2 are independent standard Gaussian noises. We begin with a reference case, A, which refers to Gaussian noise. Case B refers to a symmetric leptokurtic (i.e., with kurtosis larger than that of the Gaussian) non-Gaussian noise. In case C the multiplicative noise depends linearly on the signal X. In case D, the signal is a nonlinear cubic homeomorphism of the real domain. For case E, the signal is nonlinear and not injective in the interval [−1, 1], thus introducing ambiguity in the relationship ( X , Y ) . Finally, in case F all the factors—non-Gaussian noise, multiplicative noise and signal ambiguity—are pooled together.
Figure 3. Graphs depicting the total MI (a), Gaussian MI (b) and non-Gaussian MI (c) of order 8 for 6 cases (A–F) of different signal-noise combinations with the signal weight s in abscissas varying from 0 up to 1. See text and Table 1 for details about the cases and their color code.
Figure 3. Graphs depicting the total MI (a), Gaussian MI (b) and non-Gaussian MI (c) of order 8 for 6 cases (A–F) of different signal-noise combinations with the signal weight s in abscissas varying from 0 up to 1. See text and Table 1 for details about the cases and their color code.
Entropy 14 01103 g003
Figure 3a,b show the graphics of the total MI, estimated by Ig + Ing,8 and of the Gaussian MI (Ig) for the six cases (A to F). The graphic of non-Gaussian MI as approximated by Ing,8 is depicted in Figure 3c for five cases (B to F). In Figure 4, we show a ‘stamp-format’ collection of the contouring of ME-PDFs of polynomial order p = 8 for all cases (A to F) and extreme and intermediate cases of the snr: s = 0.1, s = 0.5 and s = 0.9. This illustrates how the snr and the nature of both the transfer function and noises influence the PDFs.
For the Gaussian noise case (A), the non-Gaussian MI is theoretically set to zero since the joint distribution of ( X , Y ) is Gaussian. In all scenarios, both Ig and the total MI Ig + Ing grow, as expected, with the snr. This is in accordance to the Bruijn’s equality stating the positiveness of the MI derivative with respect to snr and established in the literature of signal processing for certain types of noise [39,40]. On the contrary, the monotonic behavior as a function of snr is not a universal characteristic of the non-Gaussian MI.
By observing Figure 3a–c, the following qualitative results are worth mentioning. We begin by comparing the total MI in three cases (A, B and C), which share the same linear signal but feature noise of different kinds (Figure 3a). Both the red (B) and blue (C) lines lie above the black line (A) for each given s, thus indicating that the total MI is lowest when the noise is Gaussian. This means that the Gaussian noise is the most signal degrading of noises with the same variance [41]. The extra MI found in the B and C cases come, respectively, from the Gaussian MI (see case B in Figure 3c) and from the non-Gaussian MI (see case C in Figure 3a), as it is also apparent by looking at ME-PDFs for cases B, C (s = 0.1) (Figure 4).
We consider now the cases A, D, E, all of which have a Gaussian noise. Their differences lie in the signals, with the one in A being linear and the ones in D and E being nonlinear. By comparing these cases it is seen that Ig is highest for the linear signal, the black curve lying above the magenta (D) and green (E) curves for each s in Figure 3b. This indicates that the Gaussian MI, measuring the degree of signal linearity, is lower when the signal introduces nonlinearity (cases D and E) than when no nonlinearity is present (case A).
Figure 4. Collection of stamp-format ME-PDFs for cases AF (see text for details) and signal weight s = 0.1 (a), s = 0.5 (b) and s = 0.9 (c) over the [−3, 3]2 support set.
Figure 4. Collection of stamp-format ME-PDFs for cases AF (see text for details) and signal weight s = 0.1 (a), s = 0.5 (b) and s = 0.9 (c) over the [−3, 3]2 support set.
Entropy 14 01103 g004
It is worth noting that, while the signals in A and D are injective, the one in E is not, thus introducing ambiguity. This will imply loss of information in E, which is visible in the total MI depicted for each s in Figure 3a. In fact, there the green curve (case E) lies lower than the black (A) and magenta (D) curves for every s. The effect of nonlinearity is quite evident in ME-PDFs, in particular for high s value (Figure 4, cases D, E, s = 0.9).
We focus now on the non-Gaussian MI, depicted in Figure 3c for each s. The curve representing the case B, with a linear signal and a state-independent noise, indicates that the non-Gaussian MI is null for both s = 0 and s = 1. The first zero of non-Gaussian MI (at s = 0) is justified by the noise being state‑independent, whereas the second zero (at s = 1) is due to the signal being linear, which means that all the MI resides in the Gaussian MI. The non-Gaussian MI is thus positive and maximum at intermediate values of s.
By looking at case C (multiplicative noise), it is seen that the non-Gaussian MI remains roughly unchanged for every s < 1. This holds even at s = 0 (pure noise), since the noise is state-dependent and thus some information is already present. At s = 1 the non-Gaussian MI is null due to the signal being linear (as in case B).
By observing the cases with Gaussian noise and nonlinear signals (D and E) in Figure 3c, it can be seen that their non-Gaussian MI grows with s (and thus with the relative weight of the signal), due to their signals being nonlinear. This gradual behavior is also reflected in the ME-PDFs (Figure 4, cases D, E along the s values).
Finally, we consider the case in which the signal is nonlinear and the noise comprises a multiplicative and a non-Gaussian additive component (case F). As compared with E (which differs from it in that the noise is Gaussian), it can been that non-Gaussian MI is always larger in F independently of s. This is due to the fact that in F there is information even at s = 0, due to the state-dependence of its noise. For all values of s, the ME-PDF exhibits quite a large deviation from Gaussianity.

7. Discussion and Conclusions

We have addressed the problem of finding the minimum mutual information (MinMI), or the least noncommittal MI between d = 2 random variables, consistent with a set of marginal and joint expectations. The MinMI is a proper MI lower bound when marginals are set to ME-PDFs through appropriate nonlinear single anamorphoses. Moreover, the MinMI increases as long as one increases number of independent cross-constraints of the bivariate ME problem. Considering a sequence of moments, we have obtained a hierarchy of lower MI bounds approximating the total MI value. The method can easily be generalized for d > 2 variables with the necessary adaptations.
One straightforward application of that principle follows from the MI estimation from ‘Gaussianized’ variables with real support, where the marginals are rendered standard Gaussian N(0,1) by Gaussian anamorphosis. This allows for the MI decomposition into two positive contributions: a Gaussian term Ig, which depends uniquely on the Gaussian correlation cg (Pearson correlation in the space of ‘Gaussianized’ variables), and a non-Gaussian term Ing depending on nonlinear correlations. This term is equal to the joint negentropy, which is invariant for any oblique or orthogonal rotation of the ‘Gaussianized’ variables and is related to the ‘compactness’ measure or the closeness of the PDF towards a low manifold deterministic relationship. The Gaussian MI is also a ‘concordance’ measure, invariant for any monotonically growing homeomorphisms of marginals and consequently expressed as a functional of the copula-density function, which is exclusively dependent on marginal cumulated probabilities. In certain extreme cases, very far from Gaussianity, the Pearson correlation among non‑Gaussian variables is not a proper measure of the mutual information. An example of that situation is given.
Cross moments under marginal standard Gaussians are bounded by Schwarz-like inequalities defining compact sets, the shape of which resemble a rounded polytope where cross moments live. The allowed moment values portray all possible joint PDFs with Gaussian marginals. Inside that set lies the so called one-dimensional Gaussian manifold, parametrized by cg, where joint Gaussinity holds. There, Ing vanishes, growing towards infinity as far as the boundary is approached, where variables satisfy a deterministic relationship and the ME problem is ill conditioned. This behavior is illustrated in cross-sections of the polytope of cross moments of total order p = 4.
In order to systematize the possible sources of Gaussian and non-Gaussian MI, we have computed it in the context of nonlinear noisy channels. The MI has been computed between a Gaussian input and a panoply of (linear and/or nonlinear) outputs contaminated by different kinds of noise for a full range of the signal-to-noise variance ratio. Sources of non-Gaussian MI include: (a) the nonlinearity of the signal transfer function, (b) multiplicative noise and (c) non-Gaussian additive noise. This paper is followed by a companion one [27] on the estimation of non-Gaussian MI from finite samples with practical applications.

Acknowledgments

This research was developed at IDL with the financial support of project PEST-OE/CTE/LA0019/2011-FCT. Thanks are due to two anonymous reviewers, Miguel Teixeira and J. Macke for some discussions and our families for the omnipresent support.

References

  1. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley & Sons, Inc.: New York, NY, USA, 1991. [Google Scholar]
  2. Ebrahimi, N.; Soofi, E.S.; Soyer, R. Information measures in perspective. Int. Stat. Rev. 2010, 78, 383–412. [Google Scholar] [CrossRef]
  3. Stögbauer, H.; Kraskov, A.; Astakhov, S.A.; Grassberger, P. Least-dependent-component analysis based on mutual information. Phys. Rev. E 2004, 70, 066123:1–066123:17. [Google Scholar] [CrossRef]
  4. Hyvärinen, A.; Oja, E. Independent Component Analysis: Algorithms and application. Neural Network. 2000, 13, 411–430. [Google Scholar] [CrossRef]
  5. Fraser, H.M.; Swinney, H.L. Independent coordinates for strange attractors from mutual information. Phys. Rev. A 1986, 33, 1134–1140. [Google Scholar] [CrossRef] [PubMed]
  6. DelSole, T. Predictability and information theory. Part I: Measures of predictability. J. Atmos. Sci. 2004, 61, 2425–2440. [Google Scholar] [CrossRef]
  7. Majda, A.; Kleeman, R.; Cai, D. A mathematical framework for quantifying predictability through relative entropy. Methods and applications of analysis. Meth. Appl. Anal. 2002, 9, 425–444. [Google Scholar]
  8. Darbellay, G.A.; Vajda, I. Entropy expressions for multivariate continuous distributions. IEEE Trans. Inform. Theor. 2000, 46, 709–712. [Google Scholar] [CrossRef]
  9. Nadarajah, S.; Zografos, K. Expressions for Rényi and Shannon entropies for bivariate distributions. Inform. Sci. 2005, 170, 173–189. [Google Scholar] [CrossRef]
  10. Khan, S.; Bandyopadhyay, S.; Ganguly, A.R.; Saigal, S.; Erickson, D.J.; Protopopescu, V.; Ostrouchov, G. Relative performance of mutual information estimation methods for quantifying the dependence among short and noisy data. Phys. Rev. E 2007, 76, 026209:1–026209:15. [Google Scholar] [CrossRef]
  11. Walters-Williams, J.; Li, Y. Estimation of mutual information: A survey. Lect. Notes Comput. Sci. 2009, 5589, 389–396. [Google Scholar]
  12. Globerson, A.; Tishby, N. The minimum information principle for discriminative learning. In Proceedings of the 20th conference on Uncertainty in artificial intelligence, Banff, Canada, 7–11 July 2004; pp. 193–200.
  13. Jaynes, E.T. On the rationale of maximum-entropy methods. Proc. IEEE 1982, 70, 939–952. [Google Scholar] [CrossRef]
  14. Wackernagel, H. Multivariate Geostatistics—An Introduction with Applications; Springer-Verlag: Berlin, Germany, 1995. [Google Scholar]
  15. Shams, S.A. Convergent iterative procedure for constructing bivariate distributions. Comm. Stat. Theor. Meth. 2010, 39, 1026–1037. [Google Scholar] [CrossRef]
  16. Ebrahimi, N.; Soofi, E.S.; Soyer, R. Multivariate maximum entropy identification, transformation, and dependence. J. Multivariate Anal. 2008, 99, 1217–1231. [Google Scholar] [CrossRef]
  17. Abramov, R. An improved algorithm for the multidimensional moment-constrained maximum entropy problem. J. Comput. Phys. 2007, 226, 621–644. [Google Scholar] [CrossRef]
  18. Abramov, R. The multidimensional moment-constrained maximum entropy problem: A BFGS algorithm with constraint scaling. J. Comput. Phys. 2009, 228, 96–108. [Google Scholar] [CrossRef]
  19. Abramov, R. The multidimensional maximum entropy moment problem: A review on numerical methods. Commun. Math. Sci. 2010, 8, 377–392. [Google Scholar] [CrossRef]
  20. Rockinger, M.; Jondeau, E. Entropy densities with an application to autoregressive conditional skewness and kurtosis. J. Econometrics 2002, 106, 119–142. [Google Scholar] [CrossRef]
  21. Pires, C.A.; Perdigão, R.A.P. Non-Gaussianity and asymmetry of the winter monthly precipitation estimation from the NAO. Mon. Wea. Rev. 2007, 135, 430–448. [Google Scholar] [CrossRef]
  22. Kraskov, A.; Stögbauer, H.; Grassberger, P. Estimating mutual information. Phys. Rev. E 2004, 69, 066138:1–066138:16. [Google Scholar] [CrossRef]
  23. Myers, J.L.; Well, A.D. Research Design and Statistical Analysis, 2nd ed.; Lawrence Erlbaum Associates: Mahwah, NJ, USA, 2003. [Google Scholar]
  24. Calsaverini, R.S.; Vicente, R. An information-theoretic approach to statistical dependence: Copula information. Europhys. Lett. 2009, 88, 68003. 1–6. [Google Scholar] [CrossRef]
  25. Monahan, A.H.; DelSole, T. Information theoretic measures of dependence, compactness, and non-Gaussianity for multivariate probability distributions. Nonlinear Proc. Geoph. 2009, 16, 57–64. [Google Scholar] [CrossRef]
  26. Guo, D.; Shamai, S.; Verdú, S. Mutual information and minimum mean-square error in Gaussian channels. IEEE Trans. Inform. Theor. 2005, 51, 1261–1283. [Google Scholar] [CrossRef]
  27. Pires, C.A.; Perdigão, R.A.P. Minimum mutual information and non-Gaussianity through the maximum entropy method: Estimation from finite samples. Entropy 2012. submitted for publication. [Google Scholar]
  28. Shore, J.E.; Johnson, R.W. Axiomatic derivation of the principle of maximum entropy and the principle of the minimum cross-entropy. IEEE Trans. Inform. Theor. 1980, 26, 26–37. [Google Scholar] [CrossRef]
  29. Koehler, K.J.; Symmanowski, J.T. Constructing multivariate distributions with specific marginal distributions. J. Multivariate Anal. 1995, 55, 261–282. [Google Scholar] [CrossRef]
  30. Cruz-Medina, I.R.; Osorio-Sánchez, M.; García-Páez, F. Generation of multivariate random variables with known marginal distribution and a specified correlation matrix. InterStat 2010, 16, 19–29. [Google Scholar]
  31. van Hulle, M.M. Edgeworth approximation of multivariate differential entropy. Neural Computat. 2005, 17, 1903–1910. [Google Scholar] [CrossRef]
  32. Comon, P. Independent component analysis, a new concept? Signal Process. 1994, 36, 287–314. [Google Scholar] [CrossRef] [Green Version]
  33. van der Vaart, A.W. Asymptotic Statistics; Cambridge University Press: New York, NY, USA, 1998. [Google Scholar]
  34. Hilbert, D. Über die Darstellung definiter Formen als Summe von Formenquadraten. Math. Ann. 1888, 32, 342–350. [Google Scholar] [CrossRef]
  35. Ahmadi, A.A.; Parrilo, P.A. A positive definite polynomial hessian that does not factor. In Proceedings of the Joint 48th IEEE Conference on Decision and Control and 28th Chinese Control Conference, Shanghai, China, 15–18 December 2009; pp. 16–18.
  36. Wolfram, S. The Mathematica Book, 3rd ed.; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  37. Bocquet, M.; Pires, C.; Lin, W. Beyond Gaussian statistical modeling in geophysical data assimilation. Mon. Wea. Rev. 2010, 138, 2997–3023. [Google Scholar] [CrossRef]
  38. Sura, P.; Newman, M.; Penland, C.; Sardeshmuck, P. Multiplicative noise and non-Gaussianity: A paradigm for atmospheric regimes? J. Atmos. Sci. 2005, 62, 1391–1406. [Google Scholar] [CrossRef]
  39. Guo, D.; Shamai, S.; Verdú, S. Mutual information and minimum mean-square error in Gaussian channels. IEEE Trans. Inform. Theor. 2005, 51, 1261–1283. [Google Scholar] [CrossRef]
  40. Rioul, O. A simple proof of the entropy-power inequality via properties of mutual information. arXiv 2007. arXiv:cs/0701050v2 [cs.IT]. [Google Scholar]
  41. Guo, D.; Wu, Y.; Shamai, S.; Verdú, S. Estimation in gaussian noise: Properties of the minimum mean-square error. IEEE T. Inform. Theory 2011, 57, 2371–2385. [Google Scholar]
  42. Gilbert, J.; Lemaréchal, C. Some numerical experiments with variable-storage quasi-Newton algorithms. Math. Program. 1989, 45, 407–435. [Google Scholar] [CrossRef]

Appendix 1
Form and Numerical Estimation of ME-PDFs

We hereby present a summary of the ME method for distributions along with the numerical algorithm for computing it. Let us consider an information moment set (2) ( T , θ ) comprising J constraints on the RVs ( X , Y ) with the support set S = S X S Y 2 . The associated ME-PDF is obtained from the absolute minimum of a control function:
L ( η , T , θ ) = log [ Z ( η ) ] i = J η i θ i
in terms of a J-dimensional vector of Lagrange multipliers:
η = ( η 1 , ... , η J ) T
where Z ( η ) = S exp [ i = 1 J η i T i ( x , y ) ] d x d y [13]. The minimum of L lies at a value of η dependent on ( T , θ ) , given by: λ ( T , θ ) = arg min η L ( η , T , θ ) . The corresponding L minimum is the value of the maximum entropy: H ρ T , θ * = L ( λ , T , θ ) .
The ME-PDF is of the form: ρ T , θ * ( x , y ) = Z ( λ ) 1 exp [ i = 1 J λ i T i ( x , y ) ] , where Z ( λ ) is the normalization partition function. Except when no analytical relationship λ ( T , θ ) exists, this function has to be estimated by iterative techniques of minimization of L ( η , T , θ ) . The numerical algorithm consists of a bivariate version of that presented in [20]. In practice, we have solved the ME problem for a finite square support set S r = [ r , r ] 2 with r large enough in order to prevent significant boundary effects on the ME-PDF, thus obtaining a good estimation of the ME-PDF asymptotic limit when r → ∞. By using the two-dimensional Leibnitz differentiation rule, it is easy to obtain the derivative of the ME H ρ * with respect to r:
d H ρ * ( X , Y ) d r = ( x , y ) S r ρ * ( x , y ) | d l |
where the line integral is always positive and computed along the boundary line S r of S r . When |x|,|y| → ∞, the bound of (A1) leads to the scaling of the logarithm of the ME-PDF as O ( r p ) , where p is the maximum total order of the constraining bivariate monomials in the information moment set ( T , θ ) . Furthermore, in order to get integrands of the order exp ( O ( 1 ) ) during the optimization process, we solve the ME problem for the scaled variables (X/r,Y/r) in the square [−1, 1]2, by taking the appropriate scaled constraints. Then, we apply the scaling entropy relationship:
H ρ * ( X , Y ) = H ρ * ( X / r , Y / r ) + 2 log ( r )
The integrals giving the functional L ( η , T , θ ) and its η-derivatives are approximated by the bivariate Gauss truncation rule with Nf weighting factors each in the interval [−1, 1]. In order to get full resolution during the minimization, and to avoid “Not-a-Number” (NAN) and infinity (INF) errors in computation, we subtract the polynomials in the arguments of exponentials from the corresponding maximum in S. Finally, the functional L is multiplied by a sufficient high factor F in order to emphasize the gradient. After some preliminary experiments, we have set r = 6, Nf = 80, F = 1000. Convergence is assumed to be reached when one gets an accuracy of 10−6 for the gradient of L . By setting the first guess of Lagrange multipliers (FGLM) to zero, convergence is reached after about 60–400 iterations. For the optimization we have used the routine M1QN3 from INRIA [42], which uses the Quasi-Newton BFGS algorithm. Convergence is slower under a higher condition number (CN) of the Hessian of L , with the convergence time growing in general with the proximity of the boundary of the domain of allowed moments as well as the total maximum order p of constraining monomials. The convergence is faster when a closer FGLM is provided to the exact solution. This is possible in sequential ME problems with quite small successive constraint differences. There, one uses a continuation technique by setting FGLM to the optimized Lagrange multipliers from the previous ME problem. This technique has been used in the computation of the graphics in Figure 3.

Appendix 2

  • Proof of Theorem 1: Since X follows the ME-PDF generated by ( T X , θ X ) and ME-congruency holds, we have D ( ρ X | | ρ Τ X , θ X * ) = D ( ρ X | | ρ Τ X 1 , θ X 1 * ) = D ( ρ X | | ρ Τ X 2 , θ X 2 * ) = 0 with similar identities for Y. Therefore, the first inequality of (7a) follow directly from (5). The second one is obtained by difference and application of Lemma 1 to T 1 T 2 (q.e.d.).
  • Proof of Theorem 2: The first equality of (11) comes from Equation (1) as I ( X , Y ) = 2 H g H ρ X Y and from the negentropy definition of Gaussianized variables J ( X , Y ) = H g ( X , Y ) H ρ X Y = 2 H g + I g ( X , Y ) H ρ X Y , where H g ( X , Y ) is the entropy of the Gaussian fit. From the entropy formula of transformed variables we get H ρ X r Y r H ρ X Y = H g ( X r , Y r ) H g ( X , Y ) = log | det A | , leading to the negentropy equality J ( X , Y ) = J ( X r , Y r ) . Finally, the last equation of (11) comes from the equality H ρ X r Y r = H ρ X r + H ρ Y r I ( X r , Y r ) and the definition of negentropy, i.e., J ( X r ) = H g H ρ X r (q.e.d.).

Share and Cite

MDPI and ACS Style

Pires, C.A.L.; Perdigão, R.A.P. Minimum Mutual Information and Non-Gaussianity Through the Maximum Entropy Method: Theory and Properties. Entropy 2012, 14, 1103-1126. https://doi.org/10.3390/e14061103

AMA Style

Pires CAL, Perdigão RAP. Minimum Mutual Information and Non-Gaussianity Through the Maximum Entropy Method: Theory and Properties. Entropy. 2012; 14(6):1103-1126. https://doi.org/10.3390/e14061103

Chicago/Turabian Style

Pires, Carlos A. L., and Rui A. P. Perdigão. 2012. "Minimum Mutual Information and Non-Gaussianity Through the Maximum Entropy Method: Theory and Properties" Entropy 14, no. 6: 1103-1126. https://doi.org/10.3390/e14061103

Article Metrics

Back to TopTop