Next Article in Journal
Twin-Field Quantum Digital Signature with Fully Discrete Phase Randomization
Next Article in Special Issue
Application of Statistical K-Means Algorithm for University Academic Evaluation
Previous Article in Journal
Belavkin–Staszewski Relative Entropy, Conditional Entropy, and Mutual Information
Previous Article in Special Issue
Robust Test Statistics Based on Restricted Minimum Rényi’s Pseudodistance Estimators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Generic Formula and Some Special Cases for the Kullback–Leibler Divergence between Central Multivariate Cauchy Distributions

1
ImhorPhen Unit, UMR INRAe IRHS, Institut Agro Rennes-Angers, Université d’Angers, 42 Rue Georges Morel, 49070 Beaucouzé, France
2
LARIS, UMR INRAe IRHS, Université d’Angers, 62 Avenue Notre Dame du Lac, 49000 Angers, France
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(6), 838; https://doi.org/10.3390/e24060838
Submission received: 22 May 2022 / Revised: 15 June 2022 / Accepted: 15 June 2022 / Published: 17 June 2022
(This article belongs to the Special Issue Information and Divergence Measures)

Abstract

:
This paper introduces a closed-form expression for the Kullback–Leibler divergence (KLD) between two central multivariate Cauchy distributions (MCDs) which have been recently used in different signal and image processing applications where non-Gaussian models are needed. In this overview, the MCDs are surveyed and some new results and properties are derived and discussed for the KLD. In addition, the KLD for MCDs is showed to be written as a function of Lauricella D-hypergeometric series F D ( p ) . Finally, a comparison is made between the Monte Carlo sampling method to approximate the KLD and the numerical value of the closed-form expression of the latter. The approximation of the KLD by Monte Carlo sampling method are shown to converge to its theoretical value when the number of samples goes to the infinity.

1. Introduction

Multivariate Cauchy distribution (MCD) belongs to the elliptical symmetric distributions [1] and is a special case of the multivariate t-distribution [2] and the multivariate stable distribution [3]. MCD has been recently used in several signal and image processing applications for which non-Gaussian models are needed. To name a few of them, in speckle denoizing, color image denoizing, watermarking, speech enhancement, among others. Sahu et al. in [4] presented a denoizing method for speckle noise removal applied to a retinal optical coherence tomography (OCT) image. The method was based on the wavelet transform where the sub-bands coefficients were modeled using a Cauchy distribution. In [5], a dual tree complex wavelet transform (DTCWT)-based despeckling algorithm was proposed for synthetic aperture radar (SAR) images, where the DTCWT coefficients in each subband were modeled with a multivariate Cauchy distribution. In [6], a new color image denoizing method in the contourlet domain was suggested for reducing noise in images corrupted by Gaussian noise where the contourlet subband coefficients were described by the heavy-tailed MCD. Sadreazami et al. in [7] put forward a novel multiplicative watermarking scheme in the contourlet domain where the watermark detector was based on the bivariate Cauchy distribution and designed to capture the across scale dependencies of the contourlet coefficients. Fontaine et al. in [8] proposed a semi-supervised multichannel speech enhancement system where both speech and noise follow the heavy-tailed multi-variate complex Cauchy distribution.
Kullback–Leibler divergence (KLD), also called relative entropy, is one of the most fundamental and important measures in information theory and statistics [9,10]. KLD was first introduced and studied by Kullback and Leibler [11] and Kullback [12] to measure the divergence between two probability mass functions in the case of discrete random variables and between two univariate or multivariate probability density functions in the case of continuous random variables. In the literature, numerous entropy and divergence measures have been suggested for measuring the similarity between probability distributions, such as Rényi [13] divergence, Sharma and Mittal [14] divergence, Bhattacharyya [15,16] divergence and Hellinger divergence measures [17]. Other general divergence families have been also introduced and studied like the ϕ -divergence family of divergence measures defined simultaneously by Csiszár [18] and Ali and Silvey [19] where the KLD measure is a special case, the Bregman family divergence [20], the R-divergences introduced by Burbea and Rao [21,22,23], the statistical f-divergences [24,25] and recently the new family of a generalized divergence called the ( h , ϕ ) -divergence measures introduced and studied in Menéndez et al. [26]. Readers are referred to [10] for details about these divergence family measures.
KLD has a specific interpretation in coding theory [27] and is therefore the most popular and widely used as well. Since information theoretic divergence and KLD in particular are ubiquitous in information sciences [28,29], it is therefore important to establish closed-form expressions of such divergence [30]. An analytical expression of the KLD between two univariate Cauchy distributions was presented in [31,32]. To date, the KLD of MCDs has no known explicit form, and it is in practice either estimated using expensive Monte Carlo stochastic integration or approximated. Monte Carlo sampling can efficiently estimate the KLD provided that a large number of independent and identically distributed samples is provided. Nevertheless, Monte Carlo integration is a too slow process to be useful in many applications. The main contribution of this paper is to derive a closed-form expression for the KLD between two central MCDs in a general case to benchmark future approaches while avoiding approximation using expensive Monte Carlo (MC) estimation techniques. The paper is organized as follows. Section 2 introduces the MCD and the KLD. Section 3 gives some definitions and propositions related to a multiple power series used to compute the closed-form expression of the KLD between two central MCDs. In Section 4 and Section 5, expressions of some expectations related to the KLD are developed by exploiting the propositions presented in the previous section. Section 6 demonstrates some final results on the KLD computed for the central MCD. Section 7 presents some particular results such as the KLD for the univariate and the bivariate Cauchy distribution. Section 8 presents the implementation procedure of the KLD and a comparison with Monte Carlo sampling method. A summary and some conclusions are provided in the final section.

2. Multivariate Cauchy Distribution and Kullback–Leibler Divergence

Let X be a random vector of R p which follows the MCD, characterized by the following probability density function (pdf) given as follows [2]
f X ( x | μ , Σ , p ) = Γ ( 1 + p 2 ) π p 2 Γ ( 1 2 ) 1 | Σ | 1 2 1 [ 1 + ( x μ ) T Σ 1 ( x μ ) ] 1 + p 2 .
This is for any x R p , where p is the dimensionality of the sample space, μ is the location vector, Σ is a symmetric, positive definite ( p × p ) scale matrix and Γ ( . ) is the Gamma function. Let X 1 and X 2 be two random vectors that follow central MCDs with pdfs f X 1 ( x | Σ 1 , p ) = f X 1 ( x | 0 , Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) = f X 2 ( x | 0 , Σ 2 , p ) given by (1). KLD provides an asymmetric measure of the similarity of the two pdfs. Indeed, the KLD between the two central MCDs is given by
KL ( X 1 | | X 2 ) = R p ln f X 1 ( x | Σ 1 , p ) f X 2 ( x | Σ 2 , p ) f X 1 ( x | Σ 1 , p ) d x
= E X 1 { ln f X 1 ( X ) } E X 1 { ln f X 2 ( X ) } .
Since the KLD is the relative entropy defined as the difference between the cross-entropy and the entropy, we have the following relation:
KL ( X 1 | | X 2 ) = H ( f X 1 , f X 2 ) H ( f X 1 )
where H ( f X 1 , f X 2 ) = E X 1 { ln f X 2 ( X ) } denotes the cross-entropy and H ( f X 1 ) = E X 1 { ln f X 1 ( X ) } the entropy. Therefore, the determination of KLD requires the expression of the entropy and the cross-entropy. It should be noted that the smaller KL ( X 1 | | X 2 ) , the more similar are f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) . The symmetric KL similarity measure between X 1 and X 2 is d KL ( X 1 , X 2 ) = KL ( X 1 | | X 2 ) + KL ( X 2 | | X 1 ) . In order to compute the KLD, we have to derive the analytical expressions of E X 1 { ln f X 1 ( X ) } and E X 1 { ln f X 2 ( X ) } which depend, respectively, on E X 1 { ln [ 1 + X T Σ 1 1 X ] } and E X 1 { ln [ 1 + X T Σ 2 1 X ] } . Consequently, the closed-form expression of the KLD between two zero-mean MCDs is given by
KL ( X 1 | | X 2 ) = 1 2 log | Σ 2 | | Σ 1 | 1 + p 2 E X 1 { ln [ 1 + X T Σ 1 1 X ] } E X 1 { ln [ 1 + X T Σ 2 1 X ] } .
To provide the expression of these two expectations, some tools based on the multiple power series are required. The next section presents some definitions and propositions used for this goal.

3. Definitions and Propositions

This section presents some definitions and exposes some propositions related to the multiple power series used to derive the closed-form expression of the expectation E X 1 { ln [ 1 + X T Σ 1 1 X ] } and E X 1 { ln [ 1 + X T Σ 2 1 X ] } , and as a consequence the KLD between two central MCDs.
Definition 1.
The Humbert series of n variables, denoted Φ 2 ( n ) , is defined for all x i C , i = 1 , , n , by the following multiple power series (Section 1.4 in [33])
Φ 2 ( n ) ( b 1 , , b n ; c ; x 1 , , x n ) = m 1 = 0 . . m n = 0 ( b 1 ) m 1 ( b n ) m n ( c ) i = 1 n m i i = 1 n x i m i m i ! .
The Pochhammer symbol ( q ) i indicates the i-th rising factorial of q, i.e., for an integer i > 0
( q ) i = q ( q + 1 ) ( q + i 1 ) = k = 0 i 1 ( q + k ) = Γ ( q + i ) Γ ( q )

3.1. Integral Representation for Φ 2 ( n )

Proposition 1.
The following integral representation is true for Real { c } > Real { i = 1 n b i } > 0 and Real { b i } > 0 where Real { . } denotes the real part of the complex coefficients
Δ 1 i = 1 n u i c i = 1 n b i 1 i = 1 n u i b i 1 e x i u i d u i = B b 1 , , b n , c i = 1 n b i Φ 2 ( n ) ( b 1 , , b n ; c ; x 1 , , x n )
where Δ = { ( u 1 , , u n ) | 0 u i 1 , i = 1 , , n ; 0 u 1 + + u n 1 } and the multivariate beta function B is the extension of beta function to more than two arguments (called also Dirichlet function) defined as (Section 1.6.1 in [34])
B ( b 1 , , b n , b n + 1 ) = i = 1 n + 1 Γ ( b i ) Γ ( i = 1 n + 1 b i ) .
Proof. 
The power series of exponential function is given by
e x i u i = m i = 0 x i m i m i ! u i m i .
By substituting the expression of the exponential into the multiple integrals we have
. . Δ 1 i = 1 n u i c i = 1 n b i 1 i = 1 n u i b i 1 e x i u i d u i = . . Δ 1 i = 1 n u i c i = 1 n b i 1 i = 1 n m i = 0 x i m i m i ! u i m i + b i 1 d u i = m 1 = 0 . . m n = 0 i = 1 n x i m i m i ! × I D
where the multivariate integral I D , which is a generalization of a beta integral, is the type-1 Dirichlet integral (Section 1.6.1 in [34]) given by
I D = Δ 1 i = 1 n u i c i = 1 n b i 1 i = 1 n u i m i + b i 1 d u i = i = 1 n Γ ( b i + m i ) Γ ( c i = 1 n b i ) Γ ( c + i = 1 n m i ) .
Knowing that Γ ( b i + m i ) = Γ ( b i ) ( b i ) m i , the expression of I D can be written otherwise
I D = i = 1 n Γ ( b i ) Γ ( c i = 1 n b i ) Γ ( c ) i = 1 n ( b i ) m i ( c ) i = 1 n m i .
Finally, plugging (13) back into (11) leads to the final result
Γ ( c i = 1 n b i ) i = 1 n Γ ( b i ) Γ ( c ) m 1 , , m n = 0 + i = 1 n ( b i ) m i ( c ) i = 1 n m i i = 1 n x i m i m i ! = B b 1 , , b n , c i = 1 n b i Φ 2 ( n ) ( b 1 , , b n ; c ; x 1 , , x n )
Given Proposition 1, we consider the particular cases n = { 1 , 2 } one by one as follows:
Case n = 1
1 B ( b 1 , c b 1 ) 0 1 u 1 b 1 1 e x 1 u 1 ( 1 u 1 ) c b 1 1 d u 1 = m 1 = 0 ( b 1 ) m 1 ( c ) m 1 x 1 m 1 m 1 ! = Φ 2 ( 1 ) ( b 1 ; c ; x 1 ) = 1 F 1 ( b 1 , c ; x 1 )
where 1 F 1 ( . ) is the confluent hypergeometric function of the first kind (Section 9.21 in [35]).
Case n = 2
1 B ( b 1 , b 2 , c b 1 b 2 ) u 1 0 , u 2 0 , u 1 + u 2 1 u 1 b 1 1 u 2 b 2 1 e x 1 u 1 + x 2 u 2 ( 1 u 1 u 2 ) c b 1 b 2 1 d u 1 d u 2 = m 1 = 0 m 2 = 0 ( b 1 ) m 1 ( b 2 ) m 2 ( c ) m 1 + m 2 x 1 m 1 m 1 ! x 2 m 2 m 2 ! = Φ 2 ( 2 ) ( b 1 , b 2 ; c ; x 1 , x 2 ) = Φ 2 ( b 1 , b 2 , c ; x 1 , x 2 )
where the double series Φ 2 is one of the components of the Humbert series of two variables [36] that generalize Kummer’s confluent hypergeometric series 1 F 1 of one variable. The double series Φ 2 converges absolutely at any x 1 , x 2 C .

3.2. Multiple Power Series F N ( n )

Definition 2.
We define a new multiple power series, denoted by F N ( n ) and given by
F N ( n ) ( a ; b 1 , , b n ; c , c n ; x 1 , , x n ) = x n a m 1 , , m n = 0 + ( a ) i = 1 n m i ( a c n + 1 ) i = 1 n m i ( a + b n c n + 1 ) i = 1 n m i i = 1 n 1 ( b i ) m i ( c ) i = 1 n 1 m i i = 1 n 1 x i x n m i 1 m i ! ( 1 x n 1 ) m n m n ! .
The multiple power series (17) is absolutely convergent on the region | x i x n 1 | + | 1 x n 1 | < 1 in C n , i { 1 , , n 1 } .
The multiple power series F N ( n ) ( . ) can also be transformed into two other expressions as follows
F N ( n ) ( a ; b 1 , , b n ; c , c n ; x 1 , , x n ) = m 1 , , m n = 0 + ( a c n + 1 ) i = 1 n 1 m i ( b n ) m n ( a ) i = 1 n m i ( a + b n c n + 1 ) i = 1 n m i i = 1 n 1 ( b i ) m i ( c ) i = 1 n 1 m i i = 1 n 1 x i m i m i ! ( 1 x n ) m n m n ! ,
= x n 1 c n m 1 , , m n = 0 + ( a c n + 1 ) i = 1 n m i ( b n c n + 1 ) m n ( a ) i = 1 n 1 m i ( a + b n c n + 1 ) i = 1 n m i i = 1 n 1 ( b i ) m i ( c ) i = 1 n 1 m i i = 1 n 1 x i m i m i ! ( 1 x n ) m n m n ! .
By Horn’s rule for the determination of the convergence region (see [37], Section 5.7.2), the multiple power series (18) and (19) are absolutely convergent on region | x i | < 1 , i { 1 , , n 1 } , | 1 x n | < 1 in C n .
Equation (18) can then be deduced from (17) by using the following development where the F N ( p ) function can be written as
F N ( n ) ( a ; b 1 , , b n ; c , c n ; x 1 , , x n ) = x n a m 1 , , m n 1 = 0 + ( a ) i = 1 n 1 m i ( a c n + 1 ) i = 1 n 1 m i ( a + b n c n + 1 ) i = 1 n 1 m i i = 1 n 1 ( b i ) m i ( c ) i = 1 n 1 m i × i = 1 n 1 x i x n m i 1 m i ! m n = 0 ( α ) m n ( α c n + 1 ) m n ( α + b n c n + 1 ) m n ( 1 x n 1 ) m n m n !
and α = a + i = 1 n 1 m i is used here to alleviate writing equations. Using the definition of Gauss’ hypergeometric series 2 F 1 ( . ) [34] and the Pfaff transformation [38], we can write
m n = 0 ( α ) m n ( α c n + 1 ) m n ( α + b n c n + 1 ) m n ( 1 x n 1 ) m n m n ! = 2 F 1 α , α c n + 1 ; α + b n c n + 1 ; 1 x n 1
= x n α 2 F 1 α , b n ; α + b n c n + 1 ; 1 x n
= x n α m p = 0 ( α ) m n ( b n ) m n ( α + b n c n + 1 ) m n ( 1 x n ) m n m n ! .
By substituting (23) into (20), and using the following two relations:
( a ) i = 1 n 1 m i ( α ) m n = ( a ) i = 1 n m i ,
( a + b n c n + 1 ) i = 1 n 1 m i ( α + b n c n + 1 ) m n = ( a + b n c n + 1 ) i = 1 n m i
we can get (18).
The second transformation is given as follows
2 F 1 α , α c n + 1 ; b n c n + α + 1 ; 1 x n 1 = x n α c n + 1 2 F 1 b n c n + 1 , α c n + 1 ; α + b n c n + 1 ; 1 x n
= x n α c n + 1 m n = 0 ( α c n + 1 ) m n ( b n c n + 1 ) m n ( α + b n c n + 1 ) m n ( 1 x n ) m n m n ! .
By substituting (27) into (20), we get (19).
Lemma 1.
The multiple power series F N ( n ) is equal to the Lauricella D-hypergeometric function F D ( n ) (see Appendix A) [39] when a c n + 1 = c and is given as follows
F N ( n ) ( a ; b 1 , , b n ; c , c n ; x 1 , , x n ) = m 1 , , m n = 0 + ( a ) i = 1 n m i i = 1 n ( b i ) m i ( a + b n c n + 1 ) i = 1 n m i i = 1 n 1 x i m i m i ! ( 1 x n ) m n m n !
= F D ( n ) ( a , b 1 , , b n ; a + b n c n + 1 ; x 1 , , x n 1 , 1 x n )
Proof. 
By using Equation (18) of the multiple power series F N ( n ) and after having simplified ( a c n + 1 ) i = 1 n 1 m i to the numerator and ( c ) i = 1 n 1 m i to the denominator, we can get the result. □

3.3. Integral Representation for F N ( n + 1 )

Proposition 2.
The following integral representation is true for Real { a } > 0 , Real { a c n + 1 + 1 } > 0 , and Real { a c n + 1 + b n + 1 + 1 } > 0
Γ ( a ) Γ ( a c n + 1 + 1 ) Γ ( a c n + 1 + b n + 1 + 1 ) F N ( n + 1 ) ( a ; b 1 , , b n + 1 ; c , c n + 1 ; x 1 , , x n + 1 ) = 0 e r r a 1 Φ 2 ( n ) ( b 1 , , b n ; c ; r x 1 , , r x n ) U ( b n + 1 , c n + 1 ; r x n + 1 ) d r
where U ( · ) is the confluent hypergeometric function of the second kind (Section 9.21 in [35]) defined for Real { b } > 0 , Real { z } > 0 by the following integral representation
U ( b , c ; z ) = 1 Γ ( b ) 0 e z t t b 1 ( 1 + t ) c b 1 d t
and Φ 2 ( n ) ( · ) is defined by Equation (6).
Proof. 
The multiple power series Φ 2 ( n ) and the confluent hypergeometric function U ( · ) are absolutely convergent on [ 0 , + ] . Using these functions in the above integral and changing the order of integration and summation, which is easily justified by absolute convergence, we get
0 e r r a 1 Φ 2 ( n ) ( b 1 , , b n ; c ; r x 1 , , r x n ) U ( b n + 1 , c n + 1 ; r x n + 1 ) d r = m 1 = 0 . . m n = 0 ( b 1 ) m 1 ( b n ) m n ( c ) i = 1 n m i i = 1 n x i m i m i ! I
where integral I is defined as follows
I = 0 e r r a 1 + i = 1 n m i U ( b n + 1 , c n + 1 ; r x n + 1 ) d r .
Substituting the integral expression of U ( · ) in the previous equation and replacing α = a + i = 1 n m i to alleviate writing equations, we have
I = 1 Γ ( b n + 1 ) 0 0 e ( 1 + x n + 1 t ) r r α 1 t b n + 1 1 ( 1 + t ) ( c n + 1 b n + 1 1 ) d r d t .
Knowing that [35]
0 e ( 1 + x n + 1 t ) r r α 1 d r = Γ ( α ) ( 1 + x n + 1 t ) α
and
0 t b n + 1 1 ( 1 + t ) c n + 1 b n + 1 1 ( 1 + x n + 1 t ) α d t = Γ ( b n + 1 ) Γ ( α c n + 1 + 1 ) Γ ( α + b n + 1 c n + 1 + 1 ) 2 F 1 α , b n + 1 ; α + b n + 1 c n + 1 + 1 ; 1 x n + 1
the new expression of I is then given by
I = Γ ( α ) Γ ( α c n + 1 + 1 ) Γ ( α + b n + 1 c n + 1 + 1 ) m n + 1 = 0 + ( α ) m n + 1 ( b n + 1 ) m n + 1 ( α + b n + 1 c n + 1 + 1 ) m n + 1 ( 1 x n + 1 ) m n + 1 m n + 1 ! .
Using the fact that Γ ( α ) = Γ ( a ) ( a ) i = 1 n m i and ( a ) i = 1 n m i ( α ) m n + 1 = ( a ) i = 1 n + 1 m i , and developing the same method to Γ ( α + b n + 1 c n + 1 + 1 ) , the final complete expression of the integral is then given by
Γ ( a ) Γ ( a c n + 1 + 1 ) Γ ( a + b n + 1 c n + 1 + 1 ) m 1 = 0 . . m n + 1 = 0 ( b 1 ) m 1 ( b n ) m n ( c ) i = 1 n m i ( a c n + 1 + 1 ) i = 1 n m i ( b n + 1 ) m n + 1 ( a ) i = 1 n + 1 m i ( a + b n + 1 c n + 1 + 1 ) i = 1 n + 1 m i i = 1 n x i m i m i ! × ( 1 x n + 1 ) m n + 1 m n + 1 ! = Γ ( a ) Γ ( a c n + 1 + 1 ) Γ ( a c n + 1 + b n + 1 + 1 ) F N ( n + 1 ) ( a ; b 1 , , b n + 1 ; c , c n + 1 ; x 1 , , x n + 1 ) .

4. Expression of E X 1 { ln [ 1 + X T Σ 1 1 X ] }

Proposition 3.
Let X 1 be a random vector that follows a central MCD with pdf given by f X 1 ( x | Σ 1 , p ) . Expectation E X 1 { ln [ 1 + X T Σ 1 1 X ] } is given as follows
E X 1 ln [ 1 + X T Σ 1 1 X ] = ψ 1 + p 2 ψ 1 2
where ψ ( . ) is the digamma function defined as the logarithmic derivative of the Gamma function (Section 8.36 in [35]).
Proof. 
Expectation E X 1 { ln [ 1 + X T Σ 1 1 X ] } is developed as follows
E X 1 { ln [ 1 + X T Σ 1 1 X ] } = A | Σ 1 | 1 2 R p ln [ 1 + x T Σ 1 1 x ] [ 1 + x T Σ 1 1 x ] 1 + p 2 d x
where A = Γ ( 1 + p 2 ) π 1 + p 2 . Utilizing the following property log ( x ) f ( x ) d x = a x a f ( x )   d x | a = 0 , as a consequence the expectation is given as follows
E X 1 { ln [ 1 + X T Σ 1 1 X ] } = A | Σ 1 | 1 2 a R p [ 1 + x T Σ 1 1 x ] a 1 + p 2 d x | a = 0
Consider the transformation y = Σ 1 1 / 2 x where y = [ y 1 , y 2 , , y p ] T . The Jacobian determinant is given by d y = | Σ 1 | 1 / 2 d x (Theorem 1.12 in [40]). The new expression of the expectation is given by
E X 1 { ln [ 1 + X T Σ 1 1 X ] } = A a R p [ 1 + y T y ] a 1 + p 2 d y | a = 0 .
Let u = y T y be a transformation where the Jacobian determinant is given by (Lemma 13.3.1 in [41])
d y = π p 2 Γ ( p 2 ) u p 2 1 d u .
The new expectation is as follows
E X 1 { ln [ 1 + X T Σ 1 1 X ] } = Γ ( 1 + p 2 ) π 1 / 2 Γ ( p 2 ) a 0 + u p 2 1 ( 1 + u ) a 1 + p 2 d u | a = 0
Using the definition of beta function, we can write that
0 + u p 2 1 ( 1 + u ) a 1 + p 2 d u = Γ ( p 2 ) Γ ( 1 2 a ) Γ ( 1 + p 2 a ) .
The derivative of the last integral w.r.t a is given by
a 0 + u p 2 1 ( 1 + u ) a 1 + p 2 d u | a = 0 = Γ ( p 2 ) Γ ( 1 2 ) Γ ( 1 + p 2 ) ψ ( 1 + p 2 ) ψ ( 1 2 )
Finally, the expression of E X 1 { ln [ 1 + X T Σ 1 1 X ] } is given by
E X 1 { ln [ 1 + X T Σ 1 1 X ] } = ψ 1 + p 2 ψ 1 2 .

5. Expression of E X 1 { ln [ 1 + X T Σ 2 1 X ] }

Proposition 4.
Let X 1 and X 2 be two random vectors that follow central MCDs with pdfs given, respectively, by f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) . Expectation E X 1 { ln [ 1 + X T Σ 2 1 X ] } is given as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = ψ 1 + p 2 ψ 1 2 + ln λ p a F D ( p ) a , 1 2 , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 .
where λ 1 ,…, λ p are the eigenvalues of the real matrix Σ 1 Σ 2 1 , and F D ( p ) ( . ) represents the Lauricella D-hypergeometric function defined for p variables.
Proof. 
To prove Proposition 4, different steps are necessary. They are described in the following:

5.1. First Step: Eigenvalue Expression

Expectation E X 1 { ln [ 1 + X T Σ 2 1 X ] } is computed as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = A | Σ 1 | 1 2 R p ln [ 1 + x T Σ 2 1 x ] [ 1 + x T Σ 1 1 x ] 1 + p 2 d x
where A = Γ ( 1 + p 2 ) π 1 + p 2 . Consider transformation y = Σ 1 1 / 2 x where y = [ y 1 , y 2 , , y p ] T . The Jacobian determinant is given by d y = | Σ 1 | 1 / 2 d x (Theorem 1.12 in [40]) and matrix Σ = Σ 1 1 2 Σ 2 1 Σ 1 1 2 is a real symmetric matrix since Σ 1 and Σ 2 are real symmetric matrixes. Then, the expectation is evaluated as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = A R p ln [ 1 + y T Σ y ] [ 1 + y T y ] 1 + p 2 d y .
Matrix Σ can be diagonalized by an orthogonal matrix P with P 1 = P T and Σ = P D P 1 where D is a diagonal matrix composed of the eigenvalues of Σ . Considering that y T Σ y = tr ( Σ y y T ) = tr ( P D P T y y T ) = tr ( D P T y y T P ) , the expectation can be written as
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = A R p ln [ 1 + tr ( D P T y y T P ) ] [ 1 + y T y ] 1 + p 2 d y .
Let z = P T y with z = [ z 1 , z 2 , , z p ] T be a transformation where the Jacobian determinant is given by d z = | P T | d y = d y . Using the fact that tr ( D P T y y T P ) = tr ( D z z T ) = z T D z and y T y = z T P T P z = z T z , then the previous expectation (51) is given as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = A R p ln [ 1 + z T D z ] [ 1 + z T z ] 1 + p 2 d z
= A R . . R ln [ 1 + i = 1 p λ i z i 2 ] [ 1 + i = 1 p z i 2 ] 1 + p 2 d z 1 d z p
where λ 1 ,…, λ p are the eigenvalues of Σ 1 Σ 2 1 .

5.2. Second Step: Polar Decomposition

Let the independent real variables z 1 , , z p be transformed to the general polar coordinates r, θ 1 , , θ p 1 as follows, where r > 0 , π / 2 < θ j π / 2 , j = 1 , , p 2 , π < θ p 1 π [40],
z 1 = r sin θ 1
z 2 = r cos θ 1 sin θ 2
z j = r cos θ 1 cos θ 2 cos θ j 1 sin θ j , j = 2 , 3 , , p 1
z p = r cos θ 1 cos θ 2 cos θ p 1 .
The Jacobian determinant according to theorem (1.24) in [40] is
d z 1 d z p = r p 1 j = 1 p 1 | cos θ j | p j 1 d r d θ j .
It is clear that with the last transformations, we get i = 1 p z i 2 = r 2 and the multiple integral in (53) is then given as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = A 0 + r p 1 [ 1 + r 2 ] 1 + p 2 π / 2 π / 2 . . π π j = 1 p 1 | cos θ j | p j 1 × ln 1 + r 2 ( λ 1 sin 2 θ 1 + + λ p cos 2 θ 1 cos 2 θ p 1 ) d r j = 1 p 1 d θ j .
By replacing the expression of sin 2 θ j by 1 cos 2 θ j , for j = 1 , , p 1 , we have the following expression
λ 1 sin 2 θ 1 + + λ p cos 2 θ 1 cos 2 θ p 1 = λ 1 + ( λ 2 λ 1 ) cos 2 θ 1 + + ( λ p λ p 1 ) cos 2 θ 1 cos 2 θ 2 cos 2 θ p 1 .
Let x i = cos 2 θ i be a transformation to use where d x i = 2 x i 1 / 2 ( 1 x i ) 1 / 2 d θ i . Then the expectation given by the multiple integral over all θ j , j = 1 , , p 1 is as follows
2 A 0 + r p 1 [ 1 + r 2 ] 1 + p 2 0 1 0 1 j = 1 p 1 x j p j 2 1 ( 1 x j ) 1 2 ln [ 1 + r 2 B p ( x 1 , , x p 1 ) ] d r d x 1 d x p 1
where B p ( x 1 , , x p 1 ) = λ 1 + ( λ 2 λ 1 ) x 1 + + ( λ p λ p 1 ) x 1 x 2 x p 1 , p 1 and B 1 = λ 1 . In the following, we use the notation B p instead of B p ( x 1 , , x p 1 ) to alleviate writing equations.
Let t = r 2 be transformation to use. Then, one can write
= A 0 + t p 2 1 [ 1 + t ] 1 + p 2 0 1 0 1 j = 1 p 1 x j p j 2 1 ( 1 x j ) 1 2 ln [ 1 + t B p ] d t d x 1 d x p 1 .
In order to solve the integral in (62), we consider the following property given by log ( x ) f ( x )   d x = a x a f ( x ) d x | a = 0 and the following equation given as follows
1 + B p t a = 1 Γ ( a ) 0 + y a 1 e ( 1 + B p t ) y d y .
Making use of the above equation, we obtain a new expression of (62) given as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = a A Γ ( a ) 0 + t p 2 1 [ 1 + t ] 1 + p 2 0 + y a 1 e ( 1 + B p t ) y 0 1 0 1 j = 1 p 1 x j p j 2 1 ( 1 x j ) 1 2 d x j d y d t | a = 0
= a A Γ ( a ) 0 + t p 2 1 [ 1 + t ] 1 + p 2 0 + y a 1 e y H ( t , y ) d y d t | a = 0
where H ( t , y ) is defined as
H ( t , y ) = 0 1 0 1 e B p t y j = 1 p 1 x j p j 2 1 ( 1 x j ) 1 2 d x j .

5.3. Third Step: Expression for H(t,y) by Humbert and Beta Functions

Let x i = 1 x i , i = 1 , , p 1 be transformations to use. Then
( λ 2 λ 1 ) x 1 = ( λ 2 λ 1 ) ( 1 x 1 )
( λ 3 λ 2 ) x 1 x 2 = ( λ 3 λ 2 ) ( 1 x 1 ) [ 1 x 2 ]
( λ 4 λ 3 ) x 1 x 2 x 3 = ( λ 4 λ 3 ) ( 1 x 1 ) ( 1 x 2 ) [ 1 x 3 ] =
( λ p λ p 1 ) i = 1 p 1 x i = ( λ p λ p 1 ) i = 1 p 1 ( 1 x i ) .
Adding equations from (67) to (70), we can state that the new expression of the function B p becomes
B p = λ p ( λ p λ 1 ) x 1 ( λ p λ 2 ) ( 1 x 1 ) x 2 ( λ p λ 3 ) ( 1 x 1 ) ( 1 x 2 ) x 3 ( λ p λ p 1 ) ( 1 x 1 ) ( 1 x p 2 ) x p 1 .
Then, the multiple integral H ( t , y ) given by (66) can be written otherwise
H ( t , y ) = 0 1 0 1 e B p t y j = 1 p 1 ( 1 x j ) p j 2 1 x j 1 2 d x 1 d x p 1 .
Let the real variables x 1 , x 2 , , x p 1 be transformed to the real variables u 1 , u 2 , , u p 1 as follows
u 1 = x 1
u 2 = ( 1 x 1 ) x 2 = ( 1 u 1 ) x 2
u 3 = ( 1 x 1 ) ( 1 x 2 ) x 3 = ( 1 u 1 u 2 ) x 3
u p 1 = i = 1 p 2 ( 1 x i ) x p 1 = ( 1 i = 1 p 2 u i ) x p 1 .
The Jacobian determinant is given by
d u 1 d u p 1 = j = 1 p 1 1 i = 1 j 1 u i d x 1 d x p 1 .
Accordingly, the new expression of B p becomes
B p = λ p i = 1 p 1 ( λ p λ i ) u i .
As a consequence, the new domain of the multiple integral (72) is Δ = { ( u 1 , u 2 , , u p 1 ) R p 1 ; 0 u 1 1 , 0 u 2 1 u 1 , 0 u 3 1 u 1 u 2 , , and 0 u p 1 1 u 1 u 2 u p 2 } , and the expression of H ( t , y ) is given as follows
H ( t , y ) = Δ e B p t y j = 1 p 1 1 i = 1 j 1 u i 1 u j 1 i = 1 j 1 u i 1 2 j = 1 p 1 1 u j 1 i = 1 j 1 u i p j 2 1 d u j
= Δ e B p t y j = 1 p 1 u j 1 2 1 i = 1 j u i p j 2 1 1 i = 1 j 1 u i 1 2 p j 2 d u 1 d u p 1
= Δ e B p t y 1 i = 1 p 1 u i p 2 p 1 2 1 j = 1 p 1 u j 1 2 d u j
= e λ p t y Δ 1 i = 1 p 1 u i 1 2 i = 1 p 1 u i 1 2 e ( λ p λ i ) u i t y d u i .
Using Proposition 1, we subsequently find that
H ( t , y ) = e λ p t y B 1 2 , , 1 2 p Φ 2 ( p 1 ) 1 2 , , 1 2 p 1 ; p 2 ; ( λ p λ 1 ) t y , ( λ p λ 2 ) t y , , ( λ p λ p 1 ) t y .
where Φ 2 ( p 1 ) ( . ) is the Humbert series of p 1 variables and B ( 1 2 , , 1 2 ) is the multivariate beta function. Applying the following successive two transformations r = t y ( d r = t d y ) and u = 1 / t ( d u = u 2 d t ), the new expression of the expectation given by (65) is written as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = a { A Γ ( a ) B 1 2 , , 1 2 p 0 + r a 1 e λ p r × Φ 2 ( p 1 ) 1 2 , , 1 2 p 1 ; p 2 ; ( λ p λ 1 ) r , , ( λ p λ p 1 ) r 0 + u a 1 2 ( 1 + u ) 1 + p 2 e r u d u d r } | a = 0 .

5.4. Final Step

The last integral is related to the confluent hypergeometric function of the second kind U ( . ) as follows
0 + u a 1 2 ( 1 + u ) 1 + p 2 e r u d u = Γ ( a + 1 2 ) U ( a + 1 2 , a + 1 p 2 , r ) .
As a consequence, the new expression is
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = a { A Γ ( a + 1 2 ) Γ ( a ) B 1 2 , , 1 2 × 0 + r a 1 e λ p r Φ 2 ( p 1 ) 1 2 , , 1 2 ; p 2 ; 1 ; ( λ p λ 1 ) r , , ( λ p λ p 1 ) r U ( a + 1 2 , a + 1 p 2 , r ) d r } | a = 0 .
Using the transformation r = λ p r and the Proposition 2, and taking into account the expression of A, the new expression becomes
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = a { B ( a + 1 2 , p 2 ) B ( 1 2 , p 2 ) λ p a × F N ( p ) a ; 1 2 , , 1 2 , a + 1 2 p ; p 2 , a p 2 + 1 ; 1 λ 1 λ p , , 1 λ p 1 λ p , λ p 1 } | a = 0
Knowing that
a B ( p 2 , a + 1 2 ) B ( p 2 , 1 2 ) | a = 0 = ψ 1 2 ψ 1 + p 2 , and
F N ( p ) a ; 1 2 , , 1 2 , a + 1 2 ; p 2 , a p 2 + 1 ; 1 λ 1 λ p , , 1 λ p 1 λ p , λ p 1 | a = 0 = 1 ,
the new expression of E X 1 { ln [ 1 + X T Σ 2 1 X ] } becomes
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = ψ 1 + p 2 ψ 1 2 a λ p a F N ( p ) a ; 1 2 , , 1 2 , a + 1 2 p ; p 2 , a p 2 + 1 ; 1 λ 1 λ p , , 1 λ p 1 λ p , λ p 1 | a = 0 .
Applying the expression given by (18) of Definition 2 and relying on Lemma 1, the final result corresponds to the D-hypergeometric function of Lauricella F D ( p ) ( . ) given by
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = ψ 1 + p 2 ψ 1 2 a λ p a m 1 , , m p = 0 + ( a ) i = 1 p m i ( a + 1 2 ) m p i = 1 p 1 ( 1 2 ) m i ( a + 1 + p 2 ) i = 1 p m i i = 1 p 1 1 λ i λ p m i 1 m i ! ( 1 λ p 1 ) m p m p ! | a = 0
= ψ 1 + p 2 ψ 1 2 a λ p a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 .
The final development of the previous expression is as follows
E X 1 { ln [ 1 + X T Σ 2 1 X ] } = ψ 1 + p 2 ψ 1 2 + ln λ p a F D ( p ) a , 1 2 , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 .
In this section, we presented the exact expression of E X 1 { ln [ 1 + X T Σ 2 1 X ] } . In addition, the multiple power series F D ( p ) which appears to be a special case of F N ( p ) provides many properties and numerous transformations (see Appendix A) that make easier the convergence of the multiple power series. In the next section, we establish the KLD closed-form expression based on the expression of the latter expectation.

6. KLD between Two Central MCDs

Plugging (39) and (93) into (5) yields the closed-form expression of the KLD between two central MCDs with pdfs f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) . This result is presented in the following theorem.
Theorem 1.
Let X 1 and X 2 be two random vectors that follow central MCDs with pdfs given, respectively, by f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) . The Kullback–Leibler divergence between central MCDs is
KL ( X 1 | | X 2 ) = 1 2 log i = 1 p λ i + 1 + p 2 [ log λ p a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 ]
where λ 1 ,…, λ p are the eigenvalues of the real matrix Σ 1 Σ 2 1 , and F D ( p ) ( . ) represents the Lauricella D-hypergeometric function defined for p variables.
Lauricella [39] gave several transformation formulas (see Appendix A), whose relations (A5)–(A7), and (A9) are applied to F D ( p ) ( . ) in (94). The results of transformation are as follows
F D ( p ) a , 1 2 , , 1 2 , a + 1 2 ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p = λ p a + p 2 i = 1 p 1 λ i 1 2 F D ( p ) 1 + p 2 , 1 2 , , 1 2 , a + 1 2 ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ p λ p 1 , 1 λ p
= λ 1 λ p a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ 2 λ 1 , 1 1 λ 1
= λ p a F D ( p ) a , 1 2 , , 1 2 ; a + 1 + p 2 ; 1 λ 1 , 1 λ 2 , , 1 λ p
= λ p a i = 1 p λ i 1 2 F D ( p ) 1 + p 2 , 1 2 , , 1 2 ; a + 1 + p 2 ; 1 1 λ 1 , 1 1 λ 2 , , 1 1 λ p .
Considering the above equations, it is easy to provide different expressions of KL ( X 1 | | X 2 ) shown in Table 1. The derivative of the Lauricella D-hypergeometric series with respect to a goes through the derivation of the following expression
a F D ( p ) a , 1 2 , 1 2 , , 1 2 , a + 1 2 ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0
= m 1 , , m p = 0 + a ( a ) i = 1 p m i ( a + 1 2 ) m p ( a + 1 + p 2 ) i = 1 p m i | a = 0 i = 1 p 1 1 2 m i 1 λ i λ p m i 1 m i ! ( 1 λ p 1 ) m p m p !
The derivative with respect to a of the Lauricella D-hypergeometric series and its transformations goes through the following expressions (see Appendix B for demonstration)
a ( a ) i = 1 p m i ( a + 1 2 ) m p ( a + 1 + p 2 ) i = 1 p m i | a = 0 = ( 1 2 ) m p ( 1 ) i = 1 p m i ( 1 + p 2 ) i = 1 p m i ( i = 1 p m i ) ,
a ( a ) i = 1 p m i ( a + 1 + p 2 ) i = 1 p m i | a = 0 = ( 1 ) i = 1 p m i ( 1 + p 2 ) i = 1 p m i ( i = 1 p m i ) ,
a ( a + 1 2 ) m p ( a + 1 + p 2 ) i = 1 p m i | a = 0 = ( 1 2 ) m p ( 1 + p 2 ) i = 1 p m i k = 0 m p 1 1 k + 1 2 k = 0 i = 1 p m i 1 1 k + 1 + p 2 ,
a 1 ( a + 1 + p 2 ) i = 1 p m i | a = 0 = 1 ( 1 + p 2 ) i = 1 p m i k = 0 i = 1 p m i 1 1 k + 1 + p 2 .
To derive the closed-form expression of d KL ( X 1 , X 2 ) we have to evaluate the expression of KL ( X 2 | | X 1 ) . The latter can be easily deduced from KL ( X 1 | | X 2 ) as follows
KL ( X 2 | | X 1 ) = 1 2 log i = 1 p λ i 1 + p 2 [ log λ p + a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ p λ p 1 , 1 λ p | a = 0 ] .
Proceeding in the same way by using Lauricella transformations, different expressions of KL ( X 2 | | X 1 ) are provided in Table 1. Finally, given the above results, it is straightforward to compute the symmetric KL similarity measure d KL ( X 1 , X 2 ) between X 1 and X 2 . Technically, any combination of the KL ( X 1 | | X 2 ) and KL ( X 2 | | X 1 ) expressions is possible to compute d KL ( X 1 , X 2 ) . However, we choose the same convergence region for the two divergences for the calculation of the distance. Some expressions of d KL ( X 1 , X 2 ) are given in Table 1.
Table 1. KLD and KL distance computed when X 1 and X 2 are two random vectors following central MCDs with pdfs f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) .
Table 1. KLD and KL distance computed when X 1 and X 2 are two random vectors following central MCDs with pdfs f X 1 ( x | Σ 1 , p ) and f X 2 ( x | Σ 2 , p ) .
                KL ( X 1 | | X 2 ) (106) = 1 2 log i = 1 p λ i + 1 + p 2 log λ p a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 (107) = 1 2 log i = 1 p λ i 1 + p 2 λ p p 2 i = 1 p 1 λ i 1 2 a F D ( p ) 1 + p 2 , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ p λ p 1 , 1 λ p | a = 0 (108) = 1 2 log i = 1 p λ i + 1 + p 2 log λ 1 a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ 2 λ 1 , 1 1 λ 1 | a = 0 (109) = 1 2 log i = 1 p λ i 1 + p 2 a F D ( p ) a , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 λ 1 , , 1 λ p | a = 0 (110) = 1 2 log i = 1 p λ i 1 + p 2 i = 1 p λ i 1 2 a F D ( p ) 1 + p 2 , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 1 λ 1 , , 1 1 λ p | a = 0
              KL ( X 2 | | X 1 ) (111) = 1 2 log i = 1 p λ i 1 + p 2 log λ p + a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ p λ 1 , , 1 λ p λ p 1 , 1 λ p | a = 0 (112) = 1 2 log i = 1 p λ i 1 + p 2 λ p p 2 i = 1 p 1 λ i 1 2 a F D ( p ) 1 + p 2 , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 (113) = 1 2 log i = 1 p λ i 1 + p 2 log λ 1 + a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ 1 λ 2 , 1 λ 1 | a = 0 (114) = 1 2 log i = 1 p λ i 1 + p 2 a F D ( p ) a , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 1 λ 1 , , 1 1 λ p | a = 0 (115) = 1 2 log i = 1 p λ i 1 + p 2 i = 1 p λ i 1 2 a F D ( p ) 1 + p 2 , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 λ 1 , , 1 λ p | a = 0
d KL ( X 1 , X 2 ) = 1 + p 2 [ log λ p a F D ( p ) a , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 λ p p 2 i = 1 p 1 λ i 1 2 (116) × a F D ( p ) 1 + p 2 , 1 2 , , 1 2 , a + 1 2 p ; a + 1 + p 2 ; 1 λ 1 λ p , , 1 λ p 1 λ p , 1 1 λ p | a = 0 ] = 1 + p 2 [ a F D ( p ) a , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 λ 1 , , 1 λ p | a = 0 + i = 1 p λ i 1 2 a { F D ( p ) ( 1 + p 2 , 1 2 , , 1 2 p ; a + 1 + p 2 ; (117) 1 λ 1 , , 1 λ p ) } | a = 0 ] = 1 + p 2 [ i = 1 p λ i 1 2 a F D ( p ) 1 + p 2 , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 1 λ 1 , , 1 1 λ p | a = 0 + a { F D ( p ) ( a , 1 2 , , 1 2 p ; a + 1 + p 2 ; (118) 1 1 λ 1 , , 1 1 λ p ) } | a = 0 ]

7. Particular Cases: Univariate and Bivariate Cauchy Distribution

7.1. Case of p = 1

This case corresponds to the univariate Cauchy distribution. The KLD is given by
KL ( X 1 | | X 2 ) = 1 2 log λ a 2 F 1 ( a , 1 2 ; a + 1 ; 1 λ ) | a = 0
where 2 F 1 is the Gauss’s hypergeometric function. The expression of the derivative of 2 F 1 is given as follows (see Appendix C.1 for details of computation)
a 2 F 1 ( a , 1 2 ; a + 1 ; 1 λ ) | a = 0 = n = 1 1 2 n 1 n ( 1 λ ) n n ! = 2 ln 1 + λ 1 / 2 2 .
Accordingly, the KLD is then expressed as
KL ( X 1 | | X 2 ) = log ( 1 + λ 1 2 ) 2 4 λ 1 2
= log ( 1 + λ 1 2 ) 2 4 λ 1 2 = KL ( X 2 | | X 1 ) .
We conclude that KLD between Cauchy densities is always symmetric. Interestingly, this is consistent with the result presented in [31].

7.2. Case of p = 2

This case corresponds to the Bivariate Cauchy distribution. The KLD is then given by
KL ( X 1 | | X 2 ) = 1 2 log λ 1 λ 2 3 2 a F 1 ( a , 1 2 , 1 2 ; a + 3 2 ; 1 λ 1 , 1 λ 2 ) | a = 0
where F 1 is the Appell’s hypergeometric function (see Appendix A). The expression of the derivative of F 1 can be further developed
a F 1 ( a , 1 2 , 1 2 ; a + 3 2 ; 1 λ 1 , 1 λ 2 ) | a = 0 = n , m = 0 + ( 1 ) m + n ( 1 2 ) n ( 1 2 ) m ( 3 2 ) m + n 1 m + n ( 1 λ 1 ) n n ! ( 1 λ 2 ) m m ! .
In addition, when the eigenvalue λ i for i = 1 , 2 takes some particular values, the expression of the KLD becomes very simple. In the following, we show some cases:
  • ( λ 1 = 1 , λ 2 1 ) or ( λ 2 = 1 , λ 1 1 )
For this particular case, we have
a F 1 ( a , 1 2 , 1 2 ; a + 3 2 ; 1 λ i , 0 ) | a = 0 = a 2 F 1 ( a , 1 2 ; a + 3 2 ; 1 λ i ) | a = 0
= ln λ i + 1 1 λ i ln 1 1 λ i 1 + 1 λ i + 2 .
The demonstration of the derivation is shown in Appendix C.2. Then, KLD becomes equal to
KL ( X 1 | | X 2 ) = ln λ i 3 2 1 1 λ i ln 1 1 λ i 1 + 1 λ i 3 .
  • λ 1 = λ 2 = λ
For this particular case, we have
a F 1 ( a , 1 2 , 1 2 ; a + 3 2 ; 1 λ , 1 λ ) | a = 0 = a 2 F 1 ( a , 1 ; a + 3 2 ; 1 λ ) | a = 0
= 2 1 λ 1 ln ( λ + λ 1 ) + 2 .
For more details about the demonstration see Appendix C.3. The KLD becomes equal to
KL ( X 1 | | X 2 ) = ln λ + 3 1 λ 1 ln ( λ + λ 1 ) 3 .
It is easy to deduce that
KL ( X 2 | | X 1 ) = ln λ + 3 1 λ ln ( λ 1 + λ 1 1 ) 3 .
This result can be demonstrated using the same process as KL ( X 1 | | X 2 ) . It is worth to notice that KL ( X 1 | | X 2 ) KL ( X 2 | | X 1 ) which leads us to conclude that the property of symmetry observed for the univariate case is no longer valid in the multivariate case. Nielsen et al. in [32] gave the same conclusion.

8. Implementation and Comparison with Monte Carlo Technique

In this section, we show how we practically compute the numerical values of the KLD, especially when we have several equivalent expressions which differ in the region of convergence. To reach this goal, the eigenvalues of Σ 1 Σ 2 1 are rearranged in a descending order λ p > λ p 1 > > λ 1 > 0 . This operation is justified by Equation (53) where it can be seen that the permutation of the eigenvalues does not affect the expectation result. Three cases can be identified from the expressions of KLD.

8.1. Case 1 > λ p > λ p 1 > > λ 1 > 0

The expression of KL ( X 1 | | X 2 ) is given by Equation (109) and KL ( X 2 | | X 1 ) is given by (115).

8.2. Case λ p > λ p 1 > > λ 1 > 1

KL ( X 1 | | X 2 ) is given by the Equation (110) and KL ( X 2 | | X 1 ) is given by (114).

8.3. Case λ p > 1 and λ 1 < 1

This case guarantees that 0 1 λ j / λ p < 1 , j = 1 , , p 1 and 0 1 1 / λ p < 1 . The expression of the KL ( X 1 | | X 2 ) is given by Equation (106) and KL ( X 2 | | X 1 ) is given by (112) or (113). To perform an evaluation of the quality of the numerical approximation of the derivative of the Lauricella series, we consider a case where an exact and simple expression of a { F D ( p ) ( . ) } | a = 0 is possible. The following case where λ 1 = = λ p = λ allows the Lauricella series to be equivalent to the Gauss hypergeometric function given as follows
F D ( p ) a , 1 2 , , 1 2 p ; a + 1 + p 2 ; 1 λ , , 1 λ = 2 F 1 a , p 2 ; a + 1 + p 2 ; 1 λ .
This relation allows us to compare the computational accuracy of the approximation of the Lauricella series with respect to the Gauss function. In addition, to compute the numerical value the indices of the series will evolve from 0 to N instead of infinity. The latter is chosen to ensure a good approximation of the Lauricella series. Table 2 shows the computation of the derivative of F D ( p ) ( . ) and 2 F 1 ( . ) , along with the absolute value of error | ϵ | , where p = 2 , N = { 20 , 30 , 40 } . The exact expression of a { 2 F 1 ( . ) } | a = 0 when p = 2 is given by Equation (129). We can deduce the following conclusions. First, the error is reasonably low and decreases as the value of N increases. Second, the error increases for values of 1 λ close to 1 as expected, which corresponds to the convergence region limit.
In the following section, we compare the Monte Carlo sampling method to approximate the KLD value with the numerical value of the closed-form expression of the latter. The Monte Carlo method involves sampling a large number of samples and using them to calculate the sum rather than the integral. Here, for each sample size, the experiment is repeated 2000 times. The elements of Σ 1 and Σ 2 are given in Table 3. Figure 1 depicts the absolute value of bias, mean square error (MSE) and box plot of the difference between the symmetric KL approximated value and its theoretical one, given versus the sample sizes. As the sample size increases, the bias and the MSE decrease. Accordingly, the approximated value will be very close to the theoretical KLD when the number of samples is very large. The computation time of the proposed approximation and the classical Monte Carlo sampling method are recorded using Matlab on a 1.6 GHz processor with 16 GB of memory. For the proposed numerical approximation, the computation time is evaluated to 1.56 s with N = 20 . The value of N can be increased to further improve the accuracy, but it will increase the computation time. For the Monte Carlo sampling method, the mean time values at sample sizes of {65,536; 131,072; 262,144} are { 2.71 ; 5.46 ; 10.78 } seconds, respectively.
To further encourage the dissemination of these results, we provide a code available as attached file to this paper. This is given in Matlab with a specific case of p = 3 . This can be easily extended to any value of p, thanks to the general closed-form expression established in this paper.
Entropy 24 00838 i001
Entropy 24 00838 i002

9. Conclusions

Since the MCDs have various applications in signal and image processing, the KLD between central MCDs tackles an important problem for future work on statistics, machine learning and other related fields in computer science. In this paper, we derived a closed-form expression of the KLD and distance between two central MCDs. The similarity measure can be expressed as function of the Lauricella D-hypergeometric series F D ( p ) . We have also proposed a simple scheme to compute easily the Lauricella series and to bypass the convergence constraints of this series. Codes and examples for numerical calculations are presented and explained in detail. Finally, a comparison is made to show how the Monte Carlo sampling method gives approximations close to the KLD theoretical value. As a final note, it is also possible to extend these results on the KLD to the case of the multivariate t-distribution since the MCD is a particular case of this multivariate distribution.

Author Contributions

Conceptualization, N.B.; methodology, N.B.; software, N.B.; writing original draft preparation, N.B.; writing review and editing, N.B. and D.R.; supervision, D.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Authors gratefully acknowledge the PHENOTIC platform node of the french national infrastructure on plant phenotyping ANR PHENOME 11-INBS-0012. The authors would like also to thank the anonymous reviewers for their helpful comments valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Lauricella Function

In 1893, G. Lauricella [39] investigated the properties of four series F A ( n ) , F B ( n ) , F C ( n ) , F D ( n ) of n variables. When n = 2 , these functions coincide with Appell’s F 2 , F 3 , F 4 , F 1 , respectively. When n = 1 , they all coincide with Gauss’ 2 F 1 . We present here only the Lauricella series F D ( n ) given as follows
F D ( n ) ( a , b 1 , , b n ; c ; x 1 , , x n ) = m 1 = 0 m n = 0 ( a ) m 1 + + m n ( b 1 ) m 1 ( b n ) m n ( c ) m 1 + + m n x 1 m 1 m 1 ! x n m n m n !
where | x 1 | , , | x n | < 1 . The Pochhammer symbol ( q ) i indicates the i-th rising factorial of q, i.e.,
( q ) i = q ( q + 1 ) ( q + i 1 ) = Γ ( q + i ) Γ ( q ) if i = 1 , 2 ,
If i = 0 , ( q ) i = 1 . Function F D ( n ) ( . ) can be expressed in terms of multiple integrals as follows [42]
F D ( n ) ( a , b 1 , , b n ; c ; x 1 , , x n ) = Γ ( c ) Γ ( c i = 1 n b i ) i = 1 n Γ ( b i ) × Ω i = 1 n u i b i 1 ( 1 i = 1 n u i ) c i = 1 n b i 1 ( 1 i = 1 n x i u i ) a i = 1 n d u i
where Ω = { ( u 1 , u 2 , , u n ) ; 0 u i 1 , i = 1 , , n , and 0 u 1 + u 2 + + u n 1 } , Real ( b i ) > 0 for i = 1 , , n and Real ( c b 1 b n ) > 0 . Lauricella’s F D can be written as a one-dimensional Euler-type integral for any number n of variables. The integral form of F D ( n ) ( . ) is given as follows when Real ( a ) > 0 and Real ( c a ) > 0
F D ( n ) ( a , b 1 , , b n ; c ; x 1 , , x n ) = Γ ( c ) Γ ( a ) Γ ( c a ) 0 1 u a 1 ( 1 u ) c a 1 ( 1 u x 1 ) b 1 ( 1 u x n ) b n d u .
Lauricella has given several transformation formulas, from which we use the two following relationships. More details can be found in Exton’s book [43] on hypergeometric equations.
F D ( n ) ( a , b 1 , , b n ; c ; x 1 , , x n )
= i = 1 n ( 1 x i ) b i F D ( n ) c a , b 1 , , b n ; c ; x 1 x 1 1 , , x n x n 1
= ( 1 x 1 ) a F D ( n ) a , c i = 1 n b i , b 2 , , b n ; c ; x 1 x 1 1 , x 1 x 2 x 1 1 , , x 1 x n x 1 1
= ( 1 x n ) a F D ( n ) a , b 1 , , b n 1 , c i = 1 n b i ; c ; x n x 1 x n 1 , x n x 2 x n 1 , , x n x n 1 x n 1 , x n x n 1
= ( 1 x 1 ) c a i = 1 n ( 1 x i ) b i F D ( n ) c a , c i = 1 n b i , b 2 , , b n ; c ; x 1 , x 2 x 1 x 2 1 , , x n x 1 x n 1
= ( 1 x n ) c a i = 1 n ( 1 x i ) b i F D ( n ) c a , b 1 , , b n 1 , c i = 1 n b i ; c ; x 1 x n x 1 1 , , x n 1 x n x n 1 1 , x n .

Appendix B. Demonstration of Derivative

Appendix B.1. Demonstration

We use the following notation α = i = 1 p m i to alleviate the writing of equations. Knowing that c ( c ) k = ( c ) k ψ ( c + k ) ψ ( c ) , ψ ( c + k ) ψ ( c ) = = 0 k 1 1 c + and ( c ) k = i = 0 k 1 ( c + i ) we can state that
a ( a ) α ( a + 1 + p 2 ) α = ( a ) α [ ψ ( a + α ) ψ ( a ) ψ ( a + 1 + p 2 + α ) + ψ ( a + 1 + p 2 ) ] ( a + 1 + p 2 ) α = k = 0 α 1 ( a + k ) k = 0 α 1 1 a + k 1 a + 1 + p 2 + k ( a + 1 + p 2 ) α .
Using the fact that
k = 0 α 1 ( a + k ) k = 0 α 1 1 a + k = k = 1 α 1 ( a + k ) + k = 0 , k 1 α 1 ( a + k ) + + k = 0 α 2 ( a + k )
we can state that
a ( a ) α ( a + 1 + p 2 ) α | a = 0 = ( α 1 ) ! ( 1 + p 2 ) α = ( 1 ) α ( 1 + p 2 ) α 1 α .

Appendix B.2. Demonstration

a ( a ) α ( a + 1 2 ) m p ( a + 1 + p 2 ) α = ( a + 1 2 ) m p ( a ) α [ ψ ( a + α ) ψ ( a ) + ψ ( a + 1 2 + m p ) ψ ( a + 1 2 ) ] ( a + 1 + p 2 ) α ( a ) α ( a + 1 2 ) m p [ ψ ( a + 1 + p 2 + α ) ψ ( a + 1 + p 2 ) ] ( a + 1 + p 2 ) α
= ( a + 1 2 ) m p k = 0 α 1 ( a + k ) k = 0 α 1 1 a + k 1 a + 1 + p 2 + k + k = 0 m p 1 1 a + 1 2 + k ( a + 1 + p 2 ) α .
By developing the previous expression we can state that
a ( a ) α ( a + 1 2 ) m p ( a + 1 + p 2 ) α | a = 0 = ( 1 2 ) m p ( α 1 ) ! ( 1 + p 2 ) α = ( 1 2 ) m p ( 1 ) α ( 1 + p 2 ) α 1 α .

Appendix B.3. Demonstration

a ( a + 1 2 ) m p ( a + 1 + p 2 ) α = ( a + 1 2 ) m p ( a + 1 + p 2 ) α k = 0 m p 1 1 a + 1 2 + k k = 0 α 1 1 a + 1 + p 2 + k .
As a consequence,
a ( a + 1 2 ) m p ( a + 1 + p 2 ) α | a = 0 = ( 1 2 ) m p ( 1 + p 2 ) α k = 0 m p 1 1 1 2 + k k = 0 α 1 1 1 + p 2 + k .

Appendix B.4. Demonstration

a 1 ( a + 1 + p 2 ) α = ψ ( a + 1 + p 2 + α ) ψ ( a + 1 + p 2 ) ( a + 1 + p 2 ) α
= 1 ( a + 1 + p 2 ) α k = 0 α 1 1 a + 1 + p 2 + k .
Finally,
a 1 ( a + 1 + p 2 ) α | a = 0 = 1 ( 1 + p 2 ) α k = 0 α 1 1 1 + p 2 + k .

Appendix C. Computations of Some Equations

Appendix C.1. Computation

Let f be a function of λ defined as follows:
f ( λ ) = n = 1 1 2 n 1 n ( 1 λ ) n n ! .
The multiplication of the derivative of f with respect to λ by ( 1 λ ) is given as follows
( 1 λ ) λ f ( λ ) = n = 1 1 2 n ( 1 λ ) n n ! = 1 λ 1 / 2 .
As a consequence,
λ f ( λ ) = 1 λ 1 / 2 1 λ = λ 1 / 2 1 + λ 1 / 2 .
Finally,
f ( λ ) = 2 ln 1 + λ 1 / 2 2 .

Appendix C.2. Computation

a 2 F 1 a , 1 2 ; a + 3 2 ; 1 λ i | a = 0 = n = 1 ( 1 2 ) n ( 1 ) n ( 3 2 ) n n ( 1 λ i ) n n ! = f ( λ i )
where f is a function of λ i . The multiplication of the derivative of f with respect to λ i by ( 1 λ i ) is given as follows
( 1 λ i ) λ i f ( λ i ) = n = 1 ( 1 2 ) n ( 1 ) n ( 3 2 ) n ( 1 λ i ) n n !
= 2 F 1 1 2 , 1 ; 3 2 ; 1 λ i + 1 .
Knowing that
2 F 1 1 2 , 1 ; 3 2 ; 1 λ i = arctan ( λ i 1 ) λ i 1
= 1 2 1 λ i ln 1 + 1 λ i 1 1 λ i
we can deduce an expression of
λ i f ( λ i ) = arctan ( λ i 1 ) ( λ i 1 ) 3 / 2 + 1 1 λ i .
Accordingly,
f ( λ i ) = ln λ i 2 arctan ( λ i 1 ) λ i 1 + 2
= ln λ i + 1 1 λ i ln 1 1 λ i 1 + 1 λ i + 2 .

Appendix C.3. Computation

a 2 F 1 a , 1 ; a + 3 2 ; 1 λ | a = 0 = n = 1 ( 1 ) n ( 1 ) n ( 3 2 ) n n ( 1 λ ) n n ! = f ( λ )
where f is a function of λ . The multiplication of the derivative of f with respect to λ by ( 1 λ ) is given as follows
( 1 λ ) λ f ( λ ) = n = 1 ( 1 ) n ( 1 ) n ( 3 2 ) n ( 1 λ ) n n !
= 2 F 1 1 , 1 ; 3 2 ; 1 λ + 1 .
Knowing that
2 F 1 1 , 1 ; 3 2 ; 1 λ = 1 λ arcsin ( 1 λ ) 1 λ
we can state that
λ f ( λ ) = 1 λ arcsin ( 1 λ ) ( 1 λ ) 3 / 2 + 1 1 λ .
As a consequence,
f ( λ ) = 2 λ arcsin ( 1 λ ) 1 λ + 2
= 2 1 λ 1 ln ( λ + λ 1 ) + 2 .

References

  1. Ollila, E.; Tyler, D.E.; Koivunen, V.; Poor, H.V. Complex Elliptically Symmetric Distributions: Survey, New Results and Applications. IEEE Trans. Signal Process. 2012, 60, 5597–5625. [Google Scholar] [CrossRef]
  2. Kotz, S.; Nadarajah, S. Multivariate T-Distributions and Their Applications; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef]
  3. Press, S. Multivariate stable distributions. J. Multivar. Anal. 1972, 2, 444–462. [Google Scholar] [CrossRef] [Green Version]
  4. Sahu, S.; Singh, H.V.; Kumar, B.; Singh, A.K. Statistical modeling and Gaussianization procedure based de-speckling algorithm for retinal OCT images. J. Ambient. Intell. Humaniz. Comput. 2018, 1–14. [Google Scholar] [CrossRef]
  5. Ranjani, J.J.; Thiruvengadam, S.J. Generalized SAR Despeckling Based on DTCWT Exploiting Interscale and Intrascale Dependences. IEEE Geosci. Remote Sens. Lett. 2011, 8, 552–556. [Google Scholar] [CrossRef]
  6. Sadreazami, H.; Ahmad, M.O.; Swamy, M.N.S. Color image denoising using multivariate cauchy PDF in the contourlet domain. In Proceedings of the 2016 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE), Vancouver, BC, Canada, 15–18 May 2016; pp. 1–4. [Google Scholar] [CrossRef]
  7. Sadreazami, H.; Ahmad, M.O.; Swamy, M.N.S. A Study of Multiplicative Watermark Detection in the Contourlet Domain Using Alpha-Stable Distributions. IEEE Trans. Image Process. 2014, 23, 4348–4360. [Google Scholar] [CrossRef]
  8. Fontaine, M.; Nugraha, A.A.; Badeau, R.; Yoshii, K.; Liutkus, A. Cauchy Multichannel Speech Enhancement with a Deep Speech Prior. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  9. Cover, T.M.; Thomas, J.A. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing); Wiley-Interscience: Hoboken, NJ, USA, 2006. [Google Scholar] [CrossRef]
  10. Pardo, L. Statistical Inference Based on Divergence Measures; CRC Press: Abingdon, UK, 2005. [Google Scholar]
  11. Kullback, S.; Leibler, R.A. On Information and Sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  12. Kullback, S. Information Theory and Statistics; Wiley: New York, NY, USA, 1959. [Google Scholar]
  13. Rényi, A. On Measures of Entropy and Information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1961; Volume 1, pp. 547–561. [Google Scholar]
  14. Sharma, B.D.; Mittal, D.P. New non-additive measures of relative information. J. Comb. Inf. Syst. Sci. 1977, 2, 122–132. [Google Scholar]
  15. Bhattacharyya, A. On a measure of divergence between two statistical populations defined by their probability distributions. Bull. Calcutta Math. Soc. 1943, 35, 99–109. [Google Scholar]
  16. Kailath, T. The Divergence and Bhattacharyya Distance Measures in Signal Selection. IEEE Trans. Commun. Technol. 1967, 15, 52–60. [Google Scholar] [CrossRef]
  17. Giet, L.; Lubrano, M. A minimum Hellinger distance estimator for stochastic differential equations: An application to statistical inference for continuous time interest rate models. Comput. Stat. Data Anal. 2008, 52, 2945–2965. [Google Scholar] [CrossRef]
  18. Csiszár, I. Eine informationstheoretische Ungleichung und ihre Anwendung auf den Beweis der Ergodizität von Markoffschen Ketten. Publ. Math. Inst. Hung. Acad. Sci. Ser. A 1963, 8, 85–108. [Google Scholar]
  19. Ali, S.M.; Silvey, S.D. A General Class of Coefficients of Divergence of One Distribution from Another. J. R. Stat. Soc. Ser. B (Methodol.) 1966, 28, 131–142. [Google Scholar] [CrossRef]
  20. Bregman, L. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming. USSR Comput. Math. Math. Phys. 1967, 7, 200–217. [Google Scholar] [CrossRef]
  21. Burbea, J.; Rao, C. On the convexity of some divergence measures based on entropy functions. IEEE Trans. Inf. Theory 1982, 28, 489–495. [Google Scholar] [CrossRef]
  22. Burbea, J.; Rao, C. On the convexity of higher order Jensen differences based on entropy functions (Corresp.). IEEE Trans. Inf. Theory 1982, 28, 961–963. [Google Scholar] [CrossRef]
  23. Burbea, J.; Rao, C. Entropy differential metric, distance and divergence measures in probability spaces: A unified approach. J. Multivar. Anal. 1982, 12, 575–596. [Google Scholar] [CrossRef] [Green Version]
  24. Csiszar, I. Information-type measures of difference of probability distributions and indirect observation. Stud. Sci. Math. Hung. 1967, 2, 229–318. [Google Scholar]
  25. Nielsen, F.; Nock, R. On the chi square and higher-order chi distances for approximating f-divergences. IEEE Signal Process. Lett. 2014, 21, 10–13. [Google Scholar] [CrossRef] [Green Version]
  26. Menéndez, M.L.; Morales, D.; Pardo, L.; Salicrú, M. Asymptotic behaviour and statistical applications of divergence measures in multinomial populations: A unified study. Stat. Pap. 1995, 36, 1–29. [Google Scholar] [CrossRef]
  27. Cover, T.M.; Thomas, J.A. Information theory and statistics. Elem. Inf. Theory 1991, 1, 279–335. [Google Scholar]
  28. MacKay, D.J.C. Information Theory, Inference and Learning Algorithms; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  29. Ruiz, F.E.; Pérez, P.S.; Bonev, B.I. Information Theory in Computer Vision and Pattern Recognition; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  30. Nielsen, F. Statistical Divergences between Densities of Truncated Exponential Families with Nested Supports: Duo Bregman and Duo Jensen Divergences. Entropy 2022, 24, 421. [Google Scholar] [CrossRef] [PubMed]
  31. Chyzak, F.; Nielsen, F. A closed-form formula for the Kullback–Leibler divergence between Cauchy distributions. arXiv 2019, arXiv:1905.10965. [Google Scholar]
  32. Nielsen, F.; Okamura, K. On f-divergences between Cauchy distributions. arXiv 2021, arXiv:2101.12459. [Google Scholar]
  33. Srivastava, H.; Karlsson, P.W. Multiple Gaussian Hypergeometric Series; Ellis Horwood Series in Mathematics and Its Applications Statistics and Operational Research, E; Horwood Halsted Press: Chichester, UK; West Sussex, UK; New York, NY, USA, 1985. [Google Scholar]
  34. Mathai, A.M.; Haubold, H.J. Special Functions for Applied Scientists; Springer Science+Business Media: New York, NY, USA, 2008. [Google Scholar]
  35. Gradshteyn, I.; Ryzhik, I. Table of Integrals, Series, and Products, 7th ed.; Academic Press is an Imprint of Elsevier: Cambridge, MA, USA, 2007. [Google Scholar]
  36. Humbert, P. The Confluent Hypergeometric Functions of Two Variables. Proc. R. Soc. Edinb. 1922, 41, 73–96. [Google Scholar] [CrossRef] [Green Version]
  37. Erdélyi, A. Higher Transcendental Functions; McGraw-Hill: New York, NY, USA, 1953; Volume I. [Google Scholar]
  38. Koepf, W. Hypergeometric Summation an Algorithmic Approach to Summation and Special Function Identities, 2nd ed.; Universitext, Springer: London, UK, 2014. [Google Scholar]
  39. Lauricella, G. Sulle funzioni ipergeometriche a piu variabili. Rend. Del Circ. Mat. Palermo 1893, 7, 111–158. [Google Scholar] [CrossRef]
  40. Mathai, A.M. Jacobians of Matrix Transformations and Functions of Matrix Argument; World Scientific: Singapore, 1997. [Google Scholar]
  41. Anderson, T.W. An Introduction to Multivariate Statistical Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  42. Hattori, A.; Kimura, T. On the Euler integral representations of hypergeometric functions in several variables. J. Math. Soc. Jpn. 1974, 26, 1–16. [Google Scholar] [CrossRef]
  43. Exton, H. Multiple Hypergeometric Functions and Applications; Wiley: New York, NY, USA, 1976. [Google Scholar]
Figure 1. Top row: Bias (left) and MSE (right) of the difference between the approximated and theoretical symmetric KL for MCD. Bottom row: Box plot of the error. The mean error is the bias. Outliers are larger than Q 3 + 1.5 × I Q R or smaller than Q 1 1.5 × I Q R , where Q 1 , Q 3 , and I Q R are the 25th, 75th percentiles, and the interquartile range, respectively.
Figure 1. Top row: Bias (left) and MSE (right) of the difference between the approximated and theoretical symmetric KL for MCD. Bottom row: Box plot of the error. The mean error is the bias. Outliers are larger than Q 3 + 1.5 × I Q R or smaller than Q 1 1.5 × I Q R , where Q 1 , Q 3 , and I Q R are the 25th, 75th percentiles, and the interquartile range, respectively.
Entropy 24 00838 g001
Table 2. Computation of A = a { 2 F 1 ( . ) } | a = 0 and B = a { F D ( p ) ( . ) } | a = 0 when p = 2 and λ 1 = = λ p = λ .
Table 2. Computation of A = a { 2 F 1 ( . ) } | a = 0 and B = a { F D ( p ) ( . ) } | a = 0 when p = 2 and λ 1 = = λ p = λ .
N = 20 N = 30 N = 40
1 λ A B | ϵ | B | ϵ | B | ϵ |
0.10.06940.06949.1309   ×   10 16 0.06949.1309   ×   10 16 0.06949.1309   ×   10 16
0.30.22910.22913.7747   ×   10 14 0.22911.1102   ×   10 16 0.22911.1102   ×   10 16
0.50.42920.42922.6707   ×   10 9 0.42921.2458   ×   10 12 0.42926.6613   ×   10 16
0.70.70220.70225.9260   ×   10 6 0.70228.2678   ×   10 8 0.70221.3911   ×   10 9
0.91.16731.16340.00381.16657.2760   ×   10 4 1.16711.6081   ×   10 4
0.991.70431.58010.12411.62670.07761.65140.0529
Table 3. Parameters Σ 1 and Σ 2 used to compute KLD for central MCD.
Table 3. Parameters Σ 1 and Σ 2 used to compute KLD for central MCD.
Σ Σ 11 , Σ 22 , Σ 33 , Σ 12 , Σ 13 , Σ 23
Σ 1 1, 1, 1, 0.6, 0.2, 0.3
Σ 2 1, 1, 1, 0.3, 0.1, 0.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bouhlel, N.; Rousseau, D. A Generic Formula and Some Special Cases for the Kullback–Leibler Divergence between Central Multivariate Cauchy Distributions. Entropy 2022, 24, 838. https://doi.org/10.3390/e24060838

AMA Style

Bouhlel N, Rousseau D. A Generic Formula and Some Special Cases for the Kullback–Leibler Divergence between Central Multivariate Cauchy Distributions. Entropy. 2022; 24(6):838. https://doi.org/10.3390/e24060838

Chicago/Turabian Style

Bouhlel, Nizar, and David Rousseau. 2022. "A Generic Formula and Some Special Cases for the Kullback–Leibler Divergence between Central Multivariate Cauchy Distributions" Entropy 24, no. 6: 838. https://doi.org/10.3390/e24060838

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop