Next Article in Journal
Mutual Information between Order Book Layers
Next Article in Special Issue
A Generalized Measure of Cumulative Residual Entropy
Previous Article in Journal
Meaningful Secret Image Sharing with Saliency Detection
Previous Article in Special Issue
Weighted Relative Group Entropies and Associated Fisher Metrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cumulative Residual q-Fisher Information and Jensen-Cumulative Residual χ2 Divergence Measures

by
Omid Kharazmi
1,2,*,
Narayanaswamy Balakrishnan
2 and
Hassan Jamali
3
1
Department of Statistics, Faculty of Mathematical Sciences, Vali-e-Asr University of Rafsanjan, Rafsanjan P.O. Box 518, Iran
2
Department of Mathematics and Statistics, McMaster University, Hamilton, ON L8S 4L8, Canada
3
Department of Mathematics, Faculty of Mathematical Sciences, Vali-e-Asr University of Rafsanjan, Rafsanjan P.O. Box 518, Iran
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(3), 341; https://doi.org/10.3390/e24030341
Submission received: 4 January 2022 / Revised: 15 February 2022 / Accepted: 24 February 2022 / Published: 27 February 2022
(This article belongs to the Special Issue Measures of Information II)

Abstract

:
In this work, we define cumulative residual q-Fisher (CRQF) information measures for the survival function (SF) of the underlying random variables as well as for the model parameter. We also propose q-hazard rate (QHR) function via q-logarithmic function as a new extension of hazard rate function. We show that CRQF information measure can be expressed in terms of the QHR function. We define further generalized cumulative residual χ 2 divergence measures between two SFs. We then examine the cumulative residual q-Fisher information for two well-known mixture models, and the corresponding results reveal some interesting connections between the cumulative residual q-Fisher information and the generalized cumulative residual χ 2 divergence measures. Further, we define Jensen-cumulative residual χ 2 (JCR- χ 2 ) measure and a parametric version of the Jensen-cumulative residual Fisher information measure and then discuss their properties and inter-connections. Finally, for illustrative purposes, we examine a real example of image processing and provide some numerical results in terms of the CRQF information measure.

1. Introduction

Entropy-type and Fisher-type information measures have attracted great attention from researchers in information theory. Although these two types of information measures have different geneses, they are complementary to each other in the study of information resources. Among them, Fisher information and Shannon entropy are the fundamental information measures and have been used very broadly. The systems with complex structure can be thoroughly described in terms of their architecture (Fisher information) and their behavior (Shannon entropy) measures. For more details, one may refer to Cover and Thomas [1] and Zegers [2]. Fisher [3] had proposed an information measure for describing the interior properties of a probabilistic model. Shannon entropy originated from the pioneering work of Shannon [4], based on a study of the global behavior of systems modeled by a probability structure. Fisher information as well as Shannon entropy are quite important and have become fundamental quantities in numerous applied disciplines. These information measures and their extensions have been considered by several researchers in recent years. In addition, some divergence measures, based on Fisher information and Shannon entropy, have been introduced for measuring similarity and dissimilarity between two statistical models. For example, Kullback-Leibler, chi-square, and relative Fisher information divergences and their extensions have been used in this regard. For pertinent details, one may refer to Nielsen and Nock [5], Zegers [2], Popescu et al. [6], Sánchez-Moreno et al. [7], and Bercher [8].
Tsallis entropy (see [9]) is a generalized form of Shannon entropy and has found key applications in the context of non-extensive thermo-statistics. Some generalized forms of Fisher information that match well in the context of non-extensive thermo-statistics have also been introduced by some authors; see Johnson and Vignat [10], Furuichi [11], and Lutwak et al. [12]. One of the common ways of generalizing information measures is through their accumulated forms. This has been done for Shannon entropy, Tsallis entropy, and their related versions.
Let X denote a continuous random variable on the support X with survival function F ¯ θ ( x ) . The cumulative residual Fisher (CRF) information measure, introduced by Kharazmi and Balakrishnan [13], is then defined as
CI θ ( F ¯ θ ) = X F ¯ θ ( x ) log F ¯ θ ( x ) θ 2 d x ,
where log stands for the natural logarithm. Throughout this paper, we will suppress X in the integration with respect to x for ease of notation, unless a distinction becomes necessary. Due to the term F ¯ θ in the integrand in (1), it can be readily seen that CRF information measure in (1) provides decreasing weights for larger values of X. Hence, this information measure will naturally be robust to the presence of outliers. Kharazmi and Balakrishnan [13] have analogously defined the CRF information measure for the survival function ( F ¯ ) as
CI ( X ) = CI ( F ¯ ) = F ¯ ( x ) log F ¯ ( x ) x 2 d x .
They also provided an interesting representation for CI ( X ) information measure in (2), based on the hazard function, r F ( x ) = f ( x ) / F ¯ ( x ) , as
CI ( X ) = E [ r F ( X ) ] .
In the present paper, our primary goal is to introduce cumulative residual q-Fisher (CRQF) information, cumulative residual generalized- χ 2 (CRG- χ 2 ) divergence, and Jensen-cumulative residual- χ 2 (JCR- χ 2 ) and a parametric version of Jensen-cumulative residual Fisher (JCRF) information measures. We then examine some properties of these information measures and their interconnections in terms of two well-known mixture models that are commonly used in reliability, economics, and survival analysis.
The organization of the rest of this paper is as follows. In Section 2, we briefly describe some key information and entropy measures that are essential for all subsequent developments. Next, in Section 3, we define a cumulative residual q-Fisher (CRQF) information measure and q-hazard rate function. It is then shown that the CRQF information measure can be expressed via expectation involving q-hazard rate (QHR) function under proportional hazard model (PH) with proportionality parameter q. In Section 4, we propose the cumulative version of a generalized chi-square measure, called cumulative residual generalized- χ 2 (CRG- χ 2 ) divergence measure. We show that the first derivative of CRG- χ 2 measure with respect to the associated parameter is connected to the cumulative residual entropy measure, and when the parameter tends to zero, it is connected to the variance of the ratio of two survival functions. Next, we obtain the CRQF information measure for two well-known mixture models in Section 5. It is shown that the CRQF information measure for arithmetic mixture and harmonic mixture models are connected to the CRG- χ 2 divergence measure. In addition, we show that the harmonic mixture model involves optimal information under three optimization problems regarding the cumulative residual chi-square divergence measure. In Section 6, we first define Jensen-cumulative residual χ 2 (JCR- χ 2 ) measure and a parametric version of the Jensen-cumulative residual Fisher information measure and then discuss some of their properties. We also show that these two information measures are connected through arithmetic mixture models. In Section 7, we consider a real example of image processing and present some numerical results in terms of the CRQF information measure. Finally, some concluding remarks are made in Section 8.

2. Preliminaries

In this section, we briefly review some information measures that will be used in the sequel. The chi-square divergence between two SFs F ¯ and G ¯ , called cumulative residual χ 2 (CR- χ 2 ) divergence, is defined as
χ 2 ( F ¯ , G ¯ ) = ( G ¯ ( x ) F ¯ ( x ) ) 2 F ¯ ( x ) d x .
The χ 2 ( G ¯ , F ¯ ) divergence can also be defined in an analogous manner. For more details, see Kharazmi and Balakrishnan [13].
For a given continuous random variable X with survival function F ¯ ( x ) , the cumulative residual entropy (CRE) was defined by Rao et al. [14] as
ξ ( X ) = log F ¯ ( x ) F ¯ ( x ) d x .
The relative cumulative residual Fisher (RCRF) information between two absolutely continuous survival functions G ¯ θ and F ¯ θ is defined as
CD ( F ¯ θ , G ¯ θ ) = log G ¯ θ ( x ) θ log F ¯ θ ( x ) θ 2 F ¯ θ ( x ) d x .
For given absolutely continuous survival functions F ¯ 1 , , F ¯ n , the Jensen-cumulative residual Fisher (JCRF) information measure was defined by Kharazmi and Balakrishnan [13] as
JCI ( F ¯ 1 , , F ¯ n ; α ) = i = 1 n α i CI ( F ¯ i ) CI i = 1 n α i F ¯ i ,
where α 1 , , α n are non-negative real values with i = 1 n α i = 1 .

3. CRQF Information Measure

Here, we first define the cumulative residual q-Fisher (CRQF) information measure and the q-hazard rate (QHR) function. We then study some properties of the CRQF information measure and its connection to the QHR function.
The q-Fisher information of a density function f, defined by Lutwak et al. [12], is given by
I q ( f ) = log q f ( x ) x 2 f ( x ) d x ,
where log q ( x ) is the q-logarithmic function defined as
log q ( x ) = x q 1 q , x , q 0 .
For more details, see Furuichi [11], Yamano [15], and Masi [16]. Using this q-logarithmic function, we now propose two cumulative versions of the q-Fisher information in (8).
Definition 1.
Let F ¯ θ denote the survival function of variable X. Then, the CRQF information about parameter θ is defined as
CI q ( θ ) = log q F ¯ θ ( x ) θ 2 F ¯ θ ( x ) d x .
Definition 2.
The CRQF information measure for survival function F ¯ associated with variable X is defined as
CI q ( X ) = CI q ( F ¯ ) = log q F ¯ ( x ) x 2 F ¯ ( x ) d x .
Example 1.
Let X have a Weibull distribution with CDF F ( x ) = 1 e λ x β , x , β , λ > 0 . Then, the CRQF information measure of variable X, for β > 1 2 , is obtained as
CI q ( X ) = β λ 1 β Γ 2 1 β 1 + 2 q 2 1 β ,
where Γ ( . ) is the complete gamma function defined by Γ ( a ) = 0 x a 1 e x d x .
Figure 1 plots the CRQF information measure in (12) for different choices of parameters, from which we observe that CI q ( X ) gets maximized when q is decreased and the parameter λ is increased.

q-Hazard Rate Function and Its Connection to CRQF Information Measure

The hazard rate (HR) function of variable X with survival function F ¯ is given by
r ( x ) = log F ¯ ( x ) x = f ( x ) F ¯ ( x ) .
The HR function is a basic concept in reliability theory; see Barlow and Proschan [17] for elaborate details.
Now, we first propose a new extension of the HR function based on the q-logarithmic function in (9) and then study its connection to the CRQF information measure.
Definition 3.
For a random variable X with an absolutely continuous survival function F ¯ , the q-hazard rate (QHR) (or q-logarithmic hazard rate) function is defined as
r q ( x ) = log q F ¯ ( x ) x = r ( x ) F ¯ q ( x ) , q 0 , r ( x ) , q = 0 ,
where r ( x ) is the hazard rate function defined in (13).
Theorem 1.
Let variable X have its absolutely continuous survival function F ¯ and q-hazard rate function r q ( x ) for x > 0 . Then, for q > 0 , we have
CI q ( F ¯ ) = 1 q + 1 E X q r q ( X ) ,
where X q has proportional hazard model corresponding to baseline variable X with proportionality parameter q.
Proof. 
From the definition of CI q ( F ¯ ) in (11), we get
CI q ( F ¯ ) = 0 log q F ¯ ( x ) x 2 F ¯ ( x ) d x = 0 r q ( x ) f ( x ) F ¯ q ( x ) d x = 1 q + 1 E X q r q ( X ) ,
as required. □
Lemma 1.
The CI q ( F ¯ ) information measure is decreasing with respect to q > 0 .
Proof. 
From the definition of CI q ( F ) in (11), upon making use of Theorem 1, for each 0 < q 1 q 2 , we have
CI q 1 ( F ¯ ) = 0 r q 1 ( x ) f ( x ) F ¯ q 1 ( x ) d x = 0 r ( x ) f ( x ) F ¯ 2 q 1 ( x ) d x 0 r ( x ) f ( x ) F ¯ 2 q 2 ( x ) d x = 0 r q 2 ( x ) f ( x ) F ¯ q 2 ( x ) d x = CI q 2 ( F ¯ ) ,
as required. □
Theorem 2.
Let the non-negative random variable X have survival function F ¯ , CRF information CI ( F ¯ ) , and CRQF information CI q ( F ¯ ) . Then, we have:
 (i) 
If q > 0 , then CI q ( F ¯ ) CI ( F ¯ ) ;
 (ii) 
If q < 0 , then CI q ( F ¯ ) CI ( F ¯ )
with equality holding if and only if q = 0 .
Proof. 
From the definition of CRQF information measure in (11), and since F ¯ q ( x ) 1 for q > 0 , we have
CI q ( F ¯ ) = 0 log q F ¯ ( x ) x 2 F ¯ ( x ) d x = 0 r q ( x ) f ( x ) F ¯ q ( x ) d x = 0 r ( x ) f ( x ) F ¯ 2 q ( x ) d x 0 r ( x ) f ( x ) d x = CI ( F ¯ ) ,
which proves Part (i). Part (ii) can be proved in a similar manner. □

4. Generalized Cumulative Residual χ 2 Divergence Measures

In this section, we first define a cumulative form of a generalized chi-square divergence measure and then examine some of its properties. A generalized version of the χ 2 divergence between two densities f and g, for α 0 , considered by Basu et al. [18], is defined as
χ α 2 ( f , g ) = 1 + α 2 g ( x ) f ( x ) 2 f 1 α ( x ) d x .
For more details, see also Ghosh et al. [19].
Definition 4.
The cumulative residual generalized χ 2 (CRG- χ 2 ) divergence between two survival functions F ¯ and G ¯ , for α 0 , is defined as
χ α 2 F ¯ , G ¯ = α + 1 2 F ¯ ( x ) G ¯ ( x ) 2 F ¯ 1 α ( x ) d x .
It is readily seen from (16) that
χ 2 ( F ¯ , G ¯ ) = lim α 0 + χ α 2 ( F ¯ , G ¯ ) 2 ,
and we also have
χ α 2 F ¯ , G ¯ α + 1 2 χ 2 ( F ¯ , G ¯ ) .
Theorem 3.
Let the variables X and Y have survival functions F ¯ and G ¯ , respectively, and χ α 2 F ¯ , G ¯ be the corresponding CRG- χ 2 divergence measure between them. Then, we have
α χ α 2 F ¯ , G ¯ | α = 0 = 1 2 χ 2 ( F ¯ , G ¯ ) ξ ( F ¯ ) + 2 K ( G ¯ , F ¯ ) + G ¯ 2 ( x ) F ¯ ( x ) log F ¯ ( x ) d x ,
where ξ ( F ¯ ) is the CRE information measure defined in (5) and K ( G ¯ , F ¯ ) is the cumulative residual inaccuracy measure given by
K ( G ¯ , F ¯ ) = G ¯ ( x ) log F ¯ ( x ) d x .
Proof. 
Upon considering the CRG- χ 2 measure in (16) and differentiating it with respect to α , we obtain
α χ α 2 F ¯ , G ¯ | α = 0 = 1 2 G ¯ ( x ) F ¯ ( x ) 2 F ¯ ( x ) d x + 1 2 G ¯ ( x ) F ¯ ( x ) 2 F ¯ ( x ) log F ¯ ( x ) d x = 1 2 χ 2 ( F ¯ , G ¯ ) + 1 2 F ¯ ( x ) 2 G ¯ ( x ) + G ¯ 2 ( x ) F ¯ ( x ) log F ¯ ( x ) d x = 1 2 χ 2 ( F ¯ , G ¯ ) + 1 2 F ¯ ( x ) log F ¯ ( x ) d x G ¯ ( x ) log F ¯ ( x ) d x + 1 2 G ¯ 2 ( x ) F ¯ ( x ) log F ¯ ( x ) d x = 1 2 χ 2 ( F ¯ , G ¯ ) ξ ( F ¯ ) + 2 K ( G ¯ , F ¯ ) + G ¯ 2 ( x ) F ¯ ( x ) log F ¯ ( x ) d x ,
as required. □
Theorem 4.
Let the non-negative continuous variables X and Y have survival functions F ¯ and G ¯ , respectively, and have a common mean μ. Then,
lim α 0 + χ α 2 ( G ¯ , F ¯ ) = μ 2 V a r f e T ( X ) ,
where T ( x ) = G ¯ ( x ) F ¯ ( x ) and f e ( x ) = F ¯ ( x ) μ is the equilibrium distribution of variable X.
Proof. 
From the definition of GCR- χ 2 divergence measure in (16) and the facts that F ¯ ( x ) d x = G ¯ ( x ) d x = μ , we obtain
2 lim α 0 + χ α 2 ( F ¯ , G ¯ ) = ( G ¯ ( x ) F ¯ ( x ) ) 2 F ¯ ( x ) d x = G ¯ 2 ( x ) F ¯ ( x ) d x + F ¯ ( x ) d x 2 G ¯ ( x ) d x = μ G ¯ ( x ) F ¯ ( x ) 2 F ¯ ( x ) μ d x μ G ¯ ( x ) F ¯ ( x ) F ¯ ( x ) μ d x 2 = μ G ¯ ( x ) F ¯ ( x ) 2 f e ( x ) d x μ G ¯ ( x ) F ¯ ( x ) f e ( x ) d x 2 = μ V a r f e G ¯ ( X ) F ¯ ( X ) = μ V a r f e T ( X ) ,
as required. □
Theorem 5.
Let the variables X and Y have survival functions F ¯ and G ¯ , respectively, and χ α 2 F ¯ , G ¯ be the corresponding CRG- χ 2 divergence measure between them. Then, for each 0 α 1 α 2 , we have
χ α 1 2 F ¯ , G ¯ α 1 + 1 α 2 + 1 χ α 2 2 F ¯ , G ¯ .
Proof. 
From the definition of CRG- χ 2 in (16) and that fact that F ¯ α 1 ( x ) F ¯ α 1 ( x ) for 0 α 1 α 2 , we have
2 α 1 + 1 χ α 1 2 F ¯ , G ¯ = G ¯ ( x ) F ¯ ( x ) 2 F ¯ 1 α 1 ( x ) d x G ¯ ( x ) F ¯ ( x ) 2 F ¯ 1 α 2 ( x ) d x = 2 α 2 + 1 χ α 2 2 F ¯ , G ¯ ,
as required. □

5. Cumulative Residual q -Fisher Information for Two Well-Known Mixture Models

In this section, we study the CRQF information measure for the well-known arithmetic mixture and harmonic mixture distributions.

5.1. Arithmetic Mixture Distribution

The arithmetic mixture distribution based on survival functions F ¯ 1 and F ¯ 2 is given by
F ¯ η ( x ) = η F ¯ 1 ( x ) + ( 1 η ) F ¯ 2 ( x ) , η ( 0 , 1 ) .
For more details about the mixture model in (18), see Marshall and Olkin [20]. The CRQF information measure about parameter η in (18) is given by
CI q ( η ) = log q F ¯ η ( x ) η 2 F ¯ η ( x ) d x = F ¯ 2 ( x ) F ¯ 1 ( x ) 2 F ¯ η 1 2 q ( x ) d x .
Theorem 6.
The CRQF information measure in (19) is given by
CI q ( η ) = 8 2 q + 1 M 1 2 χ 2 q 2 ( F ¯ η , F ¯ 1 ) , χ 2 q 2 ( F ¯ η , F ¯ 2 ) ,
where M 1 2 ( . , . ) is power mean with exponent 1 2 , defined as M 1 2 ( x , y ) = x 1 2 2 + y 1 2 2 2 for positive x and y.
Proof. 
From the mixture model in (18), we have
F ¯ 2 ( x ) F ¯ 1 ( x ) = F ¯ η ( x ) F ¯ 1 ( x ) 1 η = F ¯ 2 ( x ) F ¯ η ( x ) η .
Now, from the definition of the CQTF information measure in (19), we find
CI q ( η ) = F ¯ 2 ( x ) F ¯ 1 ( x ) 2 F ¯ η 1 2 q ( x ) d x = 2 ( 1 + 2 q ) ( 1 η ) 2 χ 2 q 2 ( F ¯ η , F ¯ ) , F ¯ = F ¯ 1 , 2 ( 1 + 2 q ) η 2 χ 2 q 2 ( F ¯ η , F ¯ ) , F ¯ = F ¯ 2 .
Then, from (21), we obtain
CI q ( η ) = 2 1 + 2 q χ 2 q 2 ( F ¯ η , F ¯ 1 ) + χ 2 q 2 ( F ¯ η , F ¯ 2 ) 2 = 8 1 + 2 q 1 2 χ 2 q 2 ( F ¯ η , F 1 ) + 1 2 χ 2 q 2 ( F ¯ η , F ¯ 2 ) 2 = 8 1 + 2 q M 1 2 χ 2 q 2 ( F ¯ η , F ¯ 1 ) , χ 2 q 2 ( F ¯ η , F ¯ 2 ) ,
as required. □
Theorem 7.
Let the non-negative random variable X have arithmetic mixture survival function in (18) and with CRF information CI ( η ) and CRQF information CI q ( η ) . Then, we have:
 (i) 
If q > 0 , then CI q ( η ) CI ( η ) ;
 (ii) 
If q < 0 , then CI q ( η ) CI ( η ) ,
with equality holding if and only if q = 0 .
Proof. 
From the definition of CRF information measure, we have
CI ( η ) = log F ¯ η ( x ) η 2 F ¯ η ( x ) d x = F ¯ 2 ( x ) F ¯ 1 ( x ) 2 F ¯ η ( x ) d x = ( 1 η ) 2 F ¯ η ( x ) F ¯ 1 ( x ) 2 F ¯ η ( x ) d x = ( 1 η ) 2 χ 2 ( F ¯ η , F ¯ 1 ) .
Furthermore, from the definition of CRTF information measure, we find
CI q ( η ) = log q F ¯ η ( x ) η 2 F ¯ η ( x ) d x = F ¯ 2 ( x ) F ¯ 1 ( x ) 2 F ¯ η ( x ) F ¯ η 2 q ( x ) d x = ( 1 η ) 2 F ¯ η ( x ) F ¯ 1 ( x ) 2 F ¯ η ( x ) F ¯ η 2 q ( x ) d x q > 0 ( 1 η ) 2 F ¯ η ( x ) F ¯ 1 ( x ) 2 F ¯ η ( x ) d x = ( 1 η ) 2 χ 2 ( F ¯ η , F ¯ 0 ) = CI ( η ) ,
which proves Part (i). Part (ii) can be proved in an analogous manner. □

5.2. Harmonic Mixture Distribution

The harmonic mixture (HM) distribution based on survival functions F ¯ 1 and F ¯ 2 is given by
F ¯ η ( x ) = F ¯ 2 ( x ) F ¯ 1 ( x ) η F ¯ 2 ( x ) + ( 1 η ) F ¯ 1 ( x ) , η ( 0 , 1 ) .
For more details about harmonic mixture distributions, one may refer to Schmidt [21]. The CRQF information measure about parameter η in (22) is given by
CI q ( η ) = log q F ¯ η ( x ) η 2 F ¯ η ( x ) d x = F ¯ 2 ( x ) F ¯ 1 ( x ) 2 η F ¯ 2 + ( 1 η ) F ¯ 1 ( x ) 1 2 q F ¯ 1 ( x ) F ¯ 2 ( x ) 2 q + 1 d x .
Theorem 8.
An upper bound for the CRQF information measure in (23), for q 0 , is given by
CI q ( η ) 8 2 q + 1 M 1 2 χ 2 q 2 ( F ¯ T , F ¯ 1 ) , χ 2 q 2 ( F ¯ T , F ¯ 2 ) ,
where
F ¯ T ( x ) = F ¯ η ( x ) F ¯ 1 ( x ) F ¯ 2 ( x ) = η F ¯ 2 ( x ) + ( 1 η ) F ¯ 1 ( x )
and M 1 2 ( . , . ) is as defined earlier in Theorem 6.
Proof. 
Because F ¯ T ( x ) = η F ¯ 2 ( x ) + ( 1 η ) F ¯ 1 ( x ) , it is readily seen that
F ¯ 2 ( x ) F ¯ T ( x ) = ( 1 η ) F ¯ 2 ( x ) F ¯ 1 ( x ) ,
F ¯ 1 ( x ) F ¯ T ( x ) = η F ¯ 1 ( x ) F ¯ 2 ( x ) .
By using these and (23), we find
CI q ( η ) 2 ( 1 η ) 2 ( 1 + 2 q ) χ 2 q 2 ( F ¯ T , F ¯ 2 ) ,
CI q ( η ) 2 η 2 ( 1 + 2 q ) χ 2 q 2 ( F ¯ T , F ¯ 1 ) .
Adding the above two inequalities, we readily get
CI q ( η ) 2 2 q + 1 χ 2 q 2 ( F ¯ T , F ¯ 1 ) + χ 2 q 2 ( F ¯ T , F ¯ 2 ) 2 = 8 2 q + 1 M 1 2 χ 2 q 2 ( F ¯ T , F ¯ 1 ) , χ 2 q 2 ( F ¯ T , F ¯ 2 ) ,
as required. □

5.3. HM Distribution Having Optimal Information under CR- χ 2 Divergence Measure

In this section, we discuss the optimal information property of the harmonic mixture survival function in (22). For this purpose, we consider the optimization problem for cumulative residual chi-square divergence under three types of constraints. For more details about optimal information properties of some mixture distributions (arithmetic, geometric, and α mixture distributions), one may refer to Asadi et al. [22] and the references therein.
Theorem 9.
Let F ¯ , F ¯ 0 , and F ¯ 1 be three survival functions. Then, the solution to the information problem
min F ¯ χ 2 ( F ¯ : F ¯ 0 ) s u b j e c t t o χ 2 ( F ¯ : F ¯ 1 ) = θ , F ¯ ( x ) d x = μ ,
is the HM distribution in (22) with mixing parameter η = 1 1 + λ 0 and λ 0 > 0 is the Lagrangian multiplier.
Proof. 
We use the Lagrange multiplier technique to find the solution of the optimization problem stated in (25). Hence, we have
L ( F ¯ , λ 0 , λ 1 ) = ( F ¯ ( x ) F ¯ 0 ( x ) ) 2 F ¯ 0 ( x ) d x + λ 0 ( F ¯ ( x ) F ¯ 1 ( x ) ) 2 F ¯ 1 ( x ) d x + λ 1 F ¯ ( x ) d x .
Now, differentiating with respect to F ¯ , we obtain
F ¯ L ( F ¯ , λ 0 , λ 1 ) = 2 F ¯ ( x ) F ¯ 0 F ¯ 0 + 2 λ 0 F ¯ ( x ) F ¯ 1 F ¯ 1 + λ 1 .
Setting (26) to zero, we get the optimal survival function as
F ¯ ( x ) = 1 η F ¯ 0 ( x ) + 1 η F ¯ 1 ( x ) ,
where η = 1 1 + λ 0 . □
Theorem 10.
Let F ¯ , F ¯ 0 , and F ¯ 1 be three survival functions. Then, the solution to the information problem
min F ¯ { w χ 2 ( F ¯ : F ¯ 1 ) + ( 1 w ) χ 2 ( F ¯ : F ¯ 2 ) } s u b j e c t t o F ¯ ( x ) d x = μ , 0 w 1 ,
is the HM distribution in (22) with mixing parameter η = w .
Proof. 
Making use of the Lagrangian multiplier technique, and proceeding in the same way as in the proof of Theorem 25, the required result can be obtained. □
Theorem 11.
Let F ¯ , F ¯ 0 , and F ¯ 1 be three survival functions and T ( X ) = F ¯ ( X ) F ¯ 2 ( x ) . Then, the solution to the information problem
min F ¯ χ 2 ( F ¯ : F ¯ 0 ) s u b j e c t t o E f e ( T ( X ) ) = θ , F ¯ ( x ) d x = μ ,
is HM model with mixing parameter η = 1 1 + λ 0 and λ 0 > 0 is the Lagrangian multiplier, where f e is the equilibrium distribution as defined in Theorem 4.
Proof. 
Making use of the Lagrangian multiplier technique, and proceeding the same way as in the proof of Theorem 25, the required result is obtained. □

6. Jensen-Cumulative Residual χ 2 and Parametric Version of Jensen-Cumulative Residual Fisher Divergence Measures

In this section, we first introduce the Jensen-cumulative residual χ 2 divergence measure and then propose a parametric version of the Jensen-cumulative residual Fisher information in (7). Next, we show that these two Jensen-type divergence measures are connected through arithmetic mixture distributions.

6.1. Jensen-Cumulative Residual χ 2 Divergence Measure

Definition 5.
Consider the survival functions F ¯ 1 , θ , , F ¯ n , θ and G ¯ θ . Then, the Jensen-cumulative residual χ 2 (JCR- χ 2 ) information measure is defined as
J χ 2 α F ¯ 1 , θ , , F ¯ n , θ ; G ¯ θ = i = 1 n α i χ 2 F ¯ i , θ , G ¯ θ χ 2 i = 1 n α i F ¯ i , θ , G ¯ θ ,
where α 1 , , α n are non-negative real values with i = 1 n α i = 1 .
Theorem 12.
The JCR- χ 2 information measure defined in (29) is non-negative.
Proof. 
From the definition of the CR- χ 2 measure, we have
i = 1 n α i χ 2 F ¯ i , θ , G ¯ = i = 1 n α i ( F ¯ i , θ ( x ) G ¯ θ ( x ) ) 2 F ¯ i , θ ( x ) d x = G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) d x + i = 1 n α i F ¯ i , θ ( x ) d x 2 G ¯ θ ( x ) d x
and
χ 2 i = 1 n α i F ¯ i , θ , G ¯ θ = i = 1 n α i F ¯ i , θ ( x ) G ¯ θ ( x ) 2 i = 1 n α i F ¯ i , θ ( x ) d x = G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) d x + i = 1 n α i F ¯ i , θ ( x ) d x 2 G ¯ θ ( x ) d x .
Upon making use of the above results, from the definition of the JCR- χ 2 measure in (29), we find
J χ 2 α F ¯ 1 , θ , , F ¯ n , θ ; G ¯ θ = i = 1 n α i χ 2 F ¯ i , θ , G ¯ χ 2 i = 1 n α i F ¯ i , θ , G ¯ θ = i = 1 n α i ( F ¯ i , θ ( x ) G ¯ θ ( x ) ) 2 F ¯ i , θ ( x ) d x i = 1 n α i F ¯ i , θ ( x ) G ¯ θ ( x ) 2 i = 1 n α i F ¯ i , θ ( x ) d x = G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) d x G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) d x = G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) 1 i = 1 n α i F ¯ i , θ ( x ) d x .
Finally, from the arithmetic mean-harmonic mean inequality (see Theorem 5.1 of Cvetkovski [23]), we get
J χ 2 α F ¯ 1 , θ , , F ¯ n , θ ; G ¯ θ = G ¯ θ 2 ( x ) i = 1 n α i F ¯ i , θ ( x ) 1 i = 1 n α i F ¯ i , θ ( x ) d x 0 ,
as required. □

6.2. Parametric Version of Jensen-Cumulative Residual Fisher Information Divergence

In this subsection, we introduce a parametric form of the JCRQF information measure in (7).
Definition 6.
Consider the survival functions F ¯ 1 , θ , , F ¯ n , θ . Then, a parametric form of JCRQF (P-JCRF) information measure about parameter θ, for non-negative real values α 1 , , α n with i = 1 n α i = 1 , is defined as
JCI ( F ¯ 1 , θ , , F ¯ n , θ , α ) = i = 1 n α i CI θ ( F ¯ i , θ ) CI θ i = 1 n α i F ¯ i , θ ,
where
CI θ ( F ¯ i , θ ) = log F ¯ i , θ ( x ) θ 2 F ¯ i , θ ( x ) d x , i = 1 , , n .
Theorem 13.
The P-JCRF information measure in (30) can be represented as a mixture of CD measures in (6) as
JCI ( F ¯ 1 , θ , , F ¯ n , θ , α ) = i = 1 n α i CD ( F ¯ i , θ , F ¯ T , θ ) ,
where F ¯ T , θ = i = 1 n α i F ¯ i , θ is the weighted survival function.
Proof. 
From the definition in (30), we have
JCI ( F ¯ 1 , θ , , F ¯ n , θ , α ) = i = 1 n α i CI θ ( F ¯ i , θ ) CI θ i = 1 n α i F ¯ i , θ = i = 1 n α i 0 log F ¯ i , θ ( x ) θ 2 F ¯ i , θ ( x ) d x 0 log i = 1 n α i F ¯ i , θ ( x ) θ 2 i = 1 n α i F ¯ i , θ ( x ) d x .
Furthermore, we have
i = 1 n α i CD ( F ¯ i , θ , F ¯ T , θ ) = i = 1 n α i 0 log F ¯ i , θ ( x ) θ log i = 1 n α i F ¯ i , θ ( x ) θ 2 F ¯ i , θ ( x ) d x = i = 1 n α i 0 { log F ¯ i , θ ( x ) θ 2 2 log F ¯ i , θ ( x ) θ log i = 1 n α i F ¯ i , θ ( x ) θ + log i = 1 n α i F ¯ i , θ ( x ) θ 2 } F ¯ i , θ ( x ) d x = i = 1 n α i 0 log F ¯ i , θ ( x ) θ 2 F ¯ i , θ ( x ) d x 0 log i = 1 n α i F ¯ i , θ ( x ) θ 2 i = 1 n α i F ¯ i , θ ( x ) d x ,
as required. □

6.3. Connection between the P-JCRF Information and JCR- χ 2 Divergence Measures

Let F ¯ 1 , , F ¯ n and G ¯ be arbitrary continuous survival functions. Consider the arithmetic mixture distributions with survival functions
H ¯ i , Λ ( x ) = Λ F ¯ i ( x ) + ( 1 Λ ) G ¯ ( x ) , i = 1 , , n ,
and
H ¯ α , Λ ( x ) = i = 1 n α i H ¯ i , Λ ( x ) = Λ i = 1 n α i F ¯ i ( x ) + ( 1 Λ ) G ¯ ( x ) ,
where α 1 , , α n are non-negative real values with i = 1 n α i = 1 and 0 Λ 1 .
The JCRQF information measure of survival functions H ¯ 1 , Λ , , H ¯ 1 , Λ , about the mixing parameter Λ , for 0 Λ 1 , is given by
JCI ( H ¯ 1 , Λ , , H ¯ n , Λ ; α ) = i = 1 n α i CI Λ ( H ¯ i , Λ ) CI Λ i = 1 n α i H ¯ i , Λ .
Theorem 14.
The connection between J χ 2 α F ¯ 1 , θ , , F ¯ n , θ ; G ¯ θ in (29) and JCI ( H ¯ 1 , Λ , , H ¯ n , Λ ; α ) in (33) is given by
JCI ( H ¯ 1 , Λ , , H ¯ n , Λ ; α ) = 1 Λ 2 J χ 2 α H ¯ 1 , Λ , , H ¯ n , Λ ; G ¯ .
Proof. 
From the definition of JCI ( H ¯ 1 , Λ , , H ¯ n , Λ ; α ) in (33), we find
JCI ( H ¯ 1 , Λ , , H ¯ n , Λ ; α ) = i = 1 n α i CI Λ ( H ¯ i , Λ ) CI Λ i = 1 n α i H ¯ i , Λ = i = 1 n α i log H ¯ i , Λ ( x ) Λ 2 H ¯ i , Λ ( x ) d x log H ¯ i , Λ ( x ) Λ 2 H ¯ i , Λ ( x ) ( x ) d x = i = 1 n α i log Λ F ¯ i ( x ) + ( 1 Λ ) G ¯ ( x ) Λ 2 H ¯ i , Λ ( x ) d x log i = 1 n α i H ¯ i , Λ ( x ) Λ 2 H ¯ α , Λ ( x ) d x = 1 Λ 2 i = 1 n α i ( H ¯ i , Λ ( x ) G ¯ ( x ) ) 2 H ¯ i , Λ ( x ) d x 1 Λ 2 ( H ¯ α , Λ ( x ) G ¯ ( x ) ) 2 H ¯ α , Λ ( x ) d x = 1 Λ 2 i = 1 n α i χ 2 H ¯ i , Λ , G ¯ χ 2 H ¯ α , Λ , G ¯ = 1 Λ 2 J χ 2 α H ¯ 1 , Λ , , H ¯ n , Λ ; G ¯ ,
as required. □

7. Application of CRQF Information Measure

We now demonstrate an application of the CRQF information measure to image processing. Let X 1 , , X n be a random sample from density f with corresponding CDF F. The kernel estimate of density f, based on kernel function K with bandwidth h > 0 , is given by
f ^ ( x ) = 1 n h i = 1 n K x X i h .
Further, the non-parametric estimate of the survival function F ¯ ( x ) , at a given point x, is given by
F ¯ ^ ( x ) = 1 n i = 1 n I X i > x ,
where I is the indicator function taking the value 1 if the condition inside brackets is satisfied and 0 otherwise. Then, the integrated non-parametric estimate of CI q ( F ¯ ) in (11) is given by
CI q ^ ( F ¯ ) = f ^ 2 ( x ) F ¯ ^ ( x ) 1 2 q d x = 1 h 2 n 1 + 2 q i = 1 n K ( x X i h ) 2 i = 1 n I ( X i > x ) 1 2 q d x .
From (36) and with the use of Gaussian kernel K ( u ) = 1 2 π e u 2 2 , we have
CI q ^ ( F ¯ ) = 1 2 π h 2 n 1 + 2 q i = 1 n e 1 2 ( x X i h ) 2 2 i = 1 n I ( X i > x ) 1 2 q d x .
Thus, from (37) and using the Cavalieri-Simpson rule for numerical integration, the empirical estimate of the CRQF information measure can be obtained.
Next, we provide an example of image processing and compute the CRQF information and Fisher information (FI) measures for the original picture and its adjusted versions. Figure 2 shows a sample picture of two parrots (original picture) labeled as X and three adjusted versions of the original picture labeled as Y (increasing brightness), Z (increasing contrast), and W (gamma corrected). The available data of the main picture are 768 × 512 cells and the gray level of each cell has a value between 0 (black) and 1 (white). In order to examine the amount of content CRQF information measure of the original picture and compare with the information values of adjusted versions, we consider three cases as Y( = X + 0.3 ), Z( = 2 X ), and W( = X ). For pertinent details, see EBImage package in R software (Pau et al. [24]).
We have plotted in Figure 3 the extracted histograms with the corresponding empirical densities for pictures X, Y, Z, and W. As we can see from Figure 2 and Figure 3, the highest degree of similarity is first related to W and then to Y, whereas Z has the highest degree of divergence with respect to the original picture X. We have presented the CRQF information (for selected values of q = 0.55 and 0.7 ) and Fisher information (FI) measures for all four pictures in Table 1. It is easily seen that both information measures get increased when the similarity is decreased with respect to the original picture. This fact coincides with the minimum Fisher information principle. Therefore, the CRQF information measure can be considered as an efficient criteria, just as the Fisher information measure, in analyzing interior properties of the complex systems.

8. Concluding Remarks

In this paper, we have proposed cumulative residual q-Fisher (CRQF) information, q-hazard rate function (QHR), cumulative residual generalized χ 2 (CRG- χ 2 ) divergence, and Jensen-cumulative residual χ 2 (JCR- χ 2 ) divergence measures. We have shown that the CRQF information measure can be expressed in terms of expectation involving the q-hazard rate function. Further, we have established some results concerning the CRQF information and CRG- χ 2 divergence measures. We have specifically shown that the first derivative of CRG- χ 2 divergence with respect to the associated parameter is connected to cumulative residual entropy measure and, when its parameter tends to zero, it is connected to the variance of the ratio of two survival functions. We have also presented some results associated with the CRQF information measure for two well-known mixture models, namely, arithmetic mixture (AM) and harmonic mixture (HM) models. We have specifically shown that the CRQF information of AM and HM models can be expressed in terms of power mean of CRG- χ 2 divergence measures. Interestingly, we have shown that the harmonic mixture model possesses optimal information under three optimization problems associated with the cumulative residual χ 2 divergence measure. We have also proposed a Jensen-cumulative residual χ 2 divergence and a parametric version of the Jensen-cumulative residual Fisher (P-JCRF) information measures and have shown that they are connected. Finally, we have described an application of the CRQF information measure by considering an example in image processing. It will naturally be of great interest to study empirical versions of these measures and their potential applications to inferential problems. We are currently looking into this problem and hope to report the findings in a future paper.

Author Contributions

All authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cover, T.M.; Thomas, J.A. Information theory and statistics. Elem. Inf. Theory 1991, 1, 279–335. [Google Scholar]
  2. Zegers, P. Fisher information properties. Entropy 2015, 17, 4918–4939. [Google Scholar] [CrossRef] [Green Version]
  3. Fisher, R.A. Tests of significance in harmonic analysis. Proc. R. Soc. Lond. Ser. A 1929, 125, 54–59. [Google Scholar]
  4. Shannon, C.E. A mathematical theory of communication. Bell. System. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  5. Nielsen, F.; Nock, R. On the chi square and higher-order chi distances for approximating f-divergences. IEEE Signal Process. Lett. 2013, 21, 10–13. [Google Scholar] [CrossRef] [Green Version]
  6. Popescu, P.G.; Preda, V.; Sluşanschi, E.I. Bounds for Jeffreys-Tsallis and Jensen-Shannon-Tsallis divergences. Phys. Stat. Mech. Its Appl. 2014, 413, 280–283. [Google Scholar] [CrossRef]
  7. Sánchez-Moreno, P.; Zarzo, A.; Dehesa, J.S. Jensen divergence based on Fisher’s information. J. Phys. A Math. Theor. 2012, 45, 125305. [Google Scholar] [CrossRef] [Green Version]
  8. Bercher, J.F. Some properties of generalized Fisher information in the context of nonextensive thermostatistics. Phys. Stat. Mech. Its Appl. 2013, 392, 3140–3154. [Google Scholar] [CrossRef] [Green Version]
  9. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  10. Johnson, O.; Vignat, C. Some results concerning maximum Rényi entropy distributions. In: Annales de l’Institut Henri Poincare (B). Probab. Stat. 2007, 43, 339–351. [Google Scholar]
  11. Furuichi, S. On the maximum entropy principle and the minimization of the Fisher information in Tsallis statistics. J. Math. Phys. 2009, 50, 013303. [Google Scholar] [CrossRef] [Green Version]
  12. Lutwak, E.; Lv, S.; Yang, D.; Zhang, G. Extensions of Fisher information and Stam’s inequality. IEEE Trans. Inf. Theory 2012, 58, 1319–1327. [Google Scholar] [CrossRef]
  13. Kharazmi, O.; Balakrishnan, N. Cumulative residual and relative cumulative residual Fisher information and their properties. IEEE Trans. Inf. Theory 2021, 67, 6306–6312. [Google Scholar] [CrossRef]
  14. Rao, M.; Chen, Y.; Vemuri, B.C.; Wang, F. Cumulative residual entropy: A new measure of information. IEEE Trans. Inf. Theory 2004, 50, 1220–1228. [Google Scholar] [CrossRef]
  15. Yamano, T. Some properties of q-logarithm and q-exponential functions in Tsallis statistics. Phys. Stat. Mech. Its Appl. 2002, 305, 486–496. [Google Scholar] [CrossRef]
  16. Masi, M. A step beyond Tsallis and Rényi entropies. Phys. Lett. A 2005, 338, 217–224. [Google Scholar] [CrossRef] [Green Version]
  17. Barlow, R.E.; Proschan, F. Statistical Theory of Reliability and Life Testing: Probability Models; Holt, Rinehart and Winston: New York, NY, USA, 1975. [Google Scholar]
  18. Basu, A.; Harris, I.R.; Hjort, N.L.; Jones, M.C. Robust and efficient estimation by minimising a density power divergence. Biometrika 1998, 85, 549–559. [Google Scholar] [CrossRef] [Green Version]
  19. Ghosh, A.; Harris, I.R.; Maji, A.; Basu, A.; Pardo, L. A generalized divergence for statistical inference. Bernoulli 2017, 23, 2746–2783. [Google Scholar] [CrossRef] [Green Version]
  20. Marshall, A.W.; Olkin, I. Life Distributions; Springer: New York, NY, USA, 2007. [Google Scholar]
  21. Schmidt, U. Axiomatic Utility Theory Under Risk: Non-Archimedean Representations and Application to Insurance Economics; Springer: Berlin, Germany, 2012. [Google Scholar]
  22. Asadi, M.; Ebrahimi, N.; Kharazmi, O.; Soofi, E.S. Mixture models, Bayes Fisher information, and divergence measures. IEEE Trans. Inf. Theory 2018, 65, 2316–2321. [Google Scholar] [CrossRef]
  23. Cvetkovski, Z. Inequalities: Theorems, Techniques and Selected Problems; Springer: New York, NY, USA, 2012. [Google Scholar]
  24. Pau, G.; Fuchs, F.; Sklyar, O.; Boutros, M.; Huber, W. EBImage-an R package for image processing with applications to cellular phenotypes. Bioinformatics 2010, 26, 979–981. [Google Scholar] [CrossRef] [PubMed]
Figure 1. 3D plots of the CRQF information measure in (12) for some selected choices of the Weibull parameters.
Figure 1. 3D plots of the CRQF information measure in (12) for some selected choices of the Weibull parameters.
Entropy 24 00341 g001
Figure 2. Sample picture of two parrots with its adjustments.
Figure 2. Sample picture of two parrots with its adjustments.
Entropy 24 00341 g002
Figure 3. The histograms and the corresponding empirical densities for pictures X, Y, Z, and W.
Figure 3. The histograms and the corresponding empirical densities for pictures X, Y, Z, and W.
Entropy 24 00341 g003
Table 1. The CRQF information and FI measures.
Table 1. The CRQF information and FI measures.
CRQF ( q = 0 . 55 )CRQF ( q = 0 . 7 )FI
X0.00680.00570.0039
Y0.00850.00660.0447
Z0.00960.00710.0475
W0.00820.00670.0052
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kharazmi, O.; Balakrishnan, N.; Jamali, H. Cumulative Residual q-Fisher Information and Jensen-Cumulative Residual χ2 Divergence Measures. Entropy 2022, 24, 341. https://doi.org/10.3390/e24030341

AMA Style

Kharazmi O, Balakrishnan N, Jamali H. Cumulative Residual q-Fisher Information and Jensen-Cumulative Residual χ2 Divergence Measures. Entropy. 2022; 24(3):341. https://doi.org/10.3390/e24030341

Chicago/Turabian Style

Kharazmi, Omid, Narayanaswamy Balakrishnan, and Hassan Jamali. 2022. "Cumulative Residual q-Fisher Information and Jensen-Cumulative Residual χ2 Divergence Measures" Entropy 24, no. 3: 341. https://doi.org/10.3390/e24030341

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop