Next Article in Journal
Intermediate-Temperature Creep Deformation and Microstructural Evolution of an Equiatomic FCC-Structured CoCrFeNiMn High-Entropy Alloy
Previous Article in Journal
Spatial Heterogeneity in the Occurrence Probability of Rainstorms over China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Some Order Preserving Inequalities for Cross Entropy and Kullback–Leibler Divergence

1
College of Intelligence and Computing, Tianjin University, Tianjin 300350, China
2
Graphics and Imaging Laboratory, University of Girona, Campus Montilivi, 17003 Girona, Spain
3
Department of Engineering Science, University of Oxford, Oxford OX1 3PJ, UK
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(12), 959; https://doi.org/10.3390/e20120959
Submission received: 8 November 2018 / Revised: 2 December 2018 / Accepted: 8 December 2018 / Published: 12 December 2018

Abstract

:
Cross entropy and Kullback–Leibler (K-L) divergence are fundamental quantities of information theory, and they are widely used in many fields. Since cross entropy is the negated logarithm of likelihood, minimizing cross entropy is equivalent to maximizing likelihood, and thus, cross entropy is applied for optimization in machine learning. K-L divergence also stands independently as a commonly used metric for measuring the difference between two distributions. In this paper, we introduce new inequalities regarding cross entropy and K-L divergence by using the fact that cross entropy is the negated logarithm of the weighted geometric mean. We first apply the well-known rearrangement inequality, followed by a recent theorem on weighted Kolmogorov means, and, finally, we introduce a new theorem that directly applies to inequalities between K-L divergences. To illustrate our results, we show numerical examples of distributions.

1. Introduction

Cross entropy and Kullback–Leibler (K-L) divergence are applied to and discussed in studies on machine learning [1], visualization (e.g., to interpret the pipeline of visualization using the cost–benefit model [2] and several other applications [3]), computer graphics [4], and many other fields. Cross entropy and K-L divergence are two fundamental quantities of information theory. Cross entropy gives the average code length needed to represent one distribution by another distribution, and the excess code needed over the optimal coding is given by the K-L divergence. As cross entropy is just the negated logarithm of likelihood, minimizing cross entropy means maximizing likelihood, and thus, cross entropy is widely used now for optimization in machine learning. K-L divergence also stands independently as a widely used metric for measuring the difference between two distributions, and it appears in many theoretical results. As examples, the definition of mutual information between two variables is a K-L divergence, and the absolute difference of K-L divergences with respect to a third distribution was recently used to define a family of metrics [5].
Recently, order invariance for inequalities between means was presented [6]. Since likelihood is the weighted geometric mean of the representing distribution weighted by the data distribution, likelihood and thus cross entropy will share the order invariance properties too. Order invariance properties are characterized by the stochastic order between weights or data distributions [7]. Here, we further develop this idea in order to compare representations of different data by the same distribution, and we extend it to a new order, K-L dominance, that allows us to define inequalities between representations of the same data by different distributions. Finally, we establish the connection between the two orders. We also give numerical examples to illustrate the theoretical results.
This paper is organized as follows. After this introduction, in Section 2, we recapitulate the meaning of cross entropy and its relationship to likelihood and K-L divergence, and we obtain the first theoretical results based on the rearrangement inequality. Next, in Section 3, we apply the invariance results for weighted means to cross entropy. In Section 4, we introduce a new order, K-L dominance, that allows us to establish inequalities between K-L divergences and give the relationship between stochastic and K-L order. Finally, in Section 5, we present our conclusions and future work.

2. Cross Entropy as a Measure of Goodness of Representation

Consider two probability distributions, { α k } k = 1 n , { x k } k = 1 n , for all k α k 0 , x k > 0 , k α k = k x k = 1 (when not explicit, the sum limits will be understood to be between 1 and n). Cross entropy C E ( α k , x k ) (to avoid cluttered notation we will write C E ( α k , x k ) instead of C E ( { α k } , { x k } ) ) is defined as C E ( α k , x k ) = k α k log 1 x k = k α k log x k and can be written as C E ( α k , x k ) = H ( α k ) + D KL ( α k , x k ) (similar notation to C E ( α k , x k ) ), where H ( α k ) = k α k log α k is the entropy of distribution { α k } , and D KL ( α k , x k ) = k α k log α k x k 0 is the Kullback–Leibler divergence of { α k } and { x k } distributions. It represents the average length of code per symbol needed to represent { α k } using { x k } . In the context of coding, we do not let any component of { x k } be zero, as the corresponding component of { α k } would not be coded; thus, D KL ( α k , x k ) is well defined. The minimum code length to code { α k } is obtained by taking { α k } { x k } , because D KL ( α k , α k ) = 0 . Minimum code length will be between H ( α k ) and H ( α k ) + 1 bit (Huffman coding, [8]).

Cross Entropy and Likelihood

Observe that cross entropy is the negated logarithm of likelihood. Let us suppose a distribution with n states, which are guessed to be { x 1 , x 2 , , x n } , and a relative frequency of realizations { α 1 , α 2 , , α n } of the different states. That is, we guess that the distribution { x k } represents the collected data { α k } . Likelihood is then used as a measure of how well the distribution represents the data. The likelihood L ( α k , x k ) of { x k } is the true distribution of the data with relative frequency { α k } , and L = x 1 α 1 x 2 α 2 x n α n = Π k x k α k . Observe that L ( α k , x k ) is the weighted geometric mean of { x 1 , x 2 , , x n } with weights { α 1 , α 2 , , α n } . The likelihood represents the probability of { x k } being the true distribution of the known data { α k } . As C E ( α k , x k ) = log L ( α k , x k ) , maximizing the likelihood is equivalent to minimizing the cross entropy.
Now, let us suppose we take an arbitrary { x k } distribution to code { α k } . Without loss of generality, we will consider two or more sequences to be equally ordered, or in the same order, when, by the same permutation of indexes, we can make it such that all the sequences are increasing. We can state the following theorem.
Theorem 1.
Consider the distribution { α k } and the sequence of strictly positive numbers { x 1 , x 2 , , x n } , k x k = 1 . When { x 1 , x 2 , , x n } is permuted so that { x k } is ordered in the same order as { α k } , then the cross entropy C E ( α k , x k ) takes a global minimum, and, correspondingly, the likelihood L ( α k , x k ) takes a global maximum.
Proof. 
Using the rearrangement inequality [6,9] and the definition of cross entropy, the minimum of the sum of the product Π k α k log 1 x k is realized when the two sequences are ordered inversely to each other. That means the { log x k } sequence will be ordered as the { α k } sequence, and, by the monotonicity of the logarithmic function, the { x k } sequence will be ordered as the { α k } sequence. □
As { α k } , and thus H ( α k ) , is fixed, D KL ( α k , x k ) also takes the global minimum when { x k } is ordered as { α k } . Also, when there are elements repeated in { α k } , there can be non-ordered { x k } distributions with the same minimum cross entropy value. The extreme case is when α k = 1 / n for all k, and all orderings of { x k } will give the same result.
Observe that by Theorem 1, if we have a series of distributions { α t , k } t = 1 m that are all ordered in the same order, which we can consider to be increasing, and want to represent them with a single distribution { x k } —to be chosen by permuting the elements in { x 1 , x 2 , , x n } —then the best distribution will be the one for which we select the elements in { x 1 , x 2 , , x n } in increasing order too.
Example 1.
For any increasing sequence { α k } , we have that
1. 
C E ( { α k } , { 2 / 9 , 3 / 9 , 4 / 9 } ) C E ( { α k } , { 4 / 9 , 3 / 9 , 2 / 9 } ) ,
2. 
C E ( { α k } , { 2 / 9 , 3 / 9 , 4 / 9 } ) C E ( { α k } , { 4 / 9 , 2 / 9 , 3 / 9 } ) , and so on.
However, what happens when we have to choose between two distributions, { x k } and { x k } , to represent a set of increasing distributions, { α t , k } t = 1 m ? Can we find which one is better? Section 4 provides some answers to this. Before answering this question, we introduce the concept of stochastic dominance in Section 3.

3. First Stochastic Order Dominance

When for all increasing distributions { x k } : k α k x k k α k x k , we say that { α k } stochastically dominates { α k } , and we denote this relation as { α k } s t { α k } [7]. This is a partial order on the set of distributions.
The necessary and sufficient conditions for a sequence { α k } to stochastically dominate another one { α k } are given in the following theorem [6].
Theorem 2.
Given that the sequences { α k } and { α k } are positive and add up to 1, the following conditions are equivalent:
(1) 
The sequence { α k } stochastically dominates the sequence { α k }
(2) 
α 1 α 1 α 1 + α 2 α 1 + α 2 α 1 + α 2 + + α n = α 1 + α 2 + + α n
(3) 
α n α n α n + α n 1 α n + α n 1 α n + α n 1 + + α 1 = α n + α n 1 + + α 1
(4) 
For all increasing sequences { x k } and for any f ( x ) strictly monotonous function, f 1 k α k f ( x k ) f 1 k α k f ( x k ) (quasi-arithmetic, or Kolmogorov, mean), whenever f ( x ) is applicable.
(5) 
There exists a strictly monotonous function f ( x ) such that for all increasing sequences { x k } :
f 1 k α k f ( x k ) f 1 k α k f ( x k )
(quasi-arithmetic, or Kolmogorov, mean), whenever f ( x ) is applicable.
A sufficient condition for { α k } s t { α k } is given by the following theorem [6,10].
Theorem 3.
Given the distributions { α k } , { α k } , a sufficient condition for { α k } s t { α k } is that whenever i j , then α i α j α i α j .
By Theorem 3, if { α k } is increasing, then { α k } s t { 1 / n } , where { 1 / n } is the uniform distribution; if { α k } is decreasing, then { 1 / n } s t { α k } ; if { α k } is increasing and { α k } is decreasing, then { α k } s t { α k } .
Observe now that, by condition (5) of Theorem 2, if Equation (3) applies for one strictly monotonous function, then, for the equivalence between conditions (4) and (5), it applies for any other strictly monotonous function. Then, by the definition of cross-entropy,
Theorem 4.
Given two distributions { α k } and { α k } , { α k } s t { α k } if and only if for all increasing distributions { x k } , C E ( α k , x k ) C E ( α k , x k ) (or L ( α k , x k ) L ( α k , x k ) ).
Theorem 4 means that, if { α k } s t { α k } , then, taking all possible increasing representations { x k } , the code length for { α k } will be shorter than for { α k } , and the likelihood for { α k } will be bigger than the likelihood for { α k } . Observe also that if { α k } is an increasing sequence, then { α k } s t { 1 / n } ; thus, the representation of { α k } by an increasing distribution { x k } will be shorter than the representation of uniform distribution { 1 / n } , and the likelihood of { α k } will be bigger than that of { 1 / n } .
Corollary 1.
For any increasing distributions { x k } and { α k } , we have that C E ( α k , x k ) C E ( 1 / n , x k ) (or L ( α k , x k ) L ( 1 / n , x k ) ) .
Indeed, the above result is valid too if, by an index permutation, we can bring { x k } , { α k } to be in the same order. Thus,
Corollary 2.
For any distributions { x k } and { α k } that are ordered equally, we have that C E ( α k , x k ) C E ( 1 / n , x k ) .
Now, suppose { α k } is decreasing; then, { 1 / n } s t { α k } . Also, for any distribution { x k } , C E ( 1 / n , x k ) C E ( 1 / n , 1 / n ) = log n , and thus, the following corollary,
Corollary 3.
For any increasing distribution { x k } and decreasing distribution { α k } , we have that log n C E ( 1 / n , x k ) C E ( α k , x k ) .
Also, by reordering the indexes, Corollary 3 can be extended to
Corollary 4.
For any distributions { x k } and { α k } that are inversely ordered to each other, we have that log n C E ( 1 / n , x k ) C E ( α k , x k ) .
Observe now that if { α k } is increasing and { α k } is decreasing, then, { α k } s t { α k } , and thus,
Corollary 5.
For any increasing distribution { α k } and decreasing distribution { α k } , we have that for any increasing distribution { x k } , C E ( α k , x k ) C E ( α k , x k ) .
Corollary 5 is also a direct consequence of Corollaries 1 and 3.
Example 2.
We have:
1. 
{ 1 / 4 , 1 / 4 , 1 / 2 } s t { 1 / 3 , 1 / 3 , 1 / 3 }
2. 
{ 1 / 2 , 1 / 4 , 1 / 4 } s t { 1 / 3 , 1 / 3 , 1 / 3 }
3. 
{ 1 / 3 , 1 / 3 , 1 / 3 } s t { 1 / 2 , 1 / 4 , 1 / 4 }
4. 
{ 1 / 2 , 1 / 4 , 1 / 4 } s t { 1 / 5 , 3 / 5 , 1 / 5 }
5. 
{ 1 / 5 , 3 / 5 , 1 / 5 } s t { 1 / 2 , 1 / 4 , 1 / 4 }
In Figure 1, we plot the sums in Equation (1) corresponding to relation 1 and relations 2–3, 4–5 in Example 2. Observe that the sums in Equation (1) correspond to the cumulative distribution function (cdf). A first-order stochastic dominance implies that there is no crossing of the respective cdf’s.

Discussion

We show above how the interpretation of cross entropy as a weighted mean allows for obtaining interesting order invariance results for the cross entropy of two data distributions represented by the same distribution (correspondingly for likelihood). Since, in our results, the distribution of the data is variable while the representative distribution is fixed, the resulting cross entropy depends on both entropy and K-L divergence. In the next section, we consider the data fixed and the representative distribution variable; thus, the comparison of cross entropies of a given data distribution for two representative distributions will be equivalent to the comparison of their K-L divergences.

4. A New Partial Order: K-L Dominance

We extend here the results of Section 2. We first define a new partial order between distributions and call it K-L dominance.
Definition 1.
K-L dominance. We say that distribution { x k } K-L-dominates distribution { x k } , written symbolically as { x k } KL { x k } , when, for all increasing distributions { α k } , D KL ( α k , x k ) D KL ( α k , x k ) .
Observe that { x k } KL { x k } { x k } s t { x k } , because, in general, D KL ( α k , x k ) D KL ( α k , x k ) D KL ( x k , α k ) D KL ( x k , α k ) .
From the definition of cross entropy, { x k } KL { x k } is equivalent to—for all increasing { α k } C E ( α k , x k ) C E ( α k , x k ) , and thus, for all increasing { α k } sequences, coding them with { x k } will always generate a shorter codification than coding them with { x k } .
The following theorem holds:
Theorem 5.
A necessary and sufficient condition for distributions { x k } , { x k } to hold { x k } KL { x k } (equivalent to—for all increasing distributions { α k } D KL ( α k , x k ) D KL ( α k , x k ) , or C E ( α k , x k ) C E ( α k , x k ) ) or L ( α k , x k ) L ( α k , x k ) is the following:
x n x n x n x n 1 x n x n 1 x n x n 1 x 1 x n x n 1 x 1
Observe that Theorem 1 is a particular case of Theorem 5, as a sequence rearranged in increasing order will always fill the second members of inequalities in Equation (4) with respect to all other rearrangements. Thus, an increasing distribution { x k } K-L-dominates all the distributions obtained by permuting the values of { x k } . Observe also that Equation (4) condition makes the KL order reflexive, antisymmetric, and transitive, and thus, it is a partial order. To prove Theorem 5, let us consider first the following lemma:
Lemma 1.
Consider the sequences of n numbers { w k } and { y k } . Then, conditions (1) and (2) are equivalent:
(1) 
for any sequence of n positive numbers in increasing order { z k } , the following inequality holds:
k = 1 n w k z k k = 1 n y k z k ,
(2) 
(2) the following inequalities hold:
w n y n w n + w n 1 y n + y n 1 w n + + w 2 y n + + y 2
w 1 + + w n 1 + w n y 1 + + y n 1 + y n .
Proof. 
That ( 1 ) ( 2 ) is immediate, considering, respectively, the z k sequences { 0 , 0 , , 1 } , { 0 , 0 , , 1 , 1 } , , { 0 , 1 , , 1 } , { 1 , 1 , , 1 } . Let us see now that ( 2 ) ( 1 ) : Define for 1 n , A k = j = 1 k y n j + 1 , A k = j = 1 k w n j + 1 , and A 0 = A 0 = 0 , z 0 = 0 . Observe that condition (2) is equivalent to—for all k A k A k 0 . Observe also that the { z k } sequence augmented with z 0 is still increasing. We have
k = 1 n y n k + 1 z n k + 1 k = 1 n w n k + 1 z n k + 1 = k = 1 n ( y n k + 1 w n k + 1 ) z n k + 1 = k = 1 n ( A k A k 1 A k + A k 1 ) z n k + 1 = k = 1 n ( A k A k ) z n k + 1 k = 1 n ( A k 1 A k 1 ) z n k + 1 = k = 1 n ( A k A k ) z n k + 1 k = 0 n 1 ( A k A k ) z n k = k = 1 n ( A k A k ) z n k + 1 k = 1 n ( A k A k ) z n k = k = 1 n ( A k A k ) ( z n k + 1 z n k ) 0 ,
as, by hypothesis for all k, A k A k 0 , and { z k } is an increasing sequence. □
The proof of Theorem 5 follows immediately from applying Lemma 1 to the inequality k α k log x k k α k log x k .
The following corollary follows from observing that, by taking { α k } { x k } , then, D KL ( x k , x k ) = 0 < D KL ( x k , x k ) , { x k } { x k } .
Corollary 6.
An increasing distribution { x k } cannot be K-L dominated by any other distribution (except trivially by itself). In particular, the uniform distribution { x k } { 1 / n } does not dominate any other increasing distribution.
Observe now that the product 1 / n n is always bigger than or equal Π k x k , as the maximum value for Π k x k subjected to the condition k x k = 1 is, for all k, x k = 1 / n . Thus, from the last inequality in Equation (4), we have the following corollary,
Corollary 7.
For all distributions { x k } , { x k } { 1 / n } , we have that { x k } KL { 1 / n } .
Observe that Corollary 7 is also a particular case of Corollary 6, considering { 1 / n } as an increasing distribution.
Corollary 8.
The uniform distribution { 1 / n } K-L-dominates all decreasing distributions.
Proof. 
For all { α k } increasing and { x k } decreasing, we apply Corollary 4, and we have C E ( α k , 1 / n ) = log n C E ( α k , x k ) . ☐
Example 3.
Some examples are:
1. 
{ 1 / 3 , 1 / 3 , 1 / 3 } KL { 4 / 9 , 3 / 9 , 2 / 9 }
2. 
{ 1 / 3 , 1 / 3 , 1 / 3 } KL { 7 / 9 , 1 / 9 , 1 / 9 }
3. 
{ 1 / 3 , 1 / 3 , 1 / 3 } KL { 1 / 9 , 1 / 9 , 7 / 9 }
4. 
{ 1 / 3 , 1 / 3 , 1 / 3 } KL { 4 / 9 , 2 / 9 , 3 / 9 }
5. 
{ 1 / 3 , 1 / 3 , 1 / 3 } KL { 3 / 9 , 2 / 9 , 4 / 9 }
6. 
{ 1 / 4 , 1 / 4 , 1 / 2 } KL { 1 / 4 , 1 / 2 , 1 / 4 }
7. 
{ 1 / 2 , 1 / 4 , 1 / 4 } KL { 5 / 8 , 2 / 8 , 1 / 8 }
The first and second relation are examples of Corollary 8. The third is an example of Corollary 6. The sixth is an example of Theorem 1. The fourth and fifth relations mean that the uniform distribution can both K-L dominate and non-dominate non-monotonous sequences, and we have to look for each case at the Condition (4) of Theorem 5. The seventh relation tells us that a decreasing sequence can dominate another decreasing sequence. In Figure 2, we plot the logarithm of products in Equation (4) for relations 1, 3, 6, and 7, respectively. If a sequence K-L-dominates another one, the plots do not cross, and its plot appears over that of the second sequence. Inversely, no dominance means that the plots will cross.

4.1. Discussion

Remember that the equivalence between cross entropy and likelihood means that the likelihood that { x k } is the true distribution, given the data distribution { α k } , and the average length of code using { x k } to represent { α k } are related by the second quantity being the negated logarithm of the first. If you have one distribution { α k } and have the choice to use { x 1 , k } or { x 2 , k } to represent it, you want to choose the one with the minimum length of code, or, alternatively, the one with maximum likelihood. In both cases, the best is the one that gives minimum Kullback–Leibler divergence. Suppose now you want to represent data { α k } and have two alternative distributions to represent it, either { x 1 , k } { 1 / n } , the uniform distribution, or { x 2 , k } , which is ordered in the same order as { α k } . To fix ideas and without loss of generality, let us consider that both { α k } and { x 2 , k } are in increasing order. It would seem reasonable to represent the data { α k } , which is increasing, as distribution { x 2 , k } , which is increasing too; thus, the order is preserved. However, that would work as long as D KL ( α k , x 2 , k ) < D KL ( α k , 1 / n ) ; otherwise, we should choose { x 1 , k } , the uniform distribution. This is clear from corollary 7, which tells us that, for any sequence, particularly { x 2 , k } , there will always be increasing { α k } sequences such that D KL ( α k , 1 / n ) < D KL ( α k , x 2 , k ) . Intuitively, if { α k } has a high increasing gradient, then we would choose { x 2 , k } , the increasing distribution to represent it, but if { α k } becomes smoother, then we would choose { x 1 , k } , the uniform distribution.
As an example, suppose the two distributions { 0.3 , 0.7 } , { 0.4 , 0.6 } have to be represented by one of the distributions { 0.33 , 0.67 } , { 0.5 , 0.5 } . To preserve the order, we would choose { 0.33 , 0.67 } for both. It is intuitively a clear choice of { 0.3 , 0.7 } (and a correct one, as it has a K-L divergence, half the one for { 0.5 , 0.5 } ), but we might have some doubts for { 0.4 , 0.6 } , where, actually, the K-L divergence is only slightly smaller for { 0.5 , 0.5 } . To represent distributions { 0.43 , 0.57 } , { 0.45 , 0.55 } , it is intuitively more clear to select the uniform distribution { 0.5 , 0.5 } rather than { 0.33 , 0.67 } , even if order is lost. In fact, the K-L divergence from both distributions to { 0.33 , 0.67 } is one and two orders of magnitude greater, respectively, than to the uniform distribution { 0.5 , 0.5 } .

4.2. Relationship between K-L Dominance and First Stochastic Dominance Orders

Summarizing K-L dominance and first stochastic dominance orders, we have
  • { α k } s t { α k } x k increasing C E ( α k , x k ) C E ( α k , x k )
  • { x k } KL { x k } α k increasing C E ( α k , x k ) C E ( α k , x k )
Is there a relationship between the two orders? When Π k x k = Π k x k or, equivalently, k log x k = k log x k (observe that Theorem 1 is a particular case), we can find a necessary and sufficient condition for { x k } K L { x k } .
Theorem 6.
Given distributions { x k } , { x k } , and Π k x k = Π k x k , then the following conditions are equivalent:
(1) 
{ x k } KL { x k }
(2) 
{ log x k k log x k } s t { log x k k log x k }
Proof. 
Observe first that { log x k k log x k } , { log x k k log x k } are also distributions (we can indeed use the simpler expressions { log x k k log x k } , { log x k k log x k } , but, here, we have left the negated ones to remind the reader that both numerator and denominator are negative), because they are positive sequences adding to 1. Observe also that when Π k x k = Π k x k , taking the logarithms in Equation (4) and changing the sign (which changes the direction of inequalities), we obtain condition (3) in Theorem 2, which is necessary and sufficient for { log x k k log x k } s t { log x k k log x k } . □
Example 4.
Consider the distributions { 1 / 4 , 1 / 4 , 1 / 2 } , { 1 / 4 , 1 / 2 , 1 / 4 } and their normalized negated logarithms { 2 / 5 , 2 / 5 , 1 / 5 } , { 2 / 5 , 1 / 5 , 2 / 5 } , respectively. We have that { 1 / 4 , 1 / 4 , 1 / 2 } K L { 1 / 4 , 1 / 2 , 1 / 4 } and { 2 / 5 , 1 / 5 , 2 / 5 } s t { 2 / 5 , 2 / 5 , 1 / 5 } .

5. Conclusions and Future Work

In this paper, we present new inequalities for cross entropy and Kullback–Leibler divergence. As cross entropy is the negated logarithm of likelihood, the inequalities also hold for likelihood, changing the sense of the inequality. First, we applied to cross entropy the rearrangement inequality and recent stochastic order invariance results for Kolmogorov weighted means, as likelihood is a weighted geometric mean. Then, we introduced another partial order, K-L dominance, that applies directly to K-L divergences, and we give the relationship between both orders.
In data retrieval, exploration, analysis, and visualization, sorting some data objects into an ordered list or display is a common task performed by users. In the past, the benefit of having a sorted list has been typically articulated qualitatively. We plan to use the theorems presented in this paper to help establish an information-theoretic explanation about the cost–benefit of ordering in data retrieval, exploration, analysis, and visualization. We will also investigate the application to machine learning. We will study whether K-L-dominance order supports invariance properties, in the same way as first stochastic order supports the order invariance for weighted means, and we will look for sufficient conditions for K-L-dominance order, similar to the one in Theorem 3 for first stochastic order. Finally, we will investigate extensions of K-L dominance order: for instance, when { α k } { f ( k ) k f ( k ) } , f ( x ) increasing and concave/convex, in the same way as the extensions of first stochastic order.

Author Contributions

Conceptualization, M.S. and M.C.; validation, J.P. and A.B.; writing, original draft preparation, M.S.; review and editing, all authors.

Funding

Mateu Sbert acknowledges the funding of National Natural Science Foundation of China under grants No.61471261 and No.61771335, and by grant TIN2016-75866-C3-3-R from Spanish Government, Jordi Poch and Anton Bardera acknowledge the funding of TIN2016-75866-C3-3-R from Spanish Government.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Hu, B.G.; He, R.; Yuan, X.T. Information-Theoretic Measures for Objective Evaluation of Classifications. Acta Autom. Sin. 2012, 38, 1170–1182. [Google Scholar] [CrossRef]
  2. Chen, M.; Golan, A. What may visualization processes optimize? IEEE Trans. Vis. Comput. Graph. 2016, 22, 2619–2632. [Google Scholar] [CrossRef] [PubMed]
  3. Chen, M.; Feixas, M.; Viola, I.; Bardera, A.; Sbert, M.; Shen, H.W. Information Theory Tools for Visualization; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  4. Sbert, M.; Feixas, M.; Rigau, J.; Chover, M.; Viola, I. Information Theory Tools for Computer Graphics; Synthesis Lectures on Computer Graphics and Animation; Morgan & Claypool Publishers: San Rafael, CA, USA, 2009; Volume 4, pp. 1–153. [Google Scholar]
  5. Galas, D.; Dewey, G.; Kunert-Graf, J.; Sakhanenko, N. Expansion of the Kullback-Leibler Divergence, and a New Class of Information Metrics. Axioms 2017, 6, 8. [Google Scholar] [CrossRef]
  6. Sbert, M.; Poch, J. A necessary and sufficient condition for the inequality of generalized weighted means. J. Inequal. Appl. 2016, 2016, 292. [Google Scholar] [CrossRef]
  7. Shaked, M.; Shanthikumar, G. Stochastic Orders; Springer: New York, NY, USA, 2007. [Google Scholar]
  8. Cover, T.M.; Thomas, J.A. Elements of Information Theory; Wiley–Interscience: New York, NY, USA, 1991. [Google Scholar]
  9. Hardy, G.; Littlewood, J.; Pólya, G. Inequalities; Cambridge University Press: Cambridge, UK, 1952. [Google Scholar]
  10. Belzunce, F.; Martinez-Riquelme, C.; Mulero, J. An Introduction to Stochastic Orders; Academic Press: Cambridge, MA, USA, 2016. [Google Scholar]
Figure 1. Plotting the sums in Equation (1) for relations 1–5 in Example 2.
Figure 1. Plotting the sums in Equation (1) for relations 1–5 in Example 2.
Entropy 20 00959 g001
Figure 2. Plotting the logarithm of products in Equation (4) for relations 1, 3, 6, and 7 in Example 3.
Figure 2. Plotting the logarithm of products in Equation (4) for relations 1, 3, 6, and 7 in Example 3.
Entropy 20 00959 g002

Share and Cite

MDPI and ACS Style

Sbert, M.; Chen, M.; Poch, J.; Bardera, A. Some Order Preserving Inequalities for Cross Entropy and Kullback–Leibler Divergence. Entropy 2018, 20, 959. https://doi.org/10.3390/e20120959

AMA Style

Sbert M, Chen M, Poch J, Bardera A. Some Order Preserving Inequalities for Cross Entropy and Kullback–Leibler Divergence. Entropy. 2018; 20(12):959. https://doi.org/10.3390/e20120959

Chicago/Turabian Style

Sbert, Mateu, Min Chen, Jordi Poch, and Anton Bardera. 2018. "Some Order Preserving Inequalities for Cross Entropy and Kullback–Leibler Divergence" Entropy 20, no. 12: 959. https://doi.org/10.3390/e20120959

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop