Next Article in Journal
Incorporating Heterogeneous Features into the Random Subspace Method for Bearing Fault Diagnosis
Previous Article in Journal
Contextuality in Collective Intelligence: Not There Yet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Entropy and the Maximum Entropy Principle of Uncertain Variables

School of Mathematics and Statistics, Anyang Normal University, Anyang 455000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2023, 25(8), 1195; https://doi.org/10.3390/e25081195
Submission received: 19 June 2023 / Revised: 6 August 2023 / Accepted: 10 August 2023 / Published: 11 August 2023
(This article belongs to the Section Entropy Reviews)

Abstract

:
A new variance formula is developed using the generalized inverse of an increasing function. Based on the variance formula, a new entropy formula for any uncertain variable is provided. Most of the entropy formulas in the literature are special cases of the new entropy formula. Using the new entropy formula, the maximum entropy distribution for unimodel entropy of uncertain variables is provided without using the Euler–Lagrange equation.

1. Introduction

The concept of entropy was first introduced by Rudolf Clausius in the mid-19th century as a measure of the disorder or randomness in a system. In 1948, Shannon [1] inaugurated information entropy in information theory, which measures the unpredictability of the realization of a random variable. Subsequently, information entropy developed into an essential tool in information theory (see Gray [2]). Entropy was introduced into dynamical systems by Kolmogorov [3] in 1958 to investigate the conjugacy problems of the Bernoulli shift. In 2009, Liu [4] proposed logarithmic entropy in uncertainty theory to quantify the difficulty in forecasting the outcome of an uncertain variable. In 2012, Knuth et al. [5] gave an interesting application of maximal joint entropy in automated learning machines. Currently, entropy is widely used in various fields, such as probability theory, ergodic theory, dynamical systems, machine learning and data analysis, uncertainty theory, and others. We refer the readers to Martin and England [6], Walters [7], Einsiedler and Thomas [8], and Liu [9] for a detailed introduction to the entropy theory in these fields.
The maximum entropy principle was first expounded by Jaynes [10,11], where he gave a natural criterion of choice, i.e., from the set of all probability distributions compatible with several mean values of several random variables, choose the one that maximizes Shannon entropy. From then on, the methods of maximum entropy became widely applied tools for constructing the probability distribution in statistical inference [12], decision-making [13,14], communication theory [15], time-series analysis [16], and reliability theory [17,18]. In 2017, Abbas et al. [19] made an elaborate comparison between the methods of maximum entropy and the maximum log-probability from the point of the Kullback–Leibler divergence. In this paper, we will discuss the entropy and the maximum entropy principle in uncertainty theory.
Uncertainty theory, founded by Liu [20] in 2007 and refined in 2015 [9], is a branch of axiomatic mathematics that deals with human’s belief degree based on normality, duality, sub-additivity, and product axioms. Within the past few years, it has attracted numerous researchers. Peng and Iwamura [21], Liu and Lio [22] provided a sufficient and necessary characterization of an uncertainty distribution. Based on the inverse uncertainty distribution, Yao [23] introduced a new approach to computing the variance. Subsequently, Sheng and Kar [24] extended Yao’s result to the moments of the uncertain variable.
After the introduction of entropy in uncertainty theory, the entropy and maximum entropy principle were researched by numerous experts. Inspired by the maximum entropy principles in information theory and statistics, Chen and Dai [25] explored the maximum entropy principle of logarithmic entropy. According to the maximum entropy principle, using the Lagrange multipliers method, Chen and Dai [25] proved that normal uncertainty distribution is the maximal entropy distribution under the prescribed constraints on the expected value and variance. In addition, using Fubini’s theorem, Dai and Chen [26] built an entropy formula for regular uncertain variables (see Definition 2). Subsequently, there have been many extensions of entropy, the entropy formula, and the corresponding maximum entropy distribution. Yao et al. [27] proposed sine entropy. Yao et al. [27] provided an entropy formula for regular uncertain variables, and the authors also obtained the maximum entropy distribution for sine entropy. Tang and Gao [28] proposed triangular entropy and provided an entropy formula for regular uncertain variables. Ning et al. [29] applied triangular entropy to portfolio selection. Subsequently, Dai [30] suggested quadratic entropy and provided an entropy formula for regular uncertain variables. In addition, Dai [30] found the maximum entropy distribution for quadratic entropy. Here, we point out, all the entropy formulas in Dai and Chen [26], Yao et al. [27], Tang and Gao [28], and Dai [30] are based on Fubini’s theorem. Subsequently, Ma [31] proposed unimodal entropy of uncertain variables and provided an entropy formula for regular uncertain variables. Ma [31] also found the maximum entropy distribution for unimodal entropy for regular uncertain variables. Nevertheless, limited by their methods, all the entropy formulas in Dai and Chen [26], Yao et al. [27], Tang and Gao [28], Dai [30], and Ma [31] are only established for regular uncertain variables. We know most uncertain variables are not regular. Therefore, it has some significance not only in theory but also in practice to prove the entropy formulas of Dai and Chen [26], Yao et al. [27], Tang and Gao [28], Dai [30], and Ma [31] are still true for any uncertain variable. In this paper, we will provide an entropy formula that is true for any uncertain variable by using the generalized inverse of an increasing function. We emphasize that the entropy formulas of Dai and Chen [26], Yao et al. [27], Tang and Gao [28], Dai [30], and Ma [31] are special cases of our entropy formula, and our method differs completely from those methods used by Dai and Chen [26], Yao et al. [27], Dai [30], and Ma [31].
Furthermore, we point out that all the maximum entropy distributions of Chen and Dai [25], Yao et al. [27], and Dai [30] were obtained based on the Lagrange multipliers method. Using our entropy formula, we will obtain the maximum entropy distribution without utilizing the Lagrange multipliers method. We also emphasize that the maximum entropy distributions of Chen and Dai [25], Yao et al. [27], and Dai [30] are special cases of ours.
The rest of the paper is arranged as follows. In Section 2, we first collect some necessary facts in uncertainty theory and some requisite results of the generalized inverse of an increasing function. Then, we review the entropy, entropy formulas, and the maximum entropy principles in uncertainty theory. In Section 3, we present our entropy formula and our maximum entropy distribution. We also demonstrate some applications of our entropy formula and maximum entropy distribution. In Section 4, we prove our entropy formula and derive the maximum entropy distribution of unimodel entropy. In Section 5, we verify that the maximum entropy distributions of logarithmic entropy in Theorem 2, sine entropy in Theorem 5, and quadratic entropy in Theorem 8 are special cases of our maximum entropy distributions. Finally, a concise discussion and further research are given in Section 6.

2. Preliminaries

In this section, first, we will collect the necessary facts in uncertainty theory. Second, we will recall some requisite results of the generalized inverse of an increasing function. Finally, we review the entropy, entropy formulas, and the maximum entropy principles in uncertainty theory.

2.1. Some Primary Facts in Uncertainty Theory

Let Γ be a non-empty set and suppose L is a σ -algebra over Γ . Each element A in L is called an event. To handle belief degree reasonably, Liu [20] introduced the uncertain measure.
Definition 1.
A function M : L [ 0 , 1 ] is called an uncertain measure provided M satisfies the following three axioms:
1.
Normality: M { Γ } = 1 ;
2.
Duality: M { A } + M { A c } = 1 for each event A;
3.
Subadditivity: for each sequence of events A 1 , A 2 , , one gets
M i = 1 A i i = 1 M { A i } .
The triplet ( Γ , L , M ) is called an uncertainty space. A measurable function X : ( Γ , L , M ) R is called an uncertain variable. Let X be an uncertain variable; the function Φ ( x ) = M { X x } is called the uncertainty distribution of X.
Definition 2
([32]). An uncertainty distribution Φ is called regular provided it is strictly increasing and continuous over { x : 0 < Φ ( x ) < 1 } , and
lim x Φ ( x ) = 0 , lim x + Φ ( x ) = 1 .
An uncertain variable is called regular if its uncertainty distribution is regular.
Obviously, a regular uncertainty distribution Φ is invertible. Now, we recall two regular uncertainty distributions that are most used in uncertainty theory. The first is the normal uncertainty distribution N ( e , σ ) [32], which is defined by
Φ ( x ) = 1 + exp π ( e x ) 3 σ 1 , x R .
Here, e and σ are real numbers with σ > 0 . We call an uncertain variable X normal provided its uncertainty distribution is normal and write X N ( e , σ ) (Figure 1).
The second is the linear uncertainty distribution L ( a , b ) [32], which is defined by
Φ ( x ) = 0 , x a x a b a , 0 < x b 1 , b < x
where a and b are real numbers with a < b (Figure 2).
Liu [20] suggested the expectation to describe the mean of an uncertain variable.
Definition 3
([20]). Let X be an uncertain variable. The expectation of X is defined by
E [ X ] = 0 M { X x } d x 0 M { X x } d x
if one or two of the integrals above are finite.
Let X be an uncertain variable with finite expectation e; the variance of X is defined by
V [ X ] = E [ ( X e ) 2 ] .
Given an uncertainty variable X with uncertainty distribution Φ and finite expectation e, the variance of X can be computed by (see Section 2.6 of Chapter 2 in [9])
V [ X ] = 0 + ( 1 Φ ( e + x ) + Φ ( e x ) ) d x .
Yao [23] provided a nice formula for calculating the variance by inverse uncertainty distribution.
Theorem 1
(Yao [23]). Assume that X is a regular uncertain variable with uncertainty distribution Φ . If E [ X ] is finite, then
V [ X ] = 0 1 ( Φ 1 ( α ) E [ X ] ) 2 d α .
We will prove the variance formula of Yao is true for any uncertain variable in Theorem 9, which is one of our main tools for establishing our entropy formula and maximum entropy principle.

2.2. Necessary Results of the Generalized Inverse of an Increasing Function

To present our results, we need to review the generalized inverse of an increasing function, which is a basic tool in statistics; see Embrechts and Hofert [33], Resnick [34], Kampke-Radermacher [35]. Here, we follow the definition of Embrechts and Hofert [33].
Definition 4
([33]). Let Φ : R [ 0 , 1 ] be an increasing function. The generalized inverse of Φ is defined by
Φ ( α ) = inf { x R : Φ ( x ) α } , α ( 0 , 1 ] ,
with the convention that inf { } = + , where ∅ is the empty set.
Specifically, we define Φ ( 0 ) = sup { x R : Φ ( x ) = 0 } with the convention that sup { } = .
Remark 1.
To stress the difference between the generalized inverse and standard inverse, we denote the generalized inverse of an increasing function Φ by Φ . If Φ is strictly increasing and continuous, Φ is just the standard inverse of Φ .
We also point out that, by the above definition, Φ ( 0 ) may be and Φ ( 1 ) may be +. Here is an example used frequently in uncertainty theory.
Example 1.
Let Φ be a normal uncertainty distribution defined by (1). Then Φ is:
Φ ( α ) = , α = 0 e 3 σ π ln 1 α α , 0 < α < 1 + , α = 1 .
We also need to introduce the generalized inverse of a left-continuous increasing function; see Appendix B in Iritani and Kuga [36].
Definition 5
([36]). Let G be a left-continuous and increasing function over ( 0 , 1 ) . The generalized inverse of G is defined by
G 1 ( x ) = sup y : G ( y ) x
for all real numbers x.
The succeeding interesting fact is that the generalized inverse of the generalized inverse for a right-continuous increasing function is just itself. To the best of our knowledge, this result was proved first by Iritani and Kuga [36] and was refined by Kampke-Radermacher [35].
Lemma 1
([35]). Let Φ : R [ 0 , 1 ] be a right-continuous and increasing function; G is the generalized inverse of Φ defined by (3), i.e., G = Φ . Then
G 1 ( x ) = Φ ( x )
for all x R , where G 1 ( x ) is the generalized inverse of G defined by (4).

2.3. Entropy Formula and the Principle of Maximum Entropy

To quantify the difficulty in predicting the outcomes of an uncertain variable, Liu [4] proposed logarithm entropy.
Definition 6
([4]). Let X be an uncertain variable with uncertainty distribution Φ . The logarithm entropy of X is defined by
H L [ X ] = L Φ ( x ) d x .
Here, L ( y ) = y ln y ( 1 y ) ln ( 1 y ) for y ( 0 , 1 ) and L ( 0 ) = L ( 1 ) = 0 .
Using the Lagrange multipliers method, Chen and Dai [25] obtained the maximum entropy distribution for the logarithmic entropy.
Theorem 2
([25]). Suppose X is an uncertain variable with finite expectation e and variance σ 2 . Then,
H L [ X ] σ π 3 ,
and the equality holds if and only if X is a normal uncertain variable N ( e , σ ) .
In addition, using Fubini’s theorem, Dai and Chen [26] attained an interesting entropy formula for regular uncertain variables.
Theorem 3
([26]). Suppose X is a regular uncertain variable with uncertainty distribution Φ . If H L [ X ] is finite, we have
H L [ X ] = 0 1 Φ 1 ( α ) ln α 1 α d α .
Subsequently, Yao et al. [27] proposed sine entropy.
Definition 7
([27]). Let X be an uncertain variable with uncertainty distribution Φ . The sine entropy of X is defined by
H S [ X ] = S Φ ( x ) d x .
Here, S ( y ) = sin ( π y ) for y [ 0 , 1 ] .
Yao et al. [27] provided an entropy formula and obtained the maximum entropy distribution of sine entropy.
Theorem 4
([27]). Suppose X is a regular uncertain variable with uncertainty distribution Φ . Then, we have
H S [ X ] = π 0 1 Φ 1 ( 1 α ) cos ( π α ) d α .
Theorem 5
([27]). Suppose X is an uncertain variable with uncertainty distribution Φ , expectation e, and variance σ 2 . Then, we have
H S [ X ] 2 2 π σ
and equality holds if and only if
Φ 1 ( α ) = e σ 2 cos ( π α ) , α ( 0 , 1 ) .
Tang and Gao [28] suggested triangular entropy and found an entropy formula.
Definition 8
([28]). Let X be an uncertain variable with uncertainty distribution Φ . The triangular entropy of X is defined by
H T [ X ] = T Φ ( x ) d x ,
where
T ( y ) = y , 0 y 0.5 , 1 y , 0.5 y 1 .
Theorem 6
([28]). Suppose X is a regular uncertain variable with uncertainty distribution Φ . Then, we have
H T [ X ] = 0 0.5 Φ 1 ( α ) d α + 0.5 1 Φ 1 ( α ) d α .
Dai [30] put forward quadratic entropy.
Definition 9
([30]). Let X be an uncertain variable with uncertainty distribution Φ . The quadratic entropy of X is defined by
H Q [ X ] = Q Φ ( x ) d x .
Here, Q ( y ) = y ( 1 y ) for y [ 0 , 1 ] .
Dai [30] also provided an entropy formula and got the maximum entropy distribution of quadratic entropy.
Theorem 7
([30]). Suppose X is a regular uncertain variable with uncertainty distribution Φ . If the entropy exists, then
H Q [ X ] = 0 1 Φ 1 ( α ) ( 2 α 1 ) d α .
Theorem 8
([30]). Suppose X is an uncertain variable with uncertainty distribution Φ , expectation e, and variance σ 2 . Then, we have
H Q [ X ] σ 3
and equality holds if and only if
Φ 1 ( α ) = e + σ 3 ( 2 α 1 ) , α ( 0 , 1 ) .
Ma [31] suggested unimodal entropy.
Definition 10
([31]). Let X be an uncertain variable with uncertainty distribution Φ . The unimodal entropy of X is defined by
H U [ X ] = U Φ ( x ) d x ,
where function U is increasing on ( 0 , 0.5 ] and decreasing on [ 0.5 , 1 ) , U ( 0 ) = U ( 1 ) = 0 .
Ma [31] also provided an entropy formula and got the maximum entropy distribution of the unimodal entropy for regular uncertain variables. The readers can refer to Liu [9], Yao et al. [27], Tang and Gao [28], Dai [30], Ma [31], and the references therein for details.
Here, we emphasize the entropy formulas of Dai and Chen [26], Yao et al. [27], Tang and Gao [28], Dai [30], and Ma [31] hold only for regular uncertain variables. We will develop an entropy formula that is true for any uncertain variable. To establish the entropy formulas, the method used by Dai and Chen [26], Yao et al. [27], Tang and Gao [28], and Dai [30] is Fubini’s theorem. We stress that Fubini’s theorem does not work in our case because the uncertainty distribution Φ may not be regular here. We need to develop new methods.
Finally, we point out, to obtain the maximum entropy distribution, the main method used by Chen and Dai [25], Yao et al. [27], and Dai [30] is the Lagrange multipliers method. Based on our entropy formula, we can obtain the maximum entropy distribution without using the Lagrange multipliers method. We also stress that the only tool we need is the Cauchy–Schwartz inequality. In the following section, we will exhibit our new entropy formula and provide the maximum entropy distribution of unimodel entropy.

3. Results

In this section, we will present our results and give some remarks. First, we point out that the variance formula of Yao [23] for regular uncertain variables is valid for any uncertain variable.
Theorem 9.
Let X be an uncertain variable with uncertainty distribution Φ , finite expectation e, and finite variance. Then
V [ X ] = 0 1 ( Φ ( α ) e ) 2 d α ,
where Φ is the generalized inverse of Φ .
We leave the verification of Theorem 9 to the next section. Based on our variance formula and some facts on the generalized inverse, we will develop an entropy formula for any uncertain variable.
Theorem 10.
Let X be an uncertain variable with finite expectation and finite variance. Assume further that unimodal function U is differentiable and satisfies
0 1 | U ( α ) | 2 d α < .
Then we have
H U [ X ] = 0 1 Φ ( α ) U ( α ) d α ,
where Φ is the generalized inverse of the uncertainty distribution Φ of X.
Remark 2.
1.
The entropy formulas in Dai and Chen [26], Yao et al. [27], Tang and Gao [28], Dai [30], and Ma [31] are valid only for regular uncertain variables. Our entropy formula holds for any uncertain variable.
2.
Fubini’s theorem, which is used by Dai and Chen [26], Yao et al. [27], Tang and Gao [28], and Dai [30] to establish the entropy formula there, does not work in our setting, because the uncertainty distribution is not necessarily regular here.
We specifically point out, under the assumption that the uncertainty distribution Φ is regular, let U ( α ) = α ln α ( 1 α ) ln ( 1 α ) , U ( α ) = sin ( π t ) , U ( α ) = α ( 1 α ) ,
U ( α ) = α , 0 α 0.5 , 1 α , 0.5 α 1
respectively. By Theorem 10, we can regain the entropy formulas of logarithm entropy in Theorem 3, sine entropy in Theorem 4, triangular entropy in Theorem 6, and quadratic entropy in Theorem 7. As the verifications are routine, we omit them.
Using our entropy formula in Theorem 10, we obtain the maximum entropy distribution of unimodal entropy.
Theorem 11.
Let X be an uncertain variable with uncertainty distribution Φ , finite expectation e, and finite variance σ 2 . Assume further that unimodal function U is differentiable and satisfies
0 1 | U ( α ) | 2 d α < .
Then, the unimodal entropy of X satisfies
H U [ X ] σ 0 1 | U ( α ) | 2 d α
and the equality holds if and only if
Φ 1 ( α ) = e c U ( α ) , α ( 0 , 1 )
where c is a real number such that
| c | 0 1 | U ( α ) | 2 d α = σ .
Remark 3.
We stress that the maximum entropy distribution of the unimodal entropy of Ma [31] is provided only for regular uncertain variables. Here, our maximum entropy distribution of unimodal entropy in Theorem 11 is obtained for any uncertain variable.
As applications of our maximum entropy distribution in Theorem 11, we can deduce the maximum entropy distribution of logarithm entropy in Theorem 2, sine entropy in Theorem 5, and quadratic entropy in Theorem 8 immediately.
Proposition 1
(Chen and Dai [25]). Let X be an uncertain variable with uncertainty distribution Φ , finite expectation e, and variance σ 2 . Then, the logarithmic entropy of X satisfies
H L [ X ] π σ 3
and the equality holds if and only if
Φ 1 ( α ) = e + σ 3 π ln α 1 α , α ( 0 , 1 ) ,
i.e., X is a normal uncertain variable N ( e , σ ) .
Proposition 2
(Yao et al. [27]). Let X be an uncertain variable with uncertainty distribution Φ , finite expectation e, and variance σ 2 . Then, its sine entropy satisfies
H S [ X ] π σ 2
and the equality holds if and only if
Φ 1 ( α ) = e σ 2 cos ( π α ) , α ( 0 , 1 ) .
Proposition 3
(Dai [30]). Let X be an uncertain variable with uncertainty distribution Φ , finite expectation e, and variance σ 2 . Then, its quadratic entropy satisfies
H Q [ X ] σ 3
and the equality holds if and only if
Φ 1 ( α ) = e + σ 3 ( 2 α 1 ) , α ( 0 , 1 ) .
We leave the proofs of Corollary 1, Corollary 2, and Corollary 3 to Section 6. In the following section, we will prove Theorem 9, Theorem 10, and Theorem 11.

4. Proof of Theorem 9, Theorem 10, and Theorem 11

To prove Theorems 9 and 10, we need a result of the substitution rule on Lebesgue–Stieltjes integrals, which is a conclusion of Proposition 2 of Falkner and Teschl [37].
Lemma 2
([37]). Assume function M is increasing over [ a , b ] and function N is left-continuous increasing over [ M ( a ) , M ( b ) ] . Then, for each bounded Borel function f over [ a , b ] , one has
a b f ( x ) d N ( M ( x ) ) = M ( a ) M ( b ) f M 1 ( y ) d N ( y ) ,
where M 1 is the generalized inverse of M defined by (4).
Here, we point out that if N ( y ) = y , this result goes back to the classical case of Lebesgue [38]. Before proving the main results, we also need a simple fact.
Lemma 3.
Suppose function g is increasing on ( 0 , 1 ) with
lim x 0 + g ( x ) = a n d lim x 1 g ( x ) = + ,
and let L be a function increasing on ( 0 , a ] and decreasing on [ a , 1 ) , respectively, for some a in ( 0 , 1 ) . Then
(a)
| 0 a g ( x ) d L ( x ) | < + means
lim A 0 + g ( A ) ( L ( A ) L ( 0 ) ) = 0 ;
(b)
| a 1 g ( x ) d L ( x ) | < + means
lim B 1 g ( B ) ( L ( 1 ) L ( B ) ) = 0 .
Because the verification of Lemma 3 is a routine computation, we omit it.
We also point out that, notwithstanding that uncertainty distribution Φ may not be right-continuous, since Φ is increasing, the assumption that Φ is right-continuous does not change the variance and the entropy of an uncertain variable. Thus, we assume uncertainty distribution Φ is right-continuous in the rest. Now, we will prove Theorem 9.
Proof of Theorem 9.
Let Y = X e and Ψ be the uncertainty distribution of Y. Then we have
E [ Y ] = 0 , V [ X ] = V [ Y ] , and Ψ = Φ e .
We only need to prove
V [ Y ] = 0 1 ( Ψ ( α ) ) 2 d α .
Recall (see equality (2))
V [ Y ] = 0 ( 1 Ψ ( x ) ) d x + 0 Ψ ( x ) d x .
We only verify
0 ( 1 Ψ ( x ) ) d x = Ψ ( 0 ) 1 ( Ψ ( α ) ) 2 d α .
The verification of
0 Ψ ( x ) d x = 0 Ψ ( 0 ) ( Ψ ( α ) ) 2 d α
is similar. First, note that
0 ( 1 Ψ ( x ) ) d x = lim B 1 0 Ψ ( B ) ( 1 Ψ ( x ) ) d x .
Letting y = x , we have
0 Ψ ( B ) ( 1 Ψ ( x ) ) d x = 0 Ψ ( B ) ( 1 Ψ ( y ) ) d y 2 = 0 Ψ ( Ψ ( 0 ) ) ( 1 Ψ ( y ) ) d y 2 + Ψ ( Ψ ( 0 ) ) Ψ ( B ) ( 1 Ψ ( y ) ) d y 2 = ( 1 Ψ ( 0 ) ) ( Ψ ( Ψ ( 0 ) ) ) 2 + Ψ ( Ψ ( 0 ) ) Ψ ( B ) ( 1 Ψ ( y ) ) d y 2 ,
where the last equality is because Ψ ( y ) Ψ ( 0 ) for y Ψ ( Ψ ( 0 ) ) , 0 .
By Lemma 2, let M = Ψ , f ( x ) = 1 x , N ( y ) = y 2 , noting that M = Ψ according to Lemma 1, we have
Ψ ( Ψ ( 0 ) ) Ψ ( B ) ( 1 Ψ ( y ) ) d y 2 = Ψ ( 0 ) B ( 1 x ) d ( Ψ ( x ) ) 2 .
By integration by parts, we have
Ψ ( 0 ) B ( 1 x ) d ( Ψ ( x ) ) 2 = ( 1 B ) ( Ψ ( B ) ) 2 ( 1 Ψ ( 0 ) ) ( Ψ ( Ψ ( 0 ) ) ) 2 + Ψ ( 0 ) B ( Ψ ( x ) ) 2 d x .
Then, we have
0 Ψ ( B ) ( 1 Ψ ( x ) ) d x = ( 1 B ) ( Ψ ( B ) ) 2 + Ψ ( 0 ) B ( Ψ ( x ) ) 2 d x .
Let
G ( B ) = Ψ ( 0 ) B ( Ψ ( x ) ) 2 d x .
Note that function G is increasing over [ Ψ ( 0 ) , 1 ] . It follows that
lim B 1 G ( B ) = lim n + G ( 1 1 n ) .
By monotone convergence theorem, we have
Ψ ( 0 ) 1 ( Ψ ( x ) ) 2 d x = lim n + Ψ ( 0 ) 1 ( Ψ ( x ) ) 2 1 [ Ψ ( 0 ) , 1 1 n ] ( x ) d x = lim n + G ( 1 1 n ) ,
where 1 [ Ψ ( 0 ) , 1 1 n ] is indicator function of the interval [ Ψ ( 0 ) , 1 1 n ] .
Thus, we have
lim B 1 Ψ ( 0 ) B ( Ψ ( x ) ) 2 d x = Ψ ( 0 ) 1 ( Ψ ( x ) ) 2 d x .
We need to prove
lim B 1 ( 1 B ) ( Ψ ( B ) ) 2 = 0 .
If
lim B 1 Ψ ( B ) < + ,
the result is obvious. If
lim B 1 Ψ ( B ) = + ,
by Lemma 3, let g ( x ) = ( Ψ ( x ) ) 2 and L ( x ) = x , we have
lim B 1 ( 1 B ) ( Ψ ( B ) ) 2 = 0 .
Thus, we finish the proof of Theorem 9.  □
Now, using Theorem 9, Lemma 2, and Lemma 3, we will prove Theorem 10, i.e., the entropy formula.
Proof of Theorem 10.
Recall
H U [ X ] = U Φ ( x ) d x .
Note that
Φ ( x ) = 0 , x Φ ( 0 ) , 1 , x Φ ( 1 )
and U ( 0 ) = U ( 1 ) = 0 . We have
U Φ ( x ) d x = Φ ( 0 ) Φ ( 1 ) U Φ ( x ) d x .
By the monotone convergence theorem, we have
Φ ( 0 ) Φ ( 1 ) U Φ ( x ) d x = lim n + Φ ( 0 ) Φ ( 1 ) U Φ ( x ) 1 [ Φ ( 1 n ) , Φ ( 1 1 n ) ] ( x ) d x ,
where
1 [ Φ ( 1 n ) , Φ ( 1 1 n ) ] ( x )
is indicator function of the interval [ Φ ( 1 n ) , Φ ( 1 1 n ) ] .
Let
G ( ϵ ) = lim ϵ 0 + Φ ( 0 ) Φ ( 1 ) U Φ ( x ) 1 [ Φ ( ϵ ) , Φ ( 1 ϵ ) ] ( x ) d x .
Note that G ( ϵ ) is decreasing over [ 0 , 1 ] . We have
lim n + G ( 1 n ) = lim ϵ 0 + G ( ϵ ) .
Thus
Φ ( 0 ) Φ ( 1 ) U Φ ( x ) d x = lim n + G ( 1 n ) = lim ϵ 0 + Φ ( ϵ ) Φ ( 1 ϵ ) U ( Φ ( x ) ) d x .
Let f = U , M = Φ , N ( x ) = x ; by Lemma 2 and integration by parts, we have
Φ ( ϵ ) Φ ( 1 ϵ ) U ( Φ ( x ) ) d x = ϵ 1 ϵ U ( α ) d Φ ( α ) = U ( 1 ϵ ) Φ ( 1 ϵ ) U ( ϵ ) Φ ( ϵ ) ϵ 1 ϵ Φ ( α ) ) d U ( α ) .
Then, we have
H U ( X ) = lim ϵ 0 + U ( 1 ϵ ) Φ ( 1 ϵ ) lim ϵ 0 + U ( ϵ ) Φ ( ϵ ) lim ϵ 0 + ϵ 1 ϵ Φ ( α ) d U ( α ) .
We only need to show
lim ϵ 0 + U ( 1 ϵ ) Φ ( 1 ϵ ) = 0 , lim ϵ 0 + U ( ϵ ) Φ ( ϵ ) = 0
and
lim ϵ 0 + ϵ 1 ϵ Φ ( α ) d U ( α ) = 0 1 Φ ( α ) U ( α ) d α .
Noting that 0 1 U ( α ) d α = 0 , by the Cauchy–Schwartz inequality and Theorem 9, we have
0 1 Φ ( α ) d U ( α ) = 0 1 ( Φ ( α ) e ) U ( α ) d α 0 1 ( Φ ( α ) e ) 2 d α 0 1 ( U ( α ) ) 2 d α = σ 0 1 ( U ( α ) ) 2 d α < + .
By monotone convergence theorem, again we have
lim ϵ 0 + ϵ 1 ϵ Φ ( α ) d U ( α ) = 0 1 Φ ( α ) U ( α ) d α .
Let L = U and g = Φ ; by Lemma 3, we have
lim ϵ 0 + U ( 1 ϵ ) Φ ( 1 ϵ ) = 0 and lim ϵ 0 + U ( ϵ ) Φ ( ϵ ) = 0 .
Then, we have
H U [ X ] = 0 1 Φ ( α ) U ( α ) d α .
Thus, we finish the proof of Theorem 10.  □
Now, based on our entropy formula, we can derive the maximum entropy distribution of unimodal entropy immediately.
Proof of Theorem 11.
Noting that 0 1 U ( α ) d α = 0 , by Theorem 10, we have
H U [ X ] = 0 1 ( Φ ( α ) e ) U ( α ) d α .
By the Cauchy–Schwartz inequality and Theorem 9, we have
H U [ X ] 0 1 ( Φ ( α ) e ) 2 d α 0 1 ( U ( α ) ) 2 d α = σ 0 1 ( U ( α ) ) 2 d α ,
and “=” holds if and only if there exists some real number c such that
Φ ( α ) e = c U ( α ) .
Then, we have
0 1 ( Φ ( α ) e ) 2 d α = c 2 0 1 ( U ( α ) ) 2 d α ,
i.e.,
| c | 0 1 | U ( α ) | 2 d α = σ .
Thus, Theorem 11 is proved.  □

5. Proof of Proposition 1, Proposition 2, and Proposition 3

In this section, we use our maximum entropy distribution to deduce Proposition 1, Proposition 2, and Proposition 3, i.e., the maximum entropy distributions of logarithm entropy, sine entropy, and quadratic entropy. As the proofs of Proposition 2 and Proposition 3 are similar to that of Proposition 1, we only prove Proposition 1.
Proof of Proposition 1.
Let
U ( α ) = α ln α ( 1 α ) ln ( 1 α ) , α ( 0 , 1 ) .
Then,
U ( α ) = ln 1 α α .
By Theorem 2 of Ma et al. [39], we have
0 1 ( U ( α ) ) 2 d α = π 2 3 .
By Theorem 11, we have
H L [ X ] π σ 3
and the equality holds if and only if
Φ 1 ( α ) e = c ln α 1 α , α ( 0 , 1 ) .
Since functions
Φ 1 ( α ) and ln α 1 α
are increasing for α ( 0 , 1 ) , the real number c is positive. By equality (23), we have
0 1 ( Φ 1 ( α ) e ) 2 d α = c 2 0 1 ( ln α 1 α ) 2 d α .
Thus
c = σ 3 π .
Then
Φ 1 ( α ) = e + σ 3 π ln α 1 α , α ( 0 , 1 ) ,
i.e., Φ is a normal uncertainty distribution with expectation e and variance σ 2 .  □

6. Discussion

In this paper, by using the generalized inverse of an increasing function, we developed a useful entropy formula that is valid for any uncertain variable and encompasses all the previous cases in the literature. In addition, based on our entropy formula, we obtained the maximum entropy distribution for unimodel entropy. We also point out that most of the entropy formulas and maximum entropy distributions of the uncertain variables in the literature are special cases of ours.
Further research can consider the following two directions:
(1)
Develop the entropy formula and the maximum entropy principle of an uncertain set;
(2)
Establish the entropy formula and research the maximum entropy distributions of uncertain random variables.

Author Contributions

Conceptualization, Y.L. and G.M.; methodology, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, G.M.; supervision, G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Project of Natural Science Foundation of Educational Committee of Henan Province (Grant No. 23A110014).

Institutional Review Board Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Shannon, C. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  2. Gray, R.M. Entropy and Information Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  3. Kolmogorov, A. New metric invariant of transitive dynamical systems and endomorphisms of lebesgue spaces. Dokl. Russ. Acad. Sci. 1958, 119, 861–864. [Google Scholar]
  4. Liu, B. Some research problems in uncertainty theory. J. Uncertain Syst. 2009, 3, 3–10. [Google Scholar]
  5. Malakar, N.K.; Knuth, K.H.; Lary, D.J. Maximum joint entropy and information-based collaboration of automated learning machines. In AIP Conference Proceedings 31st; American Institute of Physics: College Park, MD, USA, 2012; Volume 1443, pp. 230–237. [Google Scholar] [CrossRef] [Green Version]
  6. Martin, N.F.; Engl, J.W. Mathematical Theory of Entropy (No.12); Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  7. Walters, P. An Introduction to Ergodic Theory; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  8. Einsiedler, M.; Thomas, W. Ergodic Theory; Springer: London, UK, 2013. [Google Scholar]
  9. Liu, B. Uncertainty Theory, 4th ed.; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  10. Jaynes, E. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  11. Jaynes, E. Information theory and statistical mechanics. ii. Phys. Rev. 1957, 108, 171–190. [Google Scholar] [CrossRef]
  12. Polpo, A.; Stern, J.; Louzada, F.; Izbicki, R.; Takada, H. Bayesian Inference and Maximum Entropy Methods in Science and Engineering; Springer Proceedings in Mathematics & Statistics: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  13. Buckley, J.J. Entropy principles in decision making under risk. Risk Anal. 1985, 5, 303–313. [Google Scholar] [CrossRef]
  14. Hoefer, M.P.; Ahmed, S.B. The Maximum Entropy Principle in Decision Making Under Uncertainty: Special Cases Applicable to Developing Technologies. Am. J. Math. Manag. Sci. 1990, 10, 261–273. [Google Scholar] [CrossRef]
  15. Clarke, B. Information optimality and Bayesian modelling. J. Econom. 2007, 138, 405–429. [Google Scholar] [CrossRef]
  16. Naidu, P.S. Modern Spectrum Analysis of Time Series: Fast Algorithms and Error Control Techniques; CRC Press: Boca Raton, FL, USA, 1995. [Google Scholar]
  17. Zhang, Y. Principle of maximum entropy for reliability analysis in the design of machine components. Front. Mech. Eng. 2019, 14, 21–32. [Google Scholar] [CrossRef]
  18. Zu, T.; Kang, R.; Wen, M.; Zhang, Q. Belief reliability distribution based on maximum entropy principle. IEEE Access 2017, 6, 1577–1582. [Google Scholar] [CrossRef]
  19. Abbas, A.E.H.; Cadenbach, A.; Salimi, E. A Kullback–Leibler view of maximum entropy and maximum log-probability methods. Entropy 2017, 19, 232. [Google Scholar] [CrossRef] [Green Version]
  20. Liu, B. Uncertainty Theory, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar] [CrossRef]
  21. Peng, Z.; Iwamura, K. A sufficient and necessary condition of uncertainty distribution. J. Interdiscip. Math. 2010, 13, 277–285. [Google Scholar] [CrossRef]
  22. Liu, Y.; Lio, W. A revision of sufficient and necessary condition of uncertainty distribution. J. Intell. Fuzzy Syst. 2020, 38, 4845–4854. [Google Scholar] [CrossRef]
  23. Yao, K. A formula to calculate the variance of uncertain variable. Soft Comput. 2015, 19, 2947–2953. [Google Scholar] [CrossRef]
  24. Sheng, Y.; Kar, S. Some results of moments of uncertain variable through inverse uncertainty distribution. Fuzzy Optim. Decis. Mak. 2015, 14, 57–76. [Google Scholar] [CrossRef]
  25. Chen, X.; Dai, W. Maximum entropy principle for uncertain variables. Int. J. Fuzzy Syst. 2011, 13, 232–236. [Google Scholar]
  26. Dai, W.; Chen, X. Entropy of function of uncertain variables. Math. Comput. Model. 2012, 5, 54–760. [Google Scholar] [CrossRef]
  27. Yao, K.; Gao, J.; Dai, W. Sine entropy for uncertain variable. Int. J. Uncertain. Fuzziness Knowl. Based Syst. 2013, 21, 743–753. [Google Scholar] [CrossRef]
  28. Tang, W.; Gao, W. Triangular entropy of uncertain variables. International Information Institute (Tokyo). Information 2013, 16, 1279. [Google Scholar]
  29. Ning, Y.; Ke, H.; Fu, Z. Triangular entropy of uncertain variables with application to portfolio selection. Soft Comput. 2015, 19, 2203–2209. [Google Scholar] [CrossRef]
  30. Dai, W. Quadratic entropy of uncertain variables. Soft Comput. 2018, 22, 5699–5706. [Google Scholar] [CrossRef]
  31. Ma, G. A remark on the maximum entropy principle in uncertainty theory. Soft Comput. 2021, 25, 13911–13920. [Google Scholar] [CrossRef]
  32. Liu, B. Uncertainty Theory: A Branch of Mathematics for Modeling Human Uncertainty; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  33. Embrechts, P.; Hofert, M. A note on generalized inverses. Math. Methods Oper. Res. 2013, 77, 423–432. [Google Scholar] [CrossRef] [Green Version]
  34. Resnick, S.I. Extreme Values, Regular Variation, and Point Processes (Vol. 4); Springer Science Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  35. Kämpke, T.; Radermacher, F.J. Income modeling and balancing. In Lecture Notes in Economics and Mathematical Systems; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  36. Iritani, J.; Kuga, K. Duality between the Lorenz curves and the income distribution functions. Econ. Stud. Q. (Tokyo 1950) 1983, 34, 9–21. [Google Scholar]
  37. Falkner, N.; Teschl, G. On the substitution rule for Lebesgue–Stieltjes integrals. Expo. Math. 2012, 30, 412–418. [Google Scholar] [CrossRef] [Green Version]
  38. Lebesgue, H. Sur l’integrale de stieltjes et sur les operérations fonctionelles linéaires. CR Acad. Sci. Paris 1910, 150, 86–88. [Google Scholar]
  39. Ma, G.; Yang, X.; Yao, X. A relation between moments of Liu process and Bernoulli numbers. Fuzzy Optim. Decis. Mak. 2021, 20, 261–272. [Google Scholar] [CrossRef]
Figure 1. Normal uncertainty distribution N ( e , σ ) .
Figure 1. Normal uncertainty distribution N ( e , σ ) .
Entropy 25 01195 g001
Figure 2. Linear uncertainty distribution L ( a , b ) .
Figure 2. Linear uncertainty distribution L ( a , b ) .
Entropy 25 01195 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Ma, G. On the Entropy and the Maximum Entropy Principle of Uncertain Variables. Entropy 2023, 25, 1195. https://doi.org/10.3390/e25081195

AMA Style

Liu Y, Ma G. On the Entropy and the Maximum Entropy Principle of Uncertain Variables. Entropy. 2023; 25(8):1195. https://doi.org/10.3390/e25081195

Chicago/Turabian Style

Liu, Yujun, and Guanzhong Ma. 2023. "On the Entropy and the Maximum Entropy Principle of Uncertain Variables" Entropy 25, no. 8: 1195. https://doi.org/10.3390/e25081195

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop