1. Introduction
How can some classic inequalities and results can be extended, refined or estimated, if we extend more intrinsic theorems under the perspective of more general functions? This paper shows an approach. Recall some classic inequalities and existing theorems first.
We denote by
and
the unweighted arithmetic, geometric, and harmonic means of positive real numbers
with
. Arithmetic mean is the most commonly used average mean in daily life. Geometric mean is applied to data with exponential growth property, like investment growth [
1], production growth, and population growth. The compound annual growth rate (CAGR)
in
n years is calculated by the following:
regarding the growth rate
in each
i-th year.
The equivalent resistance value of two parallel resistance, is two times the harmonic means of a resitance value of two. The overall velocity
v for a distance
, is the harmonic mean of velocity
in the first half distance
l and velocity
in the second half distance
l, calculated as follows:
The fundamental inequalities are written as follows:
These formulas have Ky Fan-type refinements. The second inequality on the right side of (
1) is due to Ky Fan in [
2] (p. 5), which is known as Ky Fan inequality in the literature, and the first inequality on the left side, known as Wang–Wang inequality, is proposed in [
3]. Inequality is written as follows:
holds for
.
Part of (
2) is proved in [
4], and the whole chain (
2) is proved in [
5]. Inequality is written as follows:
holds for
.
They are followed by [
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15].
In [
16], the following Conclusion 1 is established for convex functions on three intervals, to generalize Levinson inequality and unify the right side of Ky Fan-type inequalities (
1) and (
2). As pointed out in the same paper, it also has some connection with the Lah–Ribarič inequality [
17], and its extension can be used to prove the Hermite–Hadamard inequality [
18].
Conclusion 1. Let and be a given function. Let and be two nonnegative functions. If is increasing in andthen the following:holds for every ϕ that is convex such that the integrals exist. Then in [
19], this Conclusion 1 is further extended and is also used to establish inequality for Csiszár
-divergence. Note that this divergence can quantify the difference between two probability distributions, making it a useful tool for hypothesis testing and statistical inference; quantify the difference between the true distribution and the estimated distribution, making it a useful tool for data compression; and quantify the difference between the true distribution and the predicted distribution, making it a useful tool for classification problems and machine learning.
Recall the notion of Csiszár
-divergence [
20,
21,
22]. Given a convex function
, the
-divergence functional, written as follows:
is a generalized measure of information, a “distance function” on the set of probability distributions
. By appropriately defining this convex function
, various divergences are derived; see Chapter 1 in [
22] and Chapter 9.2 in [
20] as well as related references. In [
19], the following comparison between two different
and
is established.
Theorem 1. Let be a convex function. If and the following:for some , then we have the following: This theorem also holds for the probability distribution function (see [
19]). This inequality looks similar to data processing inequality in a discrete case, but we do not have an assumption for some
, further supposing the ratio of
.
In this paper, we will give identity expressions for these two inequalities (
5) and (
6), and by letting the function be convex, we obtain inequalities in previous papers as special cases. The identity can also establish the known trapezoid-type inequality (
7) (which can be seen as an estimation for the Hermite–Hadamard-type inequality). As the first application, we use these identities to extend Levinson inequality and give new refinements and reverses of Ky Fan-type inequalities (
1) and (
2). As the second application, we use the identity to obtain new inequalities for special examples of distance functions. Throughout this paper, we use the following notation:
2. Identity Extension
In this section, we will give identity expressions for Conclusion 1 and Theorem 1. Some special cases and corollary are also mentioned.
Theorem 2. Let and be a given function. Let and be two nonnegative functions. If is increasing in and (
3)
, (
4)
are satisfied, then for such that is absolutely continuous, we have the following:where the following is obtained:Furthermore, we have . Proof. In (
14) of Theorem 14, set
and the following:
in Theorem 2, taking (
3) and (
4) into consideration, we obtain the identity.
As for the nonnegativity of , this can be divided into three situations.
1. If
, then we obtain the following:
2. If
, then we obtain the following:
for a certain
.
2.1. If
, then we obtain the following:
2.2. If
, then according to (
3) and (
4) we have the following:
3. If
, then we obtain the following:
for a certain
, according to (
3) and (
4) we have the following:
Hence, in all situations we affirm
. □
From Theorem 2 we can prove the following theorem, which gives the upper and lower bounds.
Theorem 3. Let and be a given function. Let and be two nonnegative functions. If is increasing in and (
3)
, (
4)
are satisfied, then for such that , we have the following: Proof. Utilize Theorem 2 to obtain the following expression:
then from
in Theorem 2 and
we have
Use Fubini’s theorem to obtain the following:
the last step is due to (
3) and (
4). □
Remark 1. It is easy to observe the following:as it transforms from . By putting in Theorem 3, we obtain Conclusion 1 for twice differentiable convex functions ϕ.
From Theorem 2 another upper bound can also be given as the following.
Theorem 4. Let and be a given function. Let and be two nonnegative functions. If is increasing in and the following:then for such that , we have the following:for , where we obtain the following: Proof. Take absolute value and use Hölder inequality for Theorem 2. □
From Theorem 4 we can deduce the following extension for trapezoid inequality, see chapter 2 in [
18].
Corollary 1. Let twice differentiable such that , we have the following:for , where B is Euler’s Beta-function. Proof. In Theorem 4, set , , it is clear all the conditions are satisfied.
Then, taking
, we have the following:
With some calculation, we obtain the following:
from which we obtain the desired result. □
Taking
as the original trapezoid inequality, written as follows:
It is very useful in numerical integration, as we cut the interval into more smaller intervals (suppose evenly
n intervals), the remainder terms (error) will be much smaller, because of the following:
Remark 2. Another corollary is , , and take , we can obtain estimation forwhich is more difficult to calculate. Then we give the identity expression for Theorem 1, without assuming to be convex first.
Theorem 5. If , then for such that is absolutely continuous, we have the following: Proof. In (
13) of Theorem 14, set
in Theorem 5, we obtain the expression for
.
Similarly, in (
13) of Theorem 14, set
in Theorem 5, we obtain the expression for
.
Take into consideration , we complete the proof. □
Then we state two lemmas below, in which, if
satisfy some further conditions, we have the following:
The proofs are in the Some Lemmas section.
Lemma 1. In Theorem 5, if we further suppose the following:for some , then we obtain the following:for all . Lemma 2. In Theorem 5, if we further suppose , which means , and (here we use q to refer either or ), then we obtain the following:(the sequence majorizes the sequence with the weight q), then we obtain the following:for all . Combine Theorem 5, Lemma 1 and Lemma 2, we have the following estimation about .
Theorem 6. Under the assumptions of Theorem 5, if condition in Lemma 1 or Lemma 2 is further satisfied, then we obtain the following:for such that . Proof. Under the assumptions of Lemma 1 or Lemma 2, we have the following:
from Theorem 5 we prove the conclusion. □
Remark 3. Set , we obtain Theorem 1 for twice differentiable convex functions ϕ.
Remark 4. If , then we have a more direct expression, written as follows:for such that . 3. Application
In this section, we first use Theorem 3 to establish Levinson and Ky Fan-type inequalities, and then use Theorem 6 and Remark 4 to give several special examples of distance functions in mathematical statistics and information theory.
In [
16], the following general inequality for 3-convex function is established, which can be used to prove Jensen inequality, Levinson inequality, refine power means inequality and unify two Ky Fan-type inequalities.
Conclusion 2. Suppose that is a continuous 3-convex function, for , the inequality is written as follows:hold. In this section, based on Theorem 3, we will further extend this conclusion and use the extension to refine Ky Fan-type inequalities. These results are generalizations of Levinson, Jensen, and Ky Fan inequalities. The condition is more general than as may be .
Theorem 7. Let be a continuous function such that . For , we have the following:And we also have the follwoing: Proof. The proof is divided into two parts.
1. Term (9) satisfies all the corresponding condition in Theorem 3. It is easy to observe the following:
where
and
, if
t belongs to and only belongs to
k interval(s) of all
, written as follows:
Then, in Theorem 3, we can set
in Theorem 7, this is feasible, as we only need to check condition (
3) and (
4) naturally hold.
First, we obtain the following:
then the following:
which satisfies condition (
4).
Second, we obtain the following:
then the following:
which satisfies condition (
3).
Thus, we can use Theorem 3 to obtain the following:
and
2. Calculate terms (8) and (10), and nonnegativity.
which is equivalent to the desired conclusion. □
Theorem 8 is powerful enough to deduce several old and new inequalities below.
Remark 5. Take in Theorem 7, we obtain Conclusion 2, if we further set we obtain the Levinson inequality.
We can also deduce a known estimation for Jensen inequality below.
Corollary 2. Let be a continuous function such that . For , we have the following: Proof. In Theorem 7, set
, and multiple the inequality with
. Let
, we obtain the following:
let
we obtain the corollary. □
Then we use Theorem 7 to establish
refinements and
reverses for Ky Fan-type inequalities (
1), (
2) mentioned in the introduction section, these special cases are new.
Theorem 8. Denoting by the arithmetic and geometric means of positive real numbers with , we have the following:for , where , and the following: Proof. Set
defined on
in Theorem 7, for
and
. As
we know the following:
Thus we can use Theorem 7. With some calculation and simplification, then taking exp we obtain the desired results. □
Remark 6. The Ky Fan inequality is equivalent to the following:for , the readers may also establish similar refinement and reverse like the proof above, for the following discussion. Consider two investment products that have inverse yield in each period i, for (e. g. ETF and its inverse ETF).
If is nonnegative in each period of time, it is clear that the overall investment yield of X is higher than Y after n periods, which is written as follows:With (
11)
we obtain a more accurate estimation, written as follows:where is the arithmetic average of X’s yields in n periods. Apply the refinement and reverse of Ky Fan inequality, we can obtain even more detailed estimations, left to readers.
Theorem 9. Denote by the arithmetic and geometric means of positive real numbers with , we have the following:for , where the following is formulated: Proof. Set
defined on
in Theorem 7, for
and
. As
we know the following:
Thus we can use Theorem 7. With some calculation and simplification, taking exp we obtain the desired results. □
Remark 7. If , then we obtain the following:for . Readers may also establish similar refinement and reverse like the proof above, replacing 1 with λ, for the following discussion. Funds with better annual yield performance often have higher management fee rates, so it is important to consider the actual investment yields after paying fees. Here we just consider a simplified model.
Suppose the “better” fund X has the annual yield , for each year i, with a fixed annual management fee rate , and the “worse” fund Y has the annual yield for each year i, without management fee. Define .
From (
12)
we have the following:thus, we obtain the following:It means, if the annual fee rate , then the actual yield of the fund X must be higher than Y after n years. Sometimes we even don’t need to know the actual value of each . For example, if we know the worse Y has a negative yield each year, then we affirm ; if we further know the annual fee rate of X satisfy , then we predict the actual yield of the fund X must be higher than Y.
Apply the refinement and reverse of Ky Fan-type inequality, we have more detailed prediction about, when the X has actual higher or lower yield than Y, concerning the annual fee rate θ. Left to readers.
For similar converses of Ky Fan inequality with the following form:
for some
C concerning
, see [
9,
23].
Then we list some examples of distance functions, and use Theorem 6, Remark 4 to give inequalities for these distance functions.
Definition 1. For the follwoing:the ϕ-divergence is calculated as follows:the Kullback–Leibler distance.
We have the following estimation for two different pairs of Kullback–Leibler distances.
Theorem 10. For with , if condition in Lemma 1 or Lemma 2 is further satisfied, then the following is calculated: Proof. As
, we have the following:
And under the assumption of Lemma 1 or Lemma 2, we have the following:
Then we can use Theorem 6 and Remark 4 to obtain the result. □
Definition 2. Let the following be true:the ϕ-divergence is as follows:the Hellinger distance.
We have the following estimation for two different pairs of Hellinger distance.
Theorem 11. For with , if condition in Lemma 1 or Lemma 2 is further satisfied, then the following is calculated: Proof. As , with similar discussion in Theorem above we obtain the result. □
Definition 3. For , let the following be true:the ϕ-divergence is written as follows:the
-
order entropy. And Rényi divergence of order α is defined by the following: We have the following estimation for two different pairs of -order entropy.
Theorem 12. For with , if condition in Lemma 1 or Lemma 2 is further satisfied, then the following is calculated: Proof. As , with similar discussion in Theorem above we obtain the result. □
Definition 4. Let the following be true:the ϕ-divergence is calculated as follows:the -
distance.
We have the following estimation for two different pairs of -distance. This is a special case of identity.
Theorem 13. For with , if condition in Lemma 1 or Lemma 2 is further satisfied, then the following is calculated: Proof. As , the inequality in Theorem 6 becomes identity. □
Remark 8. If we let in the four theorems above, we obtain estimation for these four types of divergence .