Abstract
By utilizing the properties of positive definite matrices, mathematical expectations, and positive linear functionals in matrix space, the Kantorovich inequality and Wielandt inequality for positive definite matrices and random variables are obtained. Some novel Kantorovich type inequalities pertaining to matrix ordinary products, Hadamard products, and mathematical expectations of random variables are provided. Furthermore, several interesting unified and generalized forms of the Wielandt inequality for positive definite matrices are also studied. These derived inequalities are then exploited to establish an inequality regarding various correlation coefficients and study some applications in the relative efficiency of parameter estimation of linear statistical models.
Keywords:
positive-definite matrix; correlation coefficient; kantorovich inequality; covariance matrix; mathematical expectation MSC:
15B48; 62H20
1. Introduction
The Kantorovich inequality and Wielandt inequality have long been recognized as fundamental mathematical tools with profound implications in various fields. Their influence and application have been extensively documented in the literature, making them keystones in mathematical analysis and statistical modeling.
Ref. [1] provided an early and comprehensive analysis of the Kantorovich inequality, exploring its theoretical underpinnings and potential applications. Ref. [2] further extended this analysis, focusing on the inequality’s use in optimization problems and its connection to other mathematical principles. Ref. [3] then bridged the theoretical gap between these inequalities and their practical applications, particularly in linear statistical models. Their work highlighted how these inequalities could be used to enhance statistical analysis and modeling.
Ref. [4] further explored the applications of these inequalities in statistical models, demonstrating their versatility and importance in complex statistical analysis. Ref. [5] provided an in-depth examination of the Wielandt inequality, emphasizing its role in matrix analysis and its connection to other fundamental matrix inequalities. Ref. [6] revisited the Kantorovich inequality, offering a fresh perspective on its use in modern mathematical problems, especially those related to compressed sensing and signal processing. Ref. [7] brought a contemporary perspective, discussing how these inequalities remain relevant in modern statistical and mathematical challenges.
Despite the rich literature surrounding the Kantorovich and Wielandt inequalities, there is still ample room for exploration and generalization. Many existing studies focus on specific applications or theoretical aspects, leaving a gap for a more unified and generalized approach. This article aims to fill that gap, building upon the foundation laid by previous scholars and pushing the boundaries of these inequalities’ applications in probability and statistics.
In conclusion, the Kantorovich and Wielandt inequalities have been extensively studied and applied in various fields. However, there is still a need for a more comprehensive and generalized understanding of these inequalities, which this article aims to provide. By leveraging past research and adopting a novel approach, we hope to offer new insights and applications for these fundamental mathematical tools.
2. Notation and Definition
Notation: We first give some notations [8,9] used in this article. Let R and C be the sets of real and complex numbers, respectively. The symbols represent complex linear spaces formed by and n order matrix. represents the set of Hermite matrices of order n. Furthermore, , designate convex cones formed by positive semidefinite matrices and positive definite matrices of order n respectively. For a matrix , the notation represent the transpose and the conjugate transpose of A. Additionally, for a matrix , let , , be the maximum, minimum eigenvalue and condition number of A. For indicate that A is a positive-definite and semipositive-definite matrix, respectively. The inequality implies that is a semipositive-definite matrix. The symbol represents the identity matrix of order n, which can also be noted simply as I without loss of generality. For a vector , the 2-norm of x is defined as . For a random variable with a finite first-order moment, its mathematical expectation is denoted by . Furthermore, assuming are p-dimensional and q-dimensional random vectors respectively, represents the covariance matrix of and . Especially, is defined as the variance of . Note that all the statistical applications in this paper are referred to the real-valued random variables.
Definition 1.
If a linear functional Φ of satisfies for any , then Φ is called a positive linear functional on . Furthermore, Φ is called a strictly positive linear functional if for any .
Lemma 1.
Let be random variables with finite first-order moment on probability space and matrix . It follows that the expectation matrix . Note that represents a matrix composed by elements
Proof.
For any , since , then . It follows that . □
Lemma 2.
Let and , then
Proof.
Lemma 2 can be obtained via . Since
Hence, , and , we have
which completes the proof. □
Lemma 3
([3]). Assume , and then
Lemma 4.
Suppose , then
Proof.
Just need to prove , and . By Rayleigh-Ritz Theorem,
where , and . Since , then , it follows that
Combined with the Cauchy-Schwarz inequality, the Arithmetic-Geometric mean inequality and , we have
This completes the proof of Lemma 4. □
Lemma 5
([2,8]). If , and , then
Lemma 6.
If Φ is a strictly positive linear functional from , and , then
Proof.
For any parameter , since , then , it follows that
This completes the proof. □
Lemma 7.
Suppose are invertible matrix, , and , then
Proof.
Since
and Lemma 4, the results can be obtained. □
3. Some Generalizations on Kantorovich Inequality
Firstly, we utilize the properties of mathematical expectations to derive the following Kantorovich type inequality regarding mathematical expectations.
Theorem 1.
Let be random variables on probability space and there exist constants such that and , we then have
and
where and .
Proof.
(1). By homogeneity, we can assume . The proof of inequality (1) can be easily obtained for . We now only need to prove the situation for case .
Since
then at least one of the terms and must be strictly positive. Without generality, we now give the proof under , the other situation can be proved similarly.
Let , obviously . Then
Note that . Hence
Equivalently, we have
From Equations (3) and (5), it can be obtained that,
Similarly with (3) and
the following results holds,
Note that and , thus
For , it follows that from inequalities (6) and (7). By (8), we have Thus the result holds. For , note that the function is decreasing on when , and is increasing on when . Hence, by (6)–(8), the result holds.
(2). The left-hand side of inequality (2) holds via Cauchy-Schwarz inequality as . Now we prove the ride-hand side of inequality (2).
Let , thus and
Since and , then
Similarly,
If , then , Hence the result follows immediatedly from (10). The same conclusion can be derived if . Now we prove that the result holds when .
Below, we use Theorem 1 and the spectral decomposition of matrices to provide some Kantorovich type inequalities for positive definite matrices.
The basic matrix version of the kantorovich inequality [2,6,8,10,11,12] can be stated as follows. Let , and , then
Corollary 1.
Let , then
and
Proof.
The proof of the above two inequalities are quite similar, thus we only give the details of (14). Let the spectral decomposition of be
where are orthogonal projection matrix, are eigenvalues of A and B, separately, and . Thus,
and
Note that
Let , be the -fields composed of all subsets of . For any , define . Hence it’s easily to verified that
and . This completes the proof. □
Corollary 2.
Let satisfying that , then
Proof.
With homogeneity on x, we assume . Since , thus there exists an unitary matrix U such that
Let , then , and
Set , be the -fields composed of all subsets of . For any , let . We can easily get that
and . This completes the proof via Theorem 1. □
Remark 1.
Now we use inequality (2) and matrix spectral decomposition to derive a Kantorovich type inequality for positive linear functionals on matrix space.
Corollary 3.
Suppose Φ is a strictly positive linear functional on and , then
Proof.
Let the spectral decomposition of A be , where are orthogonal projection matrix, are eigenvalues of A. Thus, .
Note that
Let , be the -fields composed of all subsets of . For any , let . Hence it’s easily to verified that
and
This completes the proof by Theorem 1. □
Corollary 4.
Let Φ be a strictly positive linear functional on , and . Then
and
4. Some Generalizations on the Wielandt Inequality
Firstly, we use Theorem 1 to provide a Wielandt type inequality regarding mathematical expectations. The above corollary can be obtained if and are replaced by in Theorem 1. Note that .
Corollary 5.
Let be random variables on probability space with and there exist constants such that , then
The basic matrix version of the Wielandt inequality [13,14] is stated as follows. Let and . Then
Ref. [3] provided the following matrix form of the Wielandt inequality. Let are full column rank matrices such that , then
By variational method, [15] proved the following useful results. For , define , then
Now, we give more generalization on the Wielandt inequality.
Theorem 2.
Suppose such that , and , then
Proof.
Remark 2.
For full column rank matrix , let . Thus and From Theorem 2, the following corollary can be obtained.
Corollary 6.
From Theorem 2, it can been shown
where and .
Remark 3.
Remark 4.
If and are both full column rank matrix such that , then it follows from (31) that
Corollary 7.
Let be two invertible matrices of order n, and , such that
Then
5. Applications in Statistics
5.1. An Application on Correlation Coefficient
For any real matrix X with order , , where denotes the transpose of matrix X. Let be the correlation between two random variables and .
Suppose the covariance matrix of is denoted by . are the corresponding standard orthogonalized eigenvectors of . Thus , are called principal components of X. Note that the correlation coefficient between any two principal components should be zero in theory. However, in real data analysis, this condition is difficult to be satisfied as rounding error in actual calculations for the standard orthogonalized eigenvectors. Hence, it’s quite important to estimate the correlation coefficient [16] in real world, where . Here the variables are assumed standardized, which are scaled and centered using unit scaling.
We now give a useful bound of correlation coefficient .
Corollary 8.
If , then
Proof.
Since then the conclusion holds by replacing with in (32). □
Remark 5.
The result of [3] can be easily obtained from a special case of Corollary 8. Suppose a and b are orthogonal, the following result holds.
which leads to the result of [3].
5.2. An Application on Parameter Estimation
Consider the general linear regression model
with , X is an column full rank real matrix, is the parameter vector to be estimated. Without loss of generality, X can be assumed orthonormal, that is, the above model can be modified as
Let , hence , which leads to the standard orthonormal model,
If V is known, the best linear estimator of parameter is . It follows that
If V is unknown, the least squares estimator of parameter is . It follows that
For any real vector , the two estimators are unbiased. Their variances equal to
How to compare the two unbiased estimators is quite important in estimation theory. We now give several indexes to compare the relative efficiency of these estimators.
where index are based on the comparion of covariance matrix of and , and other index are based on the various comparion of variance of and . According to Corollary 8, we have the general results of the above index of relative efficiency as follows.
Corollary 9.
For index , we have
Proof.
The first two inequalities can be obtained by setting in Corollary 4. Similarly, the third and forth inequalities can be proved by setting Note that , then the last inequality can be inferred. □
It’s quite interesting to discuss whether the methods can be applied to partial least squares regression and penalized regression techniques, such as LASSO. However, the discussion becomes quite complex at the moment, and it leads to another direction for our future research.
Author Contributions
Conceptualization, Y.Z. and X.C.; Methodology, Y.Z. and J.L.; Validation, X.C.; Investigation, Y.Z., X.G. and J.L.; Writing—original draft, Y.Z.; Writing—review & editing, X.G., J.L. and X.C.; Supervision, J.L. and X.C.; Funding acquisition, X.C. All authors have read and agreed to the published version of the manuscript.
Funding
The work is supported by Natural Science Foundation of China (No. 12271270), Natural Science Foundation of Jiangsu Province of China (No. BK20200108), the third level training object of the sixth “333 project” in Jiangsu Province and the Zhongwu Youth Innovative Talent Program of Jiangsu University of Technology.
Data Availability Statement
The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Gustafson, K. The geometrical meaning of the Kantorovich-Wielandt inequalities. Linear Algebra Appl. 1999, 296, 143–151. [Google Scholar] [CrossRef][Green Version]
- Marshall, A.W.; Olkin, I. Matrix versions of Cauchy and Kantorovich inequalities. Aequationes Math. 1990, 40, 89–93. [Google Scholar] [CrossRef]
- Wang, S.G.; Cheung, I.W. A matrix version of the Wielandt inequality and its applications to statistics. Linear Algebra Appl. 1999, 296, 171–181. [Google Scholar] [CrossRef]
- Liu, S.Z.; Neudecker, H. A survey of Cauchy-Schwarz and Kantorovich-type matrix inequalities. Stat Pap. 1999, 40, 55–73. [Google Scholar] [CrossRef]
- Zhang, F. Equivalence of the Wielandt inequality and the Kantorovich inequality. Linear Multilinear Algebra 2001, 48, 275–279. [Google Scholar] [CrossRef]
- Moradi, H.R.; Gumus, I.H.; Heydarbeygi, Z. A glimpse at the operator Kantorovich inequality. Linear Multilinear Algebra 2019, 67, 1–6. [Google Scholar] [CrossRef]
- Chen, X.P.; Liu, J.Z.; Chen, J.D. A new result on recovery sparse signals using orthogonal matching pursuit. Stat. Theory Relat. Fields 2022, 6, 220–226. [Google Scholar] [CrossRef]
- Bhatia, R. Positive Definite Matrices; Princeton University Press: Princeton, NJ, USA, 2007; pp. 141–176. [Google Scholar]
- Horn, R.A.; Johnson, C.R. Matrix Analysis; Cambridge University Press: Cambridge, UK, 2012; pp. 108–126. [Google Scholar]
- Greub, W.; Rheinboldt, W. On a generalization of an inequality of LV Kantorovich. Proc. Am. Math. Soc. 1959, 10, 407–415. [Google Scholar] [CrossRef]
- Lin, M. On an operator Kantorovich inequality for positive linear maps. J. Math. Anal. Appl. 2013, 402, 127–132. [Google Scholar] [CrossRef]
- Sabancigil, P.; Kara, M.; Mahmudov, N.I. Higher order Kantorovich-type Szász—CMirakjan operators. J. Inequal. Appl. 2022, 2022, 1–15. [Google Scholar] [CrossRef]
- Wang, L.T.; Yang, H. Matrix Euclidean norm Wielandt inequalities and their applications to statistics. Stat. Pap. 2012, 53, 521–530. [Google Scholar] [CrossRef]
- Liu, S.Z.; Lu, C.Y.; Puntanen, S. Matrix trace Wielandt inequalities with statistical applications. J. Stat. Plan Inference 2009, 139, 2254–2260. [Google Scholar] [CrossRef]
- Yan, Z.Z. A unified version of Cauchy-Schwarz and Wielandt inequalities. Linear Algebra Appl. 2008, 428, 2079–2084. [Google Scholar] [CrossRef][Green Version]
- Liu, S.Z. Efficiency comparisons between the OLSE and the BLUE in a singular linear model. J. Stat. Plan Inference 2000, 84, 191–200. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).