1. Introduction
Random matrices have been widely used in many problems, e.g., compressed sensing [
1], high-dimensional data analysis [
2], matrix approximation [
3,
4] and dimension reduction [
5]. In the literature, one of main research issues is to study the deviation behavior of eigenvalues (or singular values) of random matrices.
In general, there are two types of deviation results studied in probability theory: one is the large-deviation inequality that describes the behavior of the probability for large t; and the other is the small-deviation (or small-ball) inequality that controls the probability for small .
The early large-deviation inequalities for sums of random matrices can be dated back to the work of Ahlswede and Winter [
6]. Tropp [
7] improved their results and developed a user-friendly framework to obtain the large-deviation inequalities for sums of random matrices. To overcome the limitation of the matrix-dimension dependence, Hsu et al. [
8] and Minsker [
9] introduced the concepts of intrinsic dimension and effective dimension to tighten large-deviation inequalities, respectively. Moreover, Zhang et al. [
10] applied a diagonalization method to obtain the dimension-free large-deviation random for largest singular value of sums of random matrices, while it remains a challenge to select the auxiliary matrices and functions. In the scenario of a single random matrix, Ledoux [
11] studied the largest eigenvalues of Gaussian unitary ensemble matrices and Vershynin [
12] studied the singular values of the sub-Gaussian and sub-exponential matrices.
Small-deviation problems stemmed from some practical applications, e.g., approximation problems [
13], Brownian pursuit problems [
14], quantization problem [
15], and convex geometry [
16]. For more details, we refer to the bibliography maintained by Lifshitz [
17]. There have been some works on the small-deviation inequalities for the specific types of random matrices. Aubrun [
18] obtained the small-deviation inequalities for the largest eigenvalue of a single Gaussian unitary ensemble matrix. Rudelson and Vershynin [
19] presented the small-deviation inequalities for the smallest singular value of the random matrix with independent entries. Volodko [
20] estimated the small-deviation probability of the determinant of the matrix
, where
is a
random matrix whose entries obey a centered joint Gaussian distribution. To the best of our knowledge, there are few works on the small-deviation inequalities for sums of random matrices.
1.1. Related Works
Let
be a finite sequence of independent random Hermitian matrices. It follows from Markov’s inequality that
where
denotes the largest eigenvalue and
stands for the expectation operation. By using Golden-Thompson inequality, Ahlswede and Winter [
6] bounded the trace of the matrix moment generating function (mgf) in the following way:
where
stands for the trace of the matrix
. By applying Lieb’s concavity theorem, Tropp [
7] achieved a tighter matrix mgf bound than the above one:
where “the eigenvalue of sum of matrices” is smaller than “the sum of eigenvalues of matrices” in the right-hand side of (
1). However, there remains a shortcoming that the result (
2) is dependent with the matrix dimension
d, and its right-hand side will become loose for high-dimensional matrices.
To overcome the shortcoming, Hsu et al. [
8] employed the intrinsic dimension
to replace the ambient dimension
d in the case of real symmetric matrices. Minsker [
9] provided a dimension-free version of Bernstein’s inequality for sequences of independent random matrices. Zhang et al. [
10] introduced a diagonalization method to obtain the tail bounds for the largest singular values of the sum of random matrices. Although their bounds are independent of the matrix dimension and overcame the aforementioned first shortcoming, there remains a challenge to select the appropriate parameters to obtain the tighter bounds.
There are also some small-deviation results on one single random matrix. Edelman [
21] presented the small-deviation behavior of the smallest singular value of a Gaussian matrix:
where
is a
random matrix whose entries are independent standard normal random variables. Rudelson and Vershynin [
22] studied the small-deviation bound of the smallest singular value of a sub-Gaussian matrix:
where
,
only depends on the sub-Gaussian moment of its entries and
is a
random matrix whose entries are i.i.d. sub-Gaussian random variables with zero mean and unit variance. However, to the best of our knowledge, there is little work on the small-deviation inequalities for sums of random matrices.
1.2. Overview of Main Results
In this paper, we present the small-deviation inequalities for the largest eigenvalue for sums of independent random Hermitian matrices, i.e., the upper bound of
In particular, we first present some basic small-deviation results of random matrices. We then obtain several types of small-deviation inequalities for the largest eigenvalue of sums of independent random positive semi-definite (PSD) matrices. In contrast to the large-deviation inequalities for random matrices, the resulting small-deviation inequalities are independent of the matrix dimension d and thus our finding are applicable to the high-dimensional and even infinite-dimensional cases.
The rest of this paper is organized as follows. In
Section 2, we introduce some useful notations and then give some basic results on small-deviation inequalities for random matrices. The small-deviation results for sums of random PSD matrices are presented in
Section 3. The last section concludes the paper.
2. Basic Small-Deviation Inequalities for Random Matrices
In this section, we first introduce the necessary notations and then present some basic small-deviation results of random matrices.
2.1. Necessary Notations
Given a Hermitian matrix , denote and as the largest and the smallest eigenvalues of , respectively. Denote and as the trace and the spectral norm of , respectively. Let be the identity matrix, be the unitary matrix and stand for the Hermitian adjoint of .
By the spectral mapping theorem, given a real-value function
, then
where
is a diagonalization of
. If
for
when the eigenvalues of
lie in
I, then there holds that
, where the semi-definite partial order ⪯ is defined as follows:
2.2. Basic Small-Deviation Inequalities for Random Matrices
Subsequently, we produce the small-deviation inequalities for random matrices. First, we consider a small-deviation bound for one single matrix:
Lemma 1. Let be a random Hermitian matrix. Then for any , Proof. The last inequality holds because the minimum eigenvalue of a positive definite (pd) matrix is dominated by the . Since this inequality holds for any , taking an infimum over completes the proof. □
Then, by using the subadditivity of the matrix cumulant generating function (see [
7], Lemma 3.4), we obtain the small-deviation bound for sums of random matrices:
Theorem 1. Let be a finite sequence of independent random Hermitian matrices. Then for any , Proof. By combining Lemma 1 and Lemma 3.4 of [
7], we have for any
,
Taking the infimum over completes the proof. □
Please note that the above small-deviation bound is independent of the matrix dimension d, and thus it is applicable to the scenarios of high-dimensional and even infinite-dimensional matrices. In addition, we also derive the following small-deviation bounds for sums of random matrices.
Corollary 1. Let be a sequence of independent random Hermitian matrices. Assume that there is a function and a sequence of fixed Hermitian matrices such that - 1.
Define the scalar parameter If , then for any , - 2.
Define the scalar parameter If , then for any ,
Proof. It follows from (
4) that
and substituting it into Theorem 1 leads to the result (
5). Then, the fact that
for any Hermitian matrix
leads to the result (
6). This completes the proof. □
By using the logarithm operation, we then obtain another small-deviation bound for sums of random matrices:
Corollary 2. Let be a sequence of independent random Hermitian matrices. Then for any , Proof. Since the matrix logarithm is operator concave, for each
, we have
According to (
3), we then arrive at
Since the trace of a matrix can be bounded by d times of its maximum eigenvalue, taking the infimum over completes the proof. □
The following presents the relationship between one random PSD matrix and a sum of PSD random matrices.
Lemma 2. Let be a sequence of independent random Hermitian PSD matrices. Then for all k and any , Proof. Since
are PSD, we have
The last inequality holds for any . This completes the proof. □
This lemma shows that the small-deviation probability for sums of random matrices can be bounded by using the small-deviation probability for one single matrix. This fact suggests that the small-deviation bound could be independent of the size of matrix sequence, while this phenomenon will not arise in the large-deviation scenario.
3. Small-Deviation Inequalities for Positive Semi-Definite Random Matrices
In this section, we present several types of small-deviation inequalities for the largest eigenvalue of sums of independent random PSD matrices. Similar to the scalar version of small-deviation inequalities, there remains a challenge to bound the term . Here, we adapt some methods to handle this issue.
First, we introduce the negative moment estimate for the largest eigenvalue to derive a small-deviation inequality for sums of random matrices:
Theorem 2. Let be a sequence of independent random Hermitian PSD matrices. Given a , if there exists a positive constant such thatthen there holds that for any , Proof. It follows from Jensen’s inequality that
Then, the Markov’s inequality yields
This completes the proof. □
In this theorem, we impose an assumption that the negative moment of is bounded. In general, this assumption is mild, and can be satisfied in most cases. The following small-deviation results are derived under that condition that the eigenvalues of the matrices are bounded:
Theorem 3. Let be a sequence of independent random Hermitian PSD matrices such that () almost surely. Then for any ,where Furthermore, there holds that for any ,where Proof. For any
and
, there holds that
According to transfer rule, we have,
By substituting (
9) into Corollary 1, we then have,
since
. The infimum is achieved at
, which leads to the result of (
7).
Moreover, the combination of Lemma 1 and (
9) leads to
Then, the result (
8) is derived from Lemma 2. This completes the proof. □
Actually, the above results are derived from the geometric point of view, where the term is bounded by the linear function for any . Finally, we study the small-deviation inequalities for random matrix series , which is a sum of fixed Hermitian pd matrices weighted by random variables .
Theorem 4. Let be a sequence of fixed Hermitian pd matrices, and be a finite sequence of independent variables. If there exist the constants and such that for all ,then there holds that for any ,where Furthermore, for any ,where Proof. According to transfer rule, we have,
By substituting (
13) into Corollary 2, we then have,
The infimum will be attained at
, and it leads to the result of (
11). Moreover, the combination of Lemma 1 and (
13) leads to
Then, the result (
12) is resulted from Lemma 2. This completes the proof. □
The above results hold under the condition (
10) that
has a power-type upper bound of
. This condition is mild, and we refer to [
23] for the details. Moreover, to keep the results (
11) and (
12) non-trivial, their right-hand sides should be less than
one, and thus we arrive at
and
, respectively.
4. Conclusions
In this paper, we present the small-deviation inequalities for the largest eigenvalues of sums of random matrices. In particular, we first give some basic results on small-deviation inequalities for random matrices. We then study the small-deviation inequalities for sums of independent random PSD matrices. In contrast to the large-deviation inequalities for random matrices, our results are independent of the matrix dimension
d and thus can be applicable to the scenarios of high-dimensional and even infinite-dimensional matrices. In addition, by using the Hermitian dilation (see Section 2.6 of [
7]), our small-deviation results can also be extended to the scenario of non-Hermitian random matrices.
Author Contributions
Conceptualization, X.G. and C.Z.; methodology, X.G. and C.Z.; validation, X.G., C.Z. and H.Z.; resources, C.Z.; writing—original draft preparation, X.G.; writing—review and editing, C.Z.; supervision, C.Z. and H.Z.; project administration, C.Z. and H.Z.; funding acquisition, C.Z.
Funding
This research was funded by National Natural Science Foundation of China grant number 61473328 and 11401076.
Acknowledgments
We are grateful to the anonymous reviewers and the editors for their valuable comments and suggestions.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Chandrasekaran, V.; Recht, B.; Parrilo, P.A.; Willsky, A.S. The convex geometry of linear inverse problems. Found. Comput. Math. 2012, 12, 805–849. [Google Scholar] [CrossRef]
- Bühlmann, P.; van de Geer, S. Statistics for High-Dimensional Data: Methods, Theory and Applications; Springer: Berlin, Germany, 2011. [Google Scholar]
- Gittens, A.; Mahoney, M.W. Revisiting the nyström method for improved large-scale machine learning. J. Mach. Learn. Res. 2016, 17, 3977–4041. [Google Scholar]
- Halko, N.; Martinsson, P.G.; Tropp, J.A. Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev. 2011, 53, 217–288. [Google Scholar] [CrossRef]
- Clarkson, K.L.; Woodruff, D.P. Low rank approximation and regression in input sparsity time. In Proceedings of the forty-fifth annual ACM symposium on Theory of computing, Palo Alto, CA, USA, 1–4 June 2013; pp. 81–90. [Google Scholar]
- Ahlswede, R.; Winter, A. Strong converse for identification via quantum channels. IEEE Trans. Inf. Theory. 2002, 48, 569–579. [Google Scholar] [CrossRef]
- Tropp, J.A. User-friendly tail bounds for sums of random matrices. Found. Comput. Math. 2012, 12, 389–434. [Google Scholar] [CrossRef]
- Hsu, D.; Kakade, S.M.; Zhang, T. Tail inequalities for sums of random matrices that depend on the intrinsic dimension. Electron. Commun. Prob. 2012, 17, 1–13. [Google Scholar] [CrossRef]
- Minsker, S. On some extensions of bernstein’s inequality for self-adjoint operators. Stat. Probabil. Lett. 2017, 127, 111–119. [Google Scholar] [CrossRef]
- Zhang, C.; Du, L.; Tao, D. Lsv-based tail inequalities for sums of random matrices. Neural. Comput. 2017, 29, 247–262. [Google Scholar] [CrossRef] [PubMed]
- Ledoux, M. Deviation inequalities on largest eigenvalues. In Geometric Aspects of Functional Analysis; Milman, V.D., Schechtman, G., Eds.; Springer: Berlin, Germany, 2007; pp. 167–219. [Google Scholar]
- Vershynin, R. Introduction to the non-asymptotic analysis of random matrices. arXiv 2010, arXiv:1011.3027. [Google Scholar]
- Li, W.V.; Linde, W. Approximation, metric entropy and small ball estimates for gaussian measures. Ann. Probab. 1999, 27, 1556–1578. [Google Scholar]
- Li, W.V.; Shao, Q.M. Capture time of brownian pursuits. Probab. Theory. Rel. 2001, 121, 30–48. [Google Scholar] [CrossRef]
- Dereich, S.; Fehringer, F.; Matoussi, A.; Scheutzow, M. On the link between small ball probabilities and the quantization problem for gaussian measures on banach spaces. J. Theor. Probab. 2003, 16, 249–265. [Google Scholar] [CrossRef]
- Klartag, B.; Vershynin, R. Small ball probability and dvoretzky’s theorem. ISR J. Math. 2007, 157, 193–207. [Google Scholar] [CrossRef]
- Lifshits, M. Bibliography of Small Deviation Probabilities. Available online: https://www.lpsm.paris/pageperso/smalldev/biblio.pdf (accessed on 15 December 2016).
- Aubrun, G. A sharp small deviation inequality for the largest eigenvalue of a random matrix. In Séminaire de Probabilités XXXVIII; Springer: Berlin, Germany, 2005; pp. 320–337. [Google Scholar]
- Rudelson, M.; Vershynin, R. Non-asymptotic theory of random matrices: extreme singular values. arXiv 2010, arXiv:1003.2990. [Google Scholar]
- Volodko, N.V. Small deviations of the determinants of random matrices with gaussian entries. Stat Probabil. Lett. 2014, 84, 48–53. [Google Scholar] [CrossRef]
- Edelman, A. Eigenvalues and condition numbers of random matrices. SIAM J. Matrix Anal. A 1988, 9, 543–560. [Google Scholar] [CrossRef]
- Rudelson, M.; Vershynin, R. The littlewood–offord problem and invertibility of random matrices. Adv. Math. 2008, 218, 600–633. [Google Scholar] [CrossRef]
- Li, W.V. Small Value Probabilities in Analysis and Mathematical Physics. Available online: https://www.math.arizona.edu/~mathphys/school_2012//WenboLi.pdf (accessed on 15 March 2012).
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).