Abstract
The paper is a survey of the recent results of the author on the perturbations of matrices. A part of the results presented in the paper is new. In particular, we suggest a bound for the difference of the determinants of two matrices which refines the well-known Bhatia inequality. We also derive new estimates for the spectral variation of a perturbed matrix with respect to a given one, as well as estimates for the Hausdorff and matching distances between the spectra of two matrices. These estimates are formulated in the terms of the entries of matrices and via so called departure from normality. In appropriate situations they improve the well-known results. We also suggest a bound for the angular sectors containing the spectra of matrices. In addition, we suggest a new bound for the similarity condition numbers of diagonalizable matrices. The paper also contains a generalization of the famous Kahan inequality on perturbations of Hermitian matrices by non-normal matrices. Finally, taking into account that any matrix having more than one eigenvalue is similar to a block-diagonal matrix, we obtain a bound for the condition numbers in the case of non-diagonalizable matrices, and discuss applications of that bound to matrix functions and spectrum perturbations. The main methodology presented in the paper is based on a combined usage of the recent norm estimates for matrix-valued functions with the traditional methods and results.
Keywords:
matrices; perturbations; spectral variation; Hausdorff distance between eigenvalues; matching distance between spectra MSC:
15A15; 15A18; 15A42
1. Introduction
This paper is a survey of the recent results of the author on perturbations of the eigenvalues and determinants of matrices.
Finding the eigenvalues of a matrix is not always an easy task. In many cases it is easier to calculate the eigenvalues of a nearby matrix and then to obtain the information about the eigenvalues of the original matrix.
The perturbation theory of matrices has been developed in the works of R. Bhatia, C. Davis, L. Elsner, A.J. Hoffman, W. Kahan, T. Kato, L. Mirsky, A. Ostrowski, G.W. Stewart, J.G. Sun, H.W. Wielandt, and many other mathematicians.
To recall some basic results of the perturbation theory, which will be discussed below, let us introduce the notations.
Let be the n-dimensional complex Euclidean space with a scalar product the norm and unit matrix I. denotes the set of complex -matrices. For an , is the adjoint matrix, is the inverse one, is the spectral norm: , are the eigenvalues of A taken with their multiplicities, is the spectrum, is the resolvent, is the trace, is the determinant, is the spectral radius, and is the Schatten-von Neumann norm; in particular, is the Hilbert-Schmidt (Frobenius) norm.
Let A and be -matrices whose eigenvalues counted with their multiplicities are and , respectively. The following result is well-known.
where , cf. [1] (p. 107). The spectral norm is unitarily invariant, but often it is not easy to compute that norm, especially if the matrix depends on many parameters. In Section 4 below we present a bound for in terms of the entries of matrices in the standard basis. That bound can be directly calculated. Moreover, under some conditions our bound is sharper than (1).
Recall some definitions from matrix perturbation theory (see [2] (p. 167)).
The spectral variation of with respect to A is .
The Hausdorff distance between the eigenvalues of A and is
The matching (optimal) distance between eigenvalues of A and is
where is taken over all permutations of .
The quantity is not a metric: it may be zero, even when the eigenvalues of A and are different (e.g., when and while ).
Geometrically, the spectral variation has the following interpretation. If
then
In other words, the eigenvalues of lie in the union of disks of radius centered at the eigenvalues of A.
The Hausdorff distance hounds the spectral variation and is actually a metric. The matching distance bounds the Hausdorff distance and is also a metric. The “smallness” of the matching distance means that the eigenvalues of a matrix and its perturbation are “close” and they can be grouped into nearby pairs. In some cases bounds on the spectral variation or the Hausdorff distance can be converted into bound on the matching distance.
One of the well-known bounds for is the Elsner inequality
References [1,2,3]. Since the right hand part of this inequality is symmetric, we have
As it was mentioned, the calculations and estimating of the spectral norm is often a not easy task. Below we suggest bounds for the spectral variation and Hausdorff distance explicitly expressed via the entries of the considered matrices. In some cases our bounds are sharper than (3).
By inequality (3) the following result called the Ostrowski–Elsner theorem has been proved:
cf. [2] (p. 170, Theorem IV.1.4). In Section 7, we consider also other bounds for .
Put
where ranges over all permutations of the integers .
One of the famous results on is the Hoffman-Wiellandt theorem proved in [4] (see also [2] (p. 189) and [5] (p. 126)), which asserts the following: for all normal matrices A and , the inequality is valid.
In [6], L. Mirsky has proved that for all Hermitian matrices A and ,
(see also [2] (p. 194) and [5] (p. 126)). In 1975, W. Kahan [7] (see also [2] (Theorem IV.5.2, p. 213)) has derived the following result: let A be a Hermitian matrix and an arbitrary one in , and
Then
Here and below , , . The Kahan theorem generalizes the Mirsky result in the case . In Section 14 we present an analogous result for a .
Furthermore, as is well-known, the Hilbert identity
plays an important role in the perturbation theory. In Section 15, we suggest a new identity for resolvents and show that it refines the results derived with the help of the Hilbert identity, if the commutator has a sufficiently small norm.
A few words about the contents of the paper. It consists of 17 Sections.
In Section 2, we recall some classical results which are needed our proofs. In Section 3, we present norm estimates for resolvents of matrices which will be applied in the sequel.
In Section 4 and Section 5, we derive the perturbation bound for determinants in terms of the entries of matrices and consider some its applications. Section 6 deals with perturbation bounds for determinants expressed via rather general norms.
Section 7, Section 8, Section 9 and Section 10 are devoted to the spectral variations. Besides, the relevant bounds are obtained in terms of the departure from normality and via the entries of matrices.
Section 11 and Section 12 deal with angular localization of matrices. The results of Section 12 are new.
Section 13 is devoted to perturbations of diagonalizable matrices. Besides, we suggest a bound for the condition numbers. Besides, Corollary 14 is new.
As it was above mentioned, in Section 14 we generalize the Kahan result.
In Section 16 and Section 17, taking into account that any matrix having more than one eigenvalue is similar to a block-diagonal matrix, we obtain a bound for the condition numbers in the case of non-diagonalizable matrices, and discuss applications of that bound to matrix functions and spectrum perturbations. The material of Section 16 and Section 17 is new.
2. Preliminaries
Recall the Schur theorem Section I.4.10.2 of [8], By that theorem there is an orthogonal normal (Schur’s) basis , in which A has the triangular representation
Schur’s basis is not unique. We can write
with a normal (diagonal) operator D defined by
and a nilpotent operator V defined by
Equality (5) is called the triangular representation of A; D and V are called the diagonal part and nilpotent part of A, respectively. Put
is called the maximal chain of the invariant projections of A. It has the properties
with and
So and D have the joint invariant subspaces. We can write
where .
Let us recall also the famous Gerschgorin theorem [2] and Section III.2.2.1 of [8], which is an important tool for the analysis of the location of the eigenvalues.
Theorem 1.
The eigenvalues of lie in the union of the discs
The Gerschgorin theorem implies the following inequality for the spectral radius:
3. Norm Estimates for Resolvents
The following quantity (the departure for normality) of A plays an essential role hereafter:
By Lemma 3.1 from [9] , where V is the nilpotent part of A (see equality (5)). Therefore, if A is a normal matrix, then . The following relations are checked in Section 3.1 of [9]:
and
By the inequality between the arithmetic and geometric means we have
Hence,
If and are commuting matrices, then . Indeed, since and commute, they can have a joint basis of the triangular representation. So the nilpotent part of is equal to where and are the nilpotent parts of and , respectively. Therefore,
We will need the following
Theorem 2
(Theorem 3.1 of [9]). Let . Then
where
This Theorem sharp: if A is a normal matrix, then and we obtain . Here and below we put .
Let us recall an additional norm estimate for the resolvent, which is sharper than Theorem 2 but more cumbersome. To this end, for an integer introduce the numbers
Here
are binomial coefficients. Evidently, for all ,
Theorem 3
(Theorem 3.10 of [9]). Let . Then
Moreover, the following result is valid.
Theorem 4
(Theorem 3.4 of [9]). Let . Then
Let us point to an inequality between the resolvent and determinant.
Theorem 5.
For any and all regular λ of A one has
For the proof see, for example Corollary 3.4 of [9].
4. Perturbation Bounds for Determinants in Terms of the Entries of Matrices
The following theorem is valid.
Theorem 6
(Reference [10]). Let , be an arbitrary orthonormal basis in and . Then
and, therefore,
Proof.
It is not hard to check that is a polynomial in and
Thanks to the Cauchy integral,
Hence,
Take into account that
Consequently,
In addition, according to (10)
Therefore, due to (13),
Taking , we get (10), as claimed. □
Obviously are directly calculated. Below we also show that in the concrete situations Theorem 6 is sharper than (1) and enables us to establish sharp upper and lower bounds for the determinants of matrices that are “close” to triangular matrices.
Furthermore, making use of the inequality between the arithmetic and geometric means, from (11) we get
Put . Then by the latter inequality
Or
where
Denote . Then
Let us check that
Indeed, the derivative of the function on the left-hand-side is
Hence it follows that the infimum is reached at . This proves (14).
So we can write
We thus arrive at our next result.
Corollary 1.
Let and be an arbitrary orthonormal basis in . Then we have
5. Perturbations of Triangular Matrices and Comparison with Inequality (1)
In this section, , , and is the standard basis. Clearly,
and
Now Theorem 6 implies
Corollary 2.
One has
and, therefore,
Furthermore, let be the upper triangular part of A, i.e.,
where if and for . Then
Clearly,
Making use of Corollary 2, we arrive at our next result.
Corollary 3.
One has
where
From this corollary we have
Moreover, if
then
Inequalities (15) and (17) are sharp: they are attained if A is triangular.
Recall that is the Frobenius norm of A.
The following lemma taken from Lemma 3.3 of [10] gives us simple conditions, under which (11) is sharper than (1).
Lemma 1.
If
then (11) is sharper than (1).
Proof.
By the Cauchy inequality,
Since , we easily have
Thus,
Now Corollary 3 implies
Since
We get
Thus, if (18) holds, then (17) improves (1). □
It should be noted that the determinants of diagonally dominant and double diagonally dominant matrices are very well explored, cf. [11,12,13,14]. At the same time the determinants of matrices “close” to triangular ones are investigated considerably less than the determinants of diagonally dominant matrices. About bounds for determinants of matrices close to the identity matrix see the papers [15].
6. Perturbation Bounds for Determinants in Terms of an Arbitrary Norm
Let be an arbitrary fixed matrix norm of , i.e., the the function from into , defined by the usual relations: for the zero matrix , if , , and
In addition, . So, . Therefore, there is a number , such that
We need the following result.
Theorem 7
(Theorem 1.7.1 of [16]). Let and condition (19) hold. Then
where
Recall that is the Schatten-von Neumann norm. Making use of the inequality between the arithmetic and geometric mean values, we obtain
Due to the Weyl inequalities
cf. Corollary II.3.1 of [17], Lemma 1.1.4 of [16], we get
So in this case
where
Now Theorem 7 implies
Corollary 4.
Let . Then for any finite ,
Note that Theorem 8.1.1 from the book [16] refines the Weyl inequality with the help of the self-commutator.
Furthermore, let
i.e., W is the off-diagonal part of A: . Then taking and making use of the previous corollary, we arrive at the following result.
Corollary 5.
Let . Then
7. Bounds for the Spectral Variations in Terms of the Departure from Normality
In this section, we estimate the spectral variation of two matrices in terms of the departure from normality introduced in Section 3. The results of the present section are based on the norm estimates for resolvents presented in Section 3 and the following technical lemma.
Lemma 2.
Let A and be linear operators in and . In addition, let
where is a monotonically increasing continuous function of a non-negative variable x, such that and . Then , where is the unique positive root of the equation
For the proof see Section 1.8 of [9]. Lemma 2 and Theorem 2 with
imply
Theorem 8.
Let A and be -matrices and . Then , where is the unique positive root of the equation
Since where (see Section 2), one can replace in (21) by
If A is normal, then , we have and, therefore, Theorem 8 gives us the well-known inequality , cf. [1,2]. Thus, Theorem 8 refines the Elsner inequality (3) if A is “close” to normal.
Equation (21) can be written as
To estimate one can apply the well-known known bounds for the roots of polynomials. For instance, consider the algebraic equation
with non-negative coefficients .
Lemma 3.
The unique positive root of (23) satisfies the inequality
Proof.
Since all the coefficients of are non-negative, it does not decrease as increases. If , then and . Hence . If , then
and , as claimed. □
Substitute into (22), assuming that A is non-normal, i.e., . Then we obtain the equation
Putting
and applying Lemma 3 for the unique positive root of (24), we obtain
But ; consequently, according to Theorem 8, we get
Furthermore, put
Then Theorem 8 implies
Corollary 6.
One has , where is the unique positive root of the equation
Replacing in Corollary 6 by , we obtain the following result.
Corollary 7.
We have
Now we are going to derive an estimate for the matching distance introduced in Section 1. To this end we need the following well-known result.
Theorem 9
(Theorem IV.1.5, p. 170 in [2]). Let and . If is a nondecreasing bound on , then
If is a nondecreasing bound on , then
Here is the integer part of .
Note that for any . By (25),
Hence,
Making use of Theorem 9, we arrive at
Corollary 8.
Let . Then
Since for a normal matrix A, , Corollary 8 refines the Ostrowski–Elsner theorem mentioned in Section 1 for matrices close to normal ones.
8. A Bound for the Spectral Variation Via the Entries of Matrices
As mentioned above, the spectral norm is unitarily invariant, but the calculations and estimating of the spectral norm is often a not easy task, especially if the matrix depends on many parameters. In the paper [18], a bound for the spectral variation has been explicitly expressed via the entries of the considered matrices. In the paper [19], we have established a new bound via the entries. In the appropriate situations it considerably improves Elsner’s inequality and the main result from [18]. In this section we present the main results from [19].
Theorem 10.
Let and be matrices. Then with the notations
one has
where
The proof of this theorem is presented in the next section. Simple calculations show that
Furthermore, let be the upper triangular part of A. i.e., , where if and for . To illustrate Theorem 10 apply it with and , taking into account that
, where . In addition, , where
Now Theorem 10 implies.
Put
Since is triangular, we have . Making use of (26), we arrive at
Corollary 9.
All the eigenvalues of lie in the set .
This corollary is sharp: if A is triangular, then , and Corollary 9 gives us the equalities .
9. Proof of Theorem 10
In this section for the brevity put and .
Lemma 4.
Let and be an arbitrary orthonormal basis in . Then for any eigenvalue of we have
where
and
Proof.
Due to Theorem 6,
Hence,
Since , (28) implies
Consequently,
as claimed. □
Proof. of Theorem 10.
Obviously,
Therefore,
Now let be the standard basis, and A and be represented in that basis by matrices and , respectively. Clearly,
So . By the Gerschgorin theorem (see Section 2), we have . Thus,
Consequently, under consideration . Now Lemma 4 implies
Since the right-hand part does not depend on j, this finishes the proof. □
10. Comments and Examples to Theorem 10
Again is the spectral norm of A. To compare Theorem 10 with the Elsner inequality (3) consider the following examples.
Example 1.
Let , .
Then . Now the Elsner inequality implies
Since , , Theorem 10 yields the inequality
Obviously, (32) is sharper than (31).
Example 2.
Let
Simple calculations give us the following results: , , and . Hence, . To apply Theorem 10 note that in the considered example . So Theorem 10 gives us the following result:
and, therefore, .
Furthermore, under consideration , and thus the Elsner inequality implies
So (33) is sharper than this result.
Example 3.
Let
By the standard calculations we get , , and . Hence, . In the considered example . Omitting simple calculations, by Theorem 8.1, we get , and, therefore, .
11. Angular Localization of the Eigenvalues of Perturbed Matrices
In this section we consider the following problem: let the eigenvalues of a matrix lie in a certain sector. In what sector do the eigenvalues of a perturbed matrix lie?
Not too many works are devoted to the angular localization of matrix spectra. The papers [20,21] should be mentioned. In these papers it is shown that the test to determine whether all eigenvalues of a complex matrix of order n lie in a certain sector can be replaced by an equivalent test to find whether all eigenvalues of a real matrix of order lie in the left half-plane. Below we also recall the well-known results from Chapter 1, Exercise 32 of [22].
To the best of our knowledge, the problem just described of angular localization of the eigenvalues of perturbed matrices was not considered in the available literature, although it is important for various applications, cf. [22].
The results of this section are adopted from the paper [23].
Again, is the spectral norm of . For a we write if Y is positive definite, i.e., .
Without loss of the generality, we assume that
If this condition does not hold, instead of A we can consider perturbations of the matrix with a constant .
By the Lyapunov theorem, cf. Theorem I.5.1 of [22], condition (34) implies that there exists a positive definite , such that . Define the angular Y-characteristic of A by
The set
will be called the Y-spectral-sector ofA. Let be an eigenvalue of A and d the corresponding eigenvector: . Then
We, thus, get
Lemma 5.
For an , let condition (34) hold and Y be a positive definite matrix, such that . Then, any eigenvalue of A lies in the Y-spectral-sector of
Example 4.
Let . Then condition (34) holds. For any commuting with A (for example ) we have and . Thus and .
So Lemma 5 is sharp.
Remark 1.
Suppose that A is invertible. Recall that the quantity defined in the finite-dimensional case by
is called the angular deviation of A, cf. Chapter 1, Exercise 32 of [22]. For example, for a positive definite operator A one has
where , are the boundary of the spectrum of A (see Chapter 1, Exercise 33 of [22]).
In Exercise 32, it is shown that the spectrum of A lies in the sector Since , Lemma 5 refines the that inequality.
Furthermore, by the above mentioned Lyapunov theorem, there exists a positive definite solving the Lyapunov equation
Hence,
Put
Now we are in a position to formulate the main result of this section.
Theorem 11.
Let , condition (34) hold and X be a solution of (35). Then, with the notation , one has
provided
The proof of this theorem is based on the following lemma
Lemma 6.
Let , condition (34) hold and X be a solution of (35). If, in addition,
then
Proof.
Put . Then and due to (35), with we obtain
In addition,
But
Hence
Now (39) yields.
provided (38) holds. Since
according to (36) we arrive at the required result. □
Proof of Theorem 11.
Note that X is representable as
Section 1.5 of [22]. Hence, we easily have . Now the latter lemma proves the theorem. □
12. An Estimate for J (A) and Examples to Theorem 11
Lemma 7.
Let condition (34) hold. Then , where
Proof.
By virtue of Example 3.2 from [9],
Then
as claimed. □
If A is normal, then and, taking we have .
The latter lemma and Theorem 11.1 imply
Corollary 10.
Let and the conditions (34) and hold. Then
Now consider the angular localization of the eigenvalues of matrices “close” to triangular ones. Let be the upper triangular part of A. i.e., , where if and for . To illustrate our results apply Corollary 10 with A instead of and with instead of A.
Since is triangular, we have ,
and . Assuming that we can write
In addition, . Now Corollary 12.2 implies.
Corollary 11.
Let and the condition
hold. Let the diagonal entries of A lie in the sector . Then the eigenvalues of A lie in the sector with ψ satisfying
Example 5.
Consider the matrix
Then
We have , where . and, therefore, . In addition, , and consequently,
Hence,
Now Corollary 11 implies that the eigenvalues of the considered matrix A lie in the sector with satisfying
The direct calculations show that .
13. Perturbations of Diagonalizable Matrices
An eigenvalue is said to be simple, if its geometric multiplicity is equal to one. In this section, we consider a matrix A whose all the eigenvalues are simple. As it is well known, in this case there is an invertible matrix T, such that
where is a normal matrix. Besides, A is called a diagonalizable matrix. The condition number is very important for various applications. We obtain a bound for the condition number and discuss applications of that bound to matrix functions and spectral variations.
If is diagonalizable, it can be written as
where are one-dimensional eigen-projections. If is a scalar function defined on the spectrum of A, then is defined as
Let
be the interpolation Lagrange-Sylvester polynomial, such that . and
cf. Section V.1 of [24]. From (40) it follows
Since is normal, . We thus arrive at
Lemma 8.
Let A be diagonalizable and be a scalar function defined on the for an . Then
In particular,
Inequality (41) and Lemma 7.1 imply.
Corollary 12.
Let and A be diagonalizable. Then
Now we are going to estimate the condition number of A assuming that all the eigenvalues of A are different:
In other words the algebraic multiplicity of each eigenvalue is is equal to one. Recall that
(see Section 3) and put
and
Theorem 12.
Let condition (42) be fulfilled. Then there is an invertible matrix T, such that (40) holds with
The proof of this theorem can be found in Theorem 6.1 of [9] and [25]. Theorem 12 is sharp: if A is normal, then and . Thus we obtain the equality .
Lemma 8 and Theorem 12 immediately imply.
Corollary 13.
Let condition (42) hold and be a scalar function defined on the for an . Then
Moreover, making use of Theorem 12 and Corollary 12, we arrive at the following result.
Corollary 14.
Let and condition (42) hold. Then
About additional inequalities for condition numbers via norms of the eigen-projections see [26,27]. About the functions of diagonalzable matrices see also [28].
14. Sums of Real Parts of Eigenvalues of Perturbed Matrices
The aim of the present section is to generalize the Kahan inequality (4). Again, put , and . Let be a sequence of positive numbers defined by the recursive relation
For a , put
As it is proved in Corollary 1.3 of [29],
Now we in a position to formulate and prove the main result of this section.
Theorem 13.
Let be a Hermitian operator and be an arbitrary matrix. Let the conditions
hold. Then for any ,
Proof.
According to the Schur theorem (see Section 2), we can write
where is an upper triangular matrix. Since and are similar, they have the same eigenvalues, and without loss of generality we can assume that is already upper triangular, i.e.,
where is the diagonal matrix and is the strictly upper triangular matrix.
Here and below denotes the spectrum of A. We have and thus, the real and imaginary part of A are
respectively. Since A and are Hermitian, by the Mirsky inequality mentioned in the Introduction, we obtain
Thus
Making use of Lemma 1.5 from [29], we get the inequality
(see also Section 3.6 of [30,31]). In addition, by (48) and, therefore,
Thanks to the above mentioned Weyl inequalities,
Thus,
Now (50) implies the inequality
So by (49) we get the desired inequality
□
The just proved theorem is sharp in the following sense: if is Hermitian, then and inequality (47) becomes the Mirsky result, presented in Section 1.
Corollary 15.
Let a matrix have the real diagonal entries. Let W be the off-diagonal part of : . Then for any ,
and, therefore,
Indeed, this result is due to the previous theorem with .
Certainly, inequality (51) has a sense only if its right-hand side is positive.
The case should be considered separately from the case , since the relations between and similar to inequality (50) are unknown if , and we could not use the arguments of the proof of Theorem 13. The case is investigated in [32].
15. An Identity for Resolvents
Let and . The Hilbert identity for resolvents mentioned in Section 1 gives the following important result: if a is regular for A and
then is also regular for . In this section we suggest a new identity for resolvents of matrices. It gives us new perturbation results which in appropriate situations improve condition (52). Put .
Theorem 14.
Let a be regular for A and . Then,
Proof.
We have
as claimed. □
Denote
Lemma 9.
Let be a regular point of A and . Then and identity (53) holds. Moreover,
Proof .
Put . Since the regular sets of operators are open, for t small enough, is a regular point of . By the previous lemma we get
Hence,
Thus, with the notation
We have
Take an integer and put . For m large enough, is a regular point of and due to (54) we can write
Hence,
where . Due to inequality (55) we can assert that . So in our arguments we can replace by and obtain the relations
Therefore, . Continuing this process for , we get . Now (54) implies the required result. □
It is clear that , where
Now the previous lemma yields the following result.
Corollary 16.
Let and Then and relation (53) holds.
Example 6.
Let us consider the matrices
with arbitrary non-zero numbers a and It is clear that , . In this example we easily have and and, therefore, Corollary 16 gives us the sharp result.
At the same time (52) gives us the invertibility condition .
Example 7.
Let us consider the block matrices
where C and B are commuting -matrices. It is simple to check that
Corollary 16 gives us the equality . At the same time, due to (52), if we can assert that only if .
If A is invertible, then due to Theorem 5,
Now Corollary 16 implies
Corollary 17.
Suppose A is invertible, and
then is also invertible.
Recall that the quantity is introduced in Section 2. Theorems 2 and Corollary 16 imply our next result.
Corollary 18.
If λ is regular for A and
then λ is regular for .
The following theorem gives us the bound for the spectral variation via the identity for resolvents considered in this section.
Theorem 15.
Let A and be matrices. Then , where is the unique positive root of the algebraic equation
Proof.
For any , due to Corollary 18 we have
Hence, it follows that where is the unique positive root of the equation
which is equivalent to (56). But . This proves the theorem. □
To estimate one can apply Lemma 13.
16. Similarity of an Arbitrary Matrix to a Block Diagonal Matrix
16.1. Preliminary Results
Again, is the spectral norm. and is the Frobenius norm of , are the different eigenvalues of A and is the algebraic multiplicity of . So
and . The aim of this section is to show that there are matrices and an invertible matrix , such that
Besides, each block has the unique eigenvalue . In addition, we obtain an estimate for the (block-condition) number and consider some applications of that estimate,
Put
By the Schur theorem (see Section 2) there is a non-unique unitary transform, such that A can be reduced to the triangular form:
Besides, the diagonal entries are the eigenvalues ordered enumerated as
Let be the corresponding orthonormal basis of the upper-triangular representation (the Schur basis). Denote
and
In addition, put and . We can see that each is an orthogonal invariant projection of A and
Besides, if , then and is one dimensional. If , then
where
In the matrix form the blocks can be written as
etc. Besides, each is a strictly upper-triangular (nilpotent) part of . So has the unique eigenvalue of the algebraic multiplicity : . We, thus, have proved the following result.
Lemma 10.
An arbitrary matrix can be reduced by a unitary transform to the block triangular form (59) with , where is either a nilpotent operator, or . Besides, has the unique eigenvalue of the algebraic multiplicity .
16.2. Statement of the Main Result
Again, put
Introduce, also, the notations
and
It is not hard to check that . Now we are in a position to formulate the main result of this section.
Theorem 16.
Let an -matrix A have different eigenvalues of the algebraic multiplicity . Then there are -matrices each of which has a unique eigenvalue , and an invertible matrix T, such that (58) holds with the block-diagonal matrix . Moreover,
This theorem is proved in the next section. Theorem 16 is sharp: if A is normal, then and . Thus we obtain the equality .
16.3. Applications of Theorem 16
Let be a scalar function, regular on . Define by the usual way via the Cauchy integral [33]. Since are mutually orthogonal, we have
Let
be the interpolation Lagrange–Sylvester polynomial such that and , cf. Section V.1 of [24].
Now (58) implies
Hence, (59) and (60) yield
Corollary 19.
Let . Then there is an invertible matrix T, such that
Due to Theorem 3.5 from the book [9] we have
Take into account that (see Section 17). Now, making use of Theorem 16.2, we arrive at the following result.
Corollary 20.
Let . Then
For example, we have
where .
About the recent results devoted to matrix-valued functions see for instance [9] and the references which are given therein.
Now consider the resolvent. Then by (58) for we have
Extending this relation analytically to all regular z and taking into account that
We get
Corollary 21.
Let . Then there is an invertible matrix T, such that
for any regular z of A.
But due to Theorem 3.2 from [9] we have
where is the distance between z and the spectrum of A. Clearly, . Now Theorem 16 and (62) imply
Corollary 22.
Let . Then
Furthermore, let A and be complex -matrices. Recall that is the spectral variation of with respect to A.
For the proof see Lemma 1.10 of [9]. Making use of Lemma 2 and Corollary 22, we obtain the inequality , where is the unique positive root of the equation
This equation is equivalent to the algebraic one
For example, if
then due to Lemma 3.17 from [9], we have . So we arrive at
Corollary 23.
Let A and be -matrices. Then . If, in addition, condition (64) holds, then .
To illustrate Corollary 23 consider the matrices
The eigenvalues of A are . So ,
, and . Hence,
where . According (59) consider the equation
So one can take where
Due to Corollary 23 we have .
Additional relevant results can be found in the papers [34,35].
17. Proof of Theorem 16
Recall that are the orthogonal invariant projections defined in Section 16.1 and ; and are also defined in Section 16.1. Put
By Lemma 10 has the unique eigenvalue and A is represented by (59). Represent and in the block-matrix form:
and
Since is a block triangular matrix, it is not hard to see that
cf. Lemma 6.2 of [9]. So due to Lemma 10,
Under this condition, the equation
has a unique solution
e.g., Section VII.2 of [33].
Lemma 11.
Let be a solution to (66). Then
Proof.
Due to (67) we can write . But . Therefore, and
Since is a projection invariant to A: , we can write . Thus, and, consequently,
Furthermore, . Hence,
Therefore,
Consequently,
Continuing this process and taking into account that , we obtain
as claimed. □
Take
According to (69)
So the matrix is inverse to . Thus,
and (68) can be written as (58). We thus arrive at
Corollary 24.
Let an -matrix A have different eigenvalues of the algebraic multiplicity . Then there are -matrices each of which has a unique eigenvalue and such that (58) holds with T defined by (70).
By the inequalities between the arithmetic and geometric means from (70) and (71) we get
and
Proof of Theorem 16
Consider the Sylvester equation
where and are given; should be found. Assume that the eigenvalues and of B and , respectively, satisfy the condition.
Then Equation (74) has a unique solution X, e.g., Section VII.2 of [33]. not mentioned in the article, plaese confirm and modifiy. Due to Corollary 5.8 of [9], the inequality
is valid and, therefore,
where .
Let us go back to Equation (66). In this case , , , , , and due to (57), . In addition, . Now (75) implies
where .
Recall that denotes the Schur basis. So
We can write with a normal (diagonal) matrix defined by and a nilpotent (strictly upper-triangular) matrix defined by , . and will be called the diagonal part and nilpotent part of A, respectively. It can be , i.e., A is normal.
Besides, . In addition, the nilpotent part of is and the nilpotent part of is . So and are orthogonal, and
Thus, from (76) it follows
It can be directly checked that
and
Since , we have
and, consequently,
Take T as is in (70). Then (72), (73), and (77) imply
and
But by the Schwarz inequality and (78),
Thus,
and . Now (68) proves the theorem. □
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Bhatia, R. Perturbation Bounds for Matrix Eigenvalues; Classics in Applied Mathematics; SIAM: Philadelphia, PA, USA, 2007; Volume 53. [Google Scholar]
- Stewart, G.W.; Sun, J.G. Matrix Perturbation Theory; Academic Press: New York, NY, USA, 1990. [Google Scholar]
- Elsner, L. On optimal bound for the spectral variation of two matrices. Linear Algebra Appl. 1985, 71, 77–80. [Google Scholar] [CrossRef]
- Hoffman, A.J.; Wiellandt, H.W. The variation of the spectrum a normal matrix. Duke Math. J. 1953, 20, 37–39. [Google Scholar] [CrossRef]
- Kato, T. Perturbation Theory for Linear Operators; Springer: Berlin, Germany, 1966. [Google Scholar]
- Mirsky, L. Symmetric gage functions and unitarily invariant norms. Q. J. Math. 1960, 11, 50–59. [Google Scholar] [CrossRef]
- Kahan, W. Spectra of nearly Hermitian matrices. Proc. Am. Math. Soc. 1975, 48, 11–17. [Google Scholar] [CrossRef]
- Marcus, M.; Minc, H. A Survey of Matrix Theory and Matrix Inequalities; Allyn and Bacon: Boston, MA, USA, 1964. [Google Scholar]
- Gil’, M.I. Operator Functions and Operator Equations; World Scientific: Hackensack, NJ, USA, 2018. [Google Scholar]
- Gil’, M.I. Perturbations of determinants of matrices. Linear Algebra Its Appl. 2020, 590, 235–242. [Google Scholar] [CrossRef]
- Li, H.B.; Huang, T.Z.; Li, H. Some new results on determinantal inequalities and applications. J. Inequalities Appl. 2010, 2010. [Google Scholar] [CrossRef]
- Li, B.; Tsatsomeros, M.J. Doubly diagonally dominant matrices. Linear Algebra Its Appl. 1997, 261, 221–235. [Google Scholar] [CrossRef]
- Wen, L.; Chen, Y. Some new two-sided bounds for determinants of diagonally dominant matrices. J. Inequal. Appl. 2012, 61, 1–9. [Google Scholar]
- Vein, R.; Dale, P. Determinants and Their Applications in Mathematical Physics; Applied Mathematical Sciences; Springer: New York, NY, USA, 1999; Volume 134. [Google Scholar]
- Brent, R.P.; Osborn, J.H.; Smith, W.D. Note on best possible bounds for determinants of matrices close to the identity matrix. J. Linear Algebra Appl. 2015, 466, 21–26. [Google Scholar] [CrossRef]
- Gil’, M.I. Bounds for Determinants of Linear Operators and Their Applications; CRC Press: Boca Raton, MA, USA; Taylor & Francis Group: London, UK, 2017. [Google Scholar]
- Gohberg, I.C.; Krein, M.G. Introduction to the Theory of Linear Nonselfadjoint Operators; American Mathematical Society: Providence, RI, USA, 1969; Volume 18. [Google Scholar]
- Gil’, M.I. A new inequality for the Hausdorff distance between spectra of two matrices. Rend. Del Circ. Mat. Palermo Ser. 2 2020, 70, 341–348. [Google Scholar] [CrossRef]
- Gil’, M.I. A refined bound for the spectral variations of matrices. Acta Sci. Math. 2021, 87, 1–6. [Google Scholar]
- Anderson, B.D.; Bose, N.K.; Jury, E.I. A simple test for zeros of a complex polynomial in a sector. IEEE Trans. Aufomat. Contr. 1974, 19, 437–438. [Google Scholar] [CrossRef]
- Anderson, B.D.; Bose, N.K.; Jury, E.I. On Eigenvalues of complex matrices in a sector. IEEE Trans. Aufomat. Contr. 1975, 20, 433. [Google Scholar] [CrossRef]
- Daleckii, Y.L.; Krein, M.G. Stability of Solutions of Differential Equations in Banach Space; American Mathematical Society: Providence, RI, USA, 1974. [Google Scholar]
- Gil’, M.I. On angular localization of spectra of perturbed operators. Extr. Math. 2020, 35, 197–204. [Google Scholar] [CrossRef]
- Gantmakher, F.R. Theory of Matrices; Nauka: Moscow, Russia, 1967. (In Russian) [Google Scholar]
- Gil’, M.I. A bound for condition numbers of matrices. Electr. J. Linear Algebra 2014, 27, 162–171. [Google Scholar] [CrossRef][Green Version]
- Gil’, M.I. Estimates for functions of finite and infinite matrices. perturbations of matrix functions. Int. J. Math. Game Theory Algebra 2013, 21, 328–392. [Google Scholar]
- Gil’, M.I. On condition numbers of spectral operators in a Hilbert space. Anal. Math. Phys. 2015, 5, 363–372. [Google Scholar] [CrossRef]
- Gil’, M.I. Norm estimates for functions of matrices with simple spectrum. Rend. Circ. Mat. Palermo 2010, 59, 215–226. [Google Scholar] [CrossRef]
- Gil’, M.I. Lower bounds for eigenvalues of Schatten-von Neumann operators. J. Inequal. Pure Appl. Math. 2007, 8, 117–122. [Google Scholar]
- Gohberg, I.C.; Krein, M.G. Theory and Applications of Volterra Operators in Hilbert Space; American Mathematical Society: Providence, RI, USA, 1970; Volume 24. [Google Scholar]
- Gil’, M.I. Operator Functions and Localization of Spectra; Lecture Notes In Mathematics; Springer: Berlin, Germany, 2003; Volume 1830. [Google Scholar]
- Gil’, M.I. Sums of real parts of perturbed matrices. Math. Ineq. 2010, 4, 517–522. [Google Scholar] [CrossRef]
- Bhatia, R. Matrix Analysis; Springer: New York, NY, USA, 1997. [Google Scholar]
- Gil’, M.I. Resolvents of operators on tensor products of Euclidean spaces. Linear Multilinear Algebra 2016, 64, 699–716. [Google Scholar] [CrossRef]
- Gil’, M.I. On Similarity of an Arbitrary Matrix to a Block Diagonal Matrix; Filomat: Nis, Serbia, 2021; (Accepted for Publication). [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).