1. Introduction
Hankel matrices are matrices constant along their antidiagonals (see
Section 2). They were named after the German mathematician Hermann Hankel. They were originally studied due to their elegant algebraic properties (cf. [
1]) and their relationship with Hankel operators (cf. [
2]), and later due to their important applications to many fields, such as signal processing, control theory, quantum mechanics and cryptography. For instance, in signal processing, the matrix structure of Hankel matrices makes them useful to represent the autocorrelation and cross-correlation of signals, as well as to analyze the spectral properties of signals. In control theory, the matrix structure of Hankel matrices makes them useful to represent the the input–output behavior of a system and to analyze and design control systems. Another class of matrices with important applications in many fields is the class of totally positive matrices, which lead to a framework of total positivity. A totally positive matrix is a matrix whose minors are all nonnegative (see
Section 2 for applications of these matrices). The relationship of Hankel matrices with the class of totally positive matrices (matrices with all their minors nonnegative) is a research field with interesting results. A related concept is given by coefficientwise totally positive matrices, where coefficientwise Hankel totally positive matrices play a key role (see [
3] and references therein). Another connection comes from the fact that the Hadamard product of Hankel strictly totally positive matrices is also strictly totally positive (cf. p. 123 of [
4] and [
5]). In addition, we note that Hankel totally positive matrices have been found in many fields, in particular (cf. [
6]), in the problem of finding algorithms with high relative accuracy (cf. [
7,
8,
9]). Algorithms with high relative accuracy have been found for some subclasses of TP matrices [
6,
10,
11,
12] in order to compute the inverse, all singular values, all eigenvalues or even the solution of some linear systems with high relative accuracy. This was obtained in [
6] for Hankel matrices that were Gramian matrices with respect to inner products defined for Hilbert spaces supported on bounded and unbounded intervals. Particular cases are Hilbert matrices that are well known in numerical linear Algebra.
A first goal of this paper is the presentation of some new contributions on the relationship of Hankel matrices and totally positive matrices. A second goal is related to the problem of the localization of eigenvalues of positive Hankel matrices. The localization of eigenvalues of a matrix is a classical subject, which is still actively considered (cf. [
13,
14,
15,
16]). Eigenvalue localization results are very useful for providing information on the eigenvalues as well providing an initial step for many numerical methods. These studies were performed in [
17] for Toeplitz matrices. Here, we perform the studies and observe that we obtain different results, which is not surprising because it is known that Hankel matrices have different properties from those of Toeplitz matrices.
The layout of the paper is as follows. In
Section 2, we introduce the classes of matrices considered in this paper as well as some basic notations. In
Section 3, we first construct an interval containing all the eigenvalues of positive Hankel matrices. With an extra condition, we also construct two intervals with an empty intersection inside the previous one and that also include all the eigenvalues. These two intervals can be considerably small and so very useful for eigenvalue localization. As we illustrate with a family of matrices, both intervals are required and they can be tight for the localization of the eigenvalues.
Section 4 analyzes the connections between positive Hankel matrices and totally positive matrices. We show that the strict total positivity of an
Hankel matrix can be characterized by the positivity of
minors. We also obtain a necessary condition for the total positivity of order 2 based on the parameters of a positive Hankel matrix. A very simple condition on the parameters that implies strict total positivity is also presented. We also relate positive Hankel matrices with a class of sign regular Toeplitz matrices. Finally, in
Section 5, we summarize the results and conclusions of this paper.
2. Basic Definitions and Notations
In this section, we define and give notations for the classes of matrices that will be considered in this work. We say that a matrix is positive if , and nonnegative if for all .
Given a sequence of
real numbers
, the
matrix
is called a
Hankel matrix. That is,
The numbers are called the parameters associated to the Hankel matrix A. Observe that Hankel matrices are symmetric matrices and so their eigenvalues are all real.
We say that a matrix with real entries is
totally positive (TP) when all its minors are nonnegative (see [
4,
18,
19,
20]). TP matrices are also called totally nonnegative matrices and they have important applications in many fields, such as approximation theory, finance, differential equations, combinatorics, economics, statistics, mechanics, quantum groups, Markov chains, biomathematics and computer-aided geometric design (cf. [
4,
18,
19,
21,
22,
23,
24,
25,
26,
27,
28,
29]). A nonnegative matrix is called TP
2 (or TP of order 2) if all its
submatrices have a nonnegative determinant. In [
19], several properties and applications of TP
2 matrices are presented. If a matrix has all its minors positive, then it is called
strictly totally positive (STP).
TP matrices also belong to the more general class of sign regular matrices, which we now introduce. A vector of signs
with
for
is called a
signature. An
matrix
A is
sign regular (SR) with signature
if, for each
, all minors of order
k have the same sign
or are zero. If
A is SR with signature
and all its minors are nonzero, then
A is called strictly SR (SSR). Sign regular matrices arise in many applications because, in the nonsingular case, they characterize the variation diminishing transformations (cf. [
18]).
The last class of matrices that will be mentioned in this paper is the class of Toeplitz matrices, which satisfy that all their diagonals are constant. Given a sequence of
real numbers
, an
Toeplitz matrix is defined by
, for all
. Toeplitz matrices have been applied to several fields, such as differential and integral equations, signal processing, statistics and physics (cf. [
30,
31])
We now introduce some matrix notations. For positive integers , , will denote the set of all increasing sequences of k natural numbers less than or equal to n. We also define the dispersion of by with the assumption for . Let A be a real matrix. For , , and for any and , denotes the submatrix of A containing rows numbered by and columns numbered by . A principal submatrix will be denoted and the leading principal submatrix of order k is given by the submatrix . An initial submatrix is either of the form with or of the form with . Using determinants, these definitions of submatrices lead to the corresponding definitions with minors.
3. Eigenvalue Localization of Positive Hankel Matrices
We start this section by introducing some basic notations useful for eigenvalue localization.
If
is a square matrix, we recall the known Gerschgorin circles (see [
16])). We define the
ith row Gerschgorin circle as
for each
. The column Gerschgorin circles can be defined in a similar way, but using columns instead of rows. By Gerschgorin’s theorem, all eigenvalues of a matrix lie in the union of all row Gerschgorin circles as well as in the union of all column Gerschgorin circles. As we have noted in a previous section, Hankel matrices are symmetric, and so their eigenvalues are real. Therefore, and taking into account that we consider symmetric and positive matrices, the row Gerschgorin circles reduce to the following intervals containing all eigenvalues of
A: for each
, the interval
In [
32], alternative inclusion regions were found for the real parts of the eigenvalues, and they can improve the information provided by Gerschgorin circles in a complementary way. Let us denote for each
and
The following result corresponds to Proposition 1 of [
17].
Proposition 1. Given a symmetric positive matrix and defined in (2), (3), all eigenvalues of A belong to the intersection between the union of row Gerschgorin intervals and the union of the intervals Moreover, for each , , .
In contrast to the previous inclusion intervals for the eigenvalues, we now consider an exclusion interval for the real eigenvalues of a real matrix presented in [
33]. From now on, for a real matrix
and for each
, we denote
Proposition 2.6 of [
33] provides the mentioned exclusion interval.
Theorem 1. Given a real matrix and of (5), we have for any real eigenvalue λ of A that The previous interval of (
6) will be named a
row exclusion interval.
Replacing A by , we can derive an exclusion column interval and derive an analogous result to the previous Theorem 1 for an exclusion column interval.
The following result corresponds to Proposition 2 of [
17].
Proposition 2. Given a symmetric positive matrix and of (5), we have that, for each , and , and the eigenvalues of A cannot belong to the interval Let us show first a very simple example showing the usefulness of Proposition 2. Let us consider the positive Hankel matrix with parameters equal to a positive real number
r. So,
A is the matrix with
k at any entry:
The eigenvalues of
A are
and 0. Then,
for all
I and the exclusion interval (
7) is
.
For general positive Hankel matrices, we shall combine the previous exclusion interval with some inclusion intervals in order to derive localization regions for the eigenvalues that can be very sharp.
Let us now take a positive Hankel matrix
A with parameters
. We need the next notations:
and also these notations indicating the minimal and maximal row sums of
A:
The next result has two parts. It gives an eigenvalue inclusion interval for all eigenvalues of a positive Hankel matrix and, in the second part, with an extra restriction, it also presents two disjoint intervals included in the previous one and such that all the eigenvalues belong to them. As we later show, both intervals are required and they can give tight localization for the eigenvalues of A.
Theorem 2. If is a positive Hankel matrix with parameters and are defined by (8) and (9), then all eigenvalues of A belong to the interval Furthermore, if we also have , then all eigenvalues of A are contained in the following union of two disjoint intervals Proof. Taking into account Proposition 1, the eigenvalues of
A belong to the intersection between the union of row Gerschgorin intervals and the union of the intervals given by (
4). Observe that the right endpoint of each interval of (
4) is the same right endpoint of the corresponding row Gerschgorin interval. Due to the positivity of
A, we have that the right endpoints of all intervals of (
4) coincide with the corresponding row sums and so they are less than or equal to Σ. Taking into account again the positivity of
A, each
ith left endpoint of each row Gerschgorin interval is provided by
minus the sum of the entries of the corresponding row and so they are greater than or equal to
. Once again, by the positivity of
A, the left endpoints of all intervals of (
4) are provided by the corresponding row sum minus
and, since the row sum is greater than or equal to
and
for all
, we deduce that all left endpoints of all intervals of (
4) are greater than or equal to
.
Let us now consider the second part of the statement. Hence, we also suppose that we have
. Given an eigenvalue
of
A,we have by Proposition 2 that
Due to the fact that
for all
, we have that
for all
. Thus,
and, since
We conclude that condition (
12) is transformed into
Taking into account our hypothesis, the exclusion interval of (
13) is nonempty and the result is proved. □
The next example provides a family of matrices depending on
such that the two intervals of (
11) converge to the two eigenvalues of the limit matrix of this family as
. This shows that both intervals are necessary and also the potential sharpness of our result.
Example 1. Let us take the family of Hankel matrices with positive parameters and , that is, As , matrices converge to the matrix with all entries equal to 1, whose eigernvalues are 0 and n. For matrices , it can be checked that the values of σ and Σ given by (9) are . This number is an eigenvalue for all matrices associated to the eigenvector . Therefore, the second interval of Theorem 2 coincides with this eigenvalue. Moreover, for all matrices , we have that . Hence, the first interval of Theorem 2 iswhich converges to , the other eigenvalue of the limit matrix, as . The following Section studies Hankel TP matrices, whose eigenvalues are all nonnegative due to their total positivity ([
18]).
4. Total Positivity and Hankel Matrices
In order to check whether a given matrix
A is STP, let us recall that a characterization considerably reducing the number of minors to consider is given by Theorem 4.1 of [
34], which only uses the initial minors defined in
Section 2. This characterization improved an old characterization by Fekete in a paper [
35] by Fekete and Polya that used the positivity of all minors
with
.
Theorem 3. An matrix A is STP if, and only if, all its initial minors are positive.
For a Hankel matrix
A, it is known that a characterization (cf. Theorem 4.4 of [
4]) shows the equivalence of the strict total positivity of
A with the positive definiteness of
A and a symmetric submatrix of
A.
Theorem 4. An Hankel matrix A is STP if, and only if, the symmetric matrices A and are positive definite.
In the case of Hankel matrices, the following result uses very few minors, reducing the characterization of strict total positivity to the positivity of minors.
Theorem 5. An Hankel matrix A is STP if, and only if, its leading principal minors and the leading principal minors of are positive.
Proof. Since an STP matrix has all its minors positive, it is sufficient to prove that the positivity of the leading principal minors of the statement implies that A is STP. As is well known, a symmetric matrix is positive definite if, and only if, all its leading principal minors are positive. Since matrices A and are symmetric, both are positive definite. Now, the result follows from Theorem 1. □
The following result for positive Hankel matrices with first parameters
provides a necessary condition for being TP
2. The condition is that the whole sequence of parameters is monotone increasing, extending the property satisfied by the two first parameters. We can see that it is a very simple condition to assure that the matrix is TP
2, which is a property that is very important in many applications (cf. [
19,
27]).
Proposition 3. Let A be a positive Hankel matrix with parameters and with . If A is TP2, then .
Proof. Since
A is TP
2,
. Since
and
, we deduce from the previous inequality that
Analogously, since , and , we can deduce that . Continuing in a similar way, and using that and , we prove that all parameters are positive and that . Hence, the result follows. □
We now present a sufficient condition for the strict total positivity of a positive Hankel matrix using a condition derived in [
36], which, in turn, slightly improves a condition of [
37] (see Corollary 2.17 of [
4]). This sufficient condition only uses
inequalities for an
Hankel matrix. Since each inequality involves two products and one quotient, this condition only requires
products,
divisions and
comparisons, and it guarantees the positivity of all minors of the matrix.
Proposition 4. If A is a positive Hankel matrix with parameters satisfying for all , then A is STP.
Proof. Observe that, due to the form of
A, the sufficient condition of Corollary 2.17 of [
4] is of the form
for all
, which follows from our hypothesis. □
Hankel TP matrices are not as closely related with Toeplitz TP matrices as one might imagine. In fact, they have many different properties. However, there is a class of sign regular (SR) matrices that is very close to the class of Hankel TP matrices. Our following result relates Hankel TP matrices with that special class of Toeplitz SR matrices. For this purpose, we shall consider the matrix P, which is the reverse of the identity matrix (i.e., it has 1 in entries and 0 elsewhere).
We also need to recall the following result (see Theorem 3.1 of [
18]), which claims that the product of SR matrices is an SR matrix.
Proposition 5. Let A and B be two SR matrices with signature and , respectively. Then, the product is a SR matrix with signature . If, in addition, A is nonsingular and B is SSR, the is SSR.
We can now give the result relating Hankel TP matrices with Toeplitz SR matrices with a special signature.
Proposition 6. An matrix A is a Hankel TP (respectively, STP) if, and only if, is Toeplitz SR (respectively, SSR) with signature given by for .
Proof. Let us observe that an TP (respectively, STP) matrix is an SR (respectively, SSR) matrix whose signature is with for all and that P is a nonsingular SR matrix whose signature is with for , where by , we denote the greatest integer less than or equal to x.
If A is Hankel, then is Toeplitz and, if A is TP (respectively, STP), then by Proposition 5, we have that is an SR (respectively, SSR) matrix with the signature of P.
For the converse, since is the identity matrix, observe that, if is Toeplitz with the signature of P, then is Hankel and again by Proposition 5, we have that A is SR with the signature vector formed by 1’s, that is, A is TP. □
The importance of this result has two aspects. On the one hand, results on Toeplitz SR matrices with the signature given in Proposition 6 lead to corresponding results on the class of Hankel TP matrices. On the other hand, new results on Hankel TP matrices imply corresponding results on Toeplitz matrices with the signature given in Proposition 6. This last aspect is even more important than the other one because the properties of SR matrices are usually much more difficult to obtain than the properties of TP matrices. For instance, there is no result analogous to Theorem 3 for the class of SSR matrices. In fact, checking that a matrix is SSR involves a considerably greater number of minors (cf. [
18,
35]) than the number of minors used in Theorem 3. In contrast, we can see that only
are required to check whether a Toeplitz matrix is SSR with the signature given in Proposition 6 when we use Theorem 5.