1. Introduction
In the free vibration of a
q-degree mass-damping-spring structure, the system of differential equations for describing the motion is [
1]
where
is a time-dependent
q-dimensional vector to signify the generalized displacements of the system.
In engineering application the mass matrix
and the stiffness matrix
are positive definite because they are related to the kinetic energy and elastic strain energy. However, the damping properties of a system reflected in the viscous damping matrix
are rarely known, making it difficult to be evaluated exactly [
2,
3].
In terms of the vibration mode
, we can express the fundamental solution of Equation (
1) as
which leads to a nonlinear eigen-equation for
:
where
is a matrix quadratic of the structure. Equation (
3) is a quadratic eigenvalue problem (QEP) to determine the eigen-pair
. In addition to the free vibrations of mechanical structures, many related systems which lead to the quadratic eigenvalue problems were discussed by Tisseur and Meerbergen [
4]. In the design of linear structure, knowing natural frequency of the structure is vital important to avoid the resonance with external exciting. The natural frequency is the imaginary part of the eigenvalue
in Equation (
3), which is a frequency to easily vibrate the structure [
5,
6].
Using the generalized Bezout method [
7] to tackle Equation (
3) one introduces a solvent matrix
determined by
where
is a
zero matrix, such that a factorization of
reads as
It gives us an opportunity to determine the eigenvalues of Equation (
3) through the eigenvalues solved from two
q-dimensional eigenvalue problems, generalized one and standard one:
The key point of the solvent matrix method is solving a nonlinear matrix Equation (
4) to obtain an accurate matrix
[
8,
9]. Most of the numerical methods that deal directly with the quadratic eigenvalue problem by solving Equations (
4) and (
6) are the variants of Newton’s method [
10,
11,
12,
13]. These Newton’s variants converge when the initial guess is close enough to the solution. But even for a good initial guess there is no guarantee that the method will converge to the desired eigenvalue. There are different methods to solve the quadratic eigenvalue problems [
14,
15,
16,
17,
18,
19].
Let
be the generalized velocity of vibration mode. We can combine Equations (
7) and (
3) together as
Equation (
8) becomes a generalized eigenvalue problem for the
n-vector
:
where
with
. Equation (
10) is used to determine the eigen-pair
, which is a linear eigen-equation associated to the pencil
, where
is an eigen-parameter.
In the linearization from Equation (
3) to Equation (
10), an unsatisfactory aspect is that the dimension of the working space is doubled to
. However, for the generalized eigenvalue problems many powerful numerical methods are available [
20,
21]. The numerical computations in [
22,
23] revealed that the methods based on the Krylov subspace can be very effective in the nonsymmetric eigenvalue problems by using the Lanczos biorthogonalization algorithm and the Arnoldi’s algorithm. The Arnoldi and nonsymmetric Lanczos methods are both of the Krylov subspace methods. Among the many algorithms to solve the matrix eigenvalue problems the Arnoldi method [
23,
24,
25,
26], the nonsymmetric Lanczos algorithm [
27], and the subspace iteration method [
28] are well known. The affine Krylov subspace method was first developed by Liu [
29] to solve linear equations system, which is however not yet used to solve the generalized eigenvalue problems. It is well known that the eig function in MATLAB is an effective code to compute the eigenvalues of the generalized eigenvalue problems. We will test it for the eigenvalue problem of highly ill-conditioned matrices and point out its limitation.
The paper sets up a new detection method for determining the real and complex eigenvalues of Equation (
10) in
Section 2, where the idea of exciting vector and excitation method (EM) are introduced with two examples being demonstrated. In
Section 3, we express the generalized eigenvalue problem (
10) in an affine Krylov subspace, and a nonhomogeneous linear equations system is derived. To precisely locate the position of the real eigenvalue in the response curve, we develop an iterative detection method (IDM) in
Section 4, where we derive new methods to compute eigenvalue and eigenvector. In
Section 5 some examples of the generalized eigenvalue problems are given. The EM and IDM are extended in
Section 6 to directly solve the quadratic eigenvalue problem (
3), where we derive IDM in the eigen-parametric plane to determine the complex eigenvalue. In
Section 7, several examples are given and solved by either linearizing them to the generalized eigenvalue problems or treating them in the original quadratic forms by the direct detection method. Finally, the conclusions are drawn in
Section 8.
2. A New Detection Method
Let
be the set of all the eigenvalues of Equation (
10), which may include the pairs of conjugate complex eigenvalues. It is known that if
then Equation (
10) has only the trivial solution with
and
. In contrast, if
then Equation (
10) has a non-trivial solution with
and
, which is called the eigenvector. However, when
n is large it is difficult to directly solve Equation (
10) to determine
and
by the manual operations. Instead, numerical methods have to be employed to solve Equation (
10), from which
is always obtained to be a zero vector no matter which
is, since the right-hand side of
is zero.
To definitely determine
to have a finite magnitude with
, we consider a variable transformation from
to
by
where
is a given nonzero vector, which being inserted into Equation (
10) generates
It is important that the right-hand side is not zero because of
, which is a given exciting vector to render a nonzero response of
and then
by Equation (
11) is available. We must emphasize that when
is near to a singular matrix, we cannot eliminate
in Equation (
12) by inverting the matrix
to obtain
.
2.1. Real Eigenvalue
Let the eigen-parameter in Equation (
12) run in an interval
, and we can solve Equation (
12) by the Gaussian elimination method to obtain
and then
. Hence, the response curve is formed by varying the magnitude
vs. the eigen-parameter
in the interval
. It is different from Equation (
10) that now we can compute
in Equation (
12) and then
by Equation (
11), when
is a given nonzero vector. Through this transformation by solving a nonhomogeneous linear system (
12), rather than the homogeneous linear system (
10), the resultant vector
can generate a nonzero finite response of
when the eigen-parameter tends to an eigenvalue, and for most eigen-parameters that not near to any eigenvalue the responses of
are very small, nearly close to zero. The technique to construct a nonzero response curve is called an excitation method (EM).
For instance, we consider Equation (
10) endowing with [
20]:
We take
and
, and we plot
vs. the eigen-parameter
in
Figure 1a, wherein only zero values of
appear.
However, under a nonzero excitation with
, we plot
vs. the eigen-parameter
in
Figure 1b, where we can observe five peaks of the response curve which signifying the locations of five real eigenvalues to be sought. When
does not locate at the peak point,
is zero from the theoretical point of view; however, the value of
as shown in
Figure 1b is not zero, which is very small due to the machinery round-off error caused by using the Gaussian elimination method to solve Equation (
12). In
Section 4, we will develop an iterative method to precisely locate those eigenvalues based on the EM.
2.2. Complex Eigenvalue
Because
and
are real matrices, the eigenvalue may be a complex number, which is assumed to be
where
, and
and
are, respectively, the real and imaginary parts of
. Correspondingly, we take
Inserting Equations (
14) and (
15) into Equation (
10), yields
Letting
Equation (
16) becomes
Similarly, by taking
where
is a given exciting vector and by Equation (
18), we have
No matter which
is, since Equation (
20) is a consistent linear system with a dimension
, we can solve it by using the Gaussian elimination method to obtain
and then
by Equation (
19).
When and take values inside a rectangle by , we can plot vs. over the eigen-parametric plane, and investigate the property of the response surface.
For instance, for
we have a pair of complex eigenvalues:
We take
,
, and with
, we plot
over the plane
in
Figure 2, where we can observe one peak near to the point
. More precise values of
will be obtained by the iterative detection method to be developed in
Section 6.
4. An Iterative Detection Method for Real Eigenvalue
It follows from Equation (
29) that
which is the projection of Equation (
10) into the affine Krylov subspace
as a nonhomogeneous projected eigen-equation. Upon comparing to the original eigen-equation
in Equation (
10), Equation (
30) is different for the appearance of a nonhomogeneous term
in the right-hand side, When we take
in the right-hand side,
is a zero vector solved by the numerical method and thus
. To excite the nonzero response of
, we must give a nonzero exciting vector
in the right-hand side.
In
Section 2, we have given an example to show that the peaks of
at the eigenvalues are happened in the response curve of
vs.
, which motivates us using a simple maximum method to determine the eigenvalue by collocating points inside an interval by
where the size of
must be sufficiently large to include at least one eigenvalue
. Therefore, the numerical procedures for determining the eigenvalue of a generalized eigenvalue problem are summarized as follows. (i) Select
m,
,
a and
b. (ii) Construct
. (iii) For collocating point
solving Equation (
30), setting
and taking the optimal
to satisfy Equation (
31).
By repeating the use of Equation (
31), we gradually reduce the size of the interval centered at the previous peak point by renewing the interval to a finer one. Give an initial interval
, and we place the collocating points by
and pick up the maximal point denoted by
. Then, a finer interval is given by
and
, which is centered at the previous peak point
and with a smaller length
. In that new interval we pick up the new maximum point denoted by
. Continuing this process until
, we can obtain the eigenvalue with high accuracy. This algorithm is shortened as an iterative detection method (IDM).
In order to construct the response curve, we choose a large interval of to include all eigenvalues, such that the rough locations of the eigenvalues can be observed in the response curve as the peaks. Then, to precisely determine the individual eigenvalue, we choose a small initial interval to include that eigenvalue as internal point. A few iterations by the IDM can compute very accurate eigenvalue.
When the eigenvalue
is obtained, if one wants to compute the eigenvector, we can normalize a nonzero
th component of
by
. Let
where
and
are the components of
and
, respectively. Then, it follows from Equation (
10) an
-dimensional linear system:
where
are constructed by
We can apply the Gaussian elimination method or the conjugate gradient method to solve
in Equation (
33). And then
is computed by
To evaluate the accuracy of the obtained eigenvalue
and eigenvector
, we can investigate the error of
to satisfy the eigen-equation (
10). We will extend the above IDM to detect the complex eigenvalue in
Section 6.
5. Examples of Generalized Eigenvalue Problems
Example 1. We considerand . The two smallest eigenvalues are . It is a highly ill-conditioned generalized eigenvalue problem due to Cond. We take
and plot
vs.
in
Figure 3, where we can observe two peaks happened at two eigenvalues. The zigzags in the response curve are due to the ill-conditioned
in Equation (
10).
Starting from
and under a convergence criterion
, with seven iterations we can obtain
, which is very close to −0.619402940600584 with a difference
. On the other hand, the error to satisfy Equation (
10) is very small with
, where
is solved from Equations (
32)–(
35) with
.
Starting from and with six iterations, we can obtain , which is very close to 1.627440079051887 with a difference , and is obtained.
Example 2. In Equation (10), and are given in Equation (13). We take and plot vs. λ in Figure 4, where five peaks in the response curve happen at five eigenvalues. Starting from
and under a convergence criterion
, through five iterations
is obtained.
is obtained, where
is solved from Equations (
32)–(
35) with
.
Starting from and under a convergence criterion , we can obtain and after seven iterations.
With and seven iterations, we can obtain and .
With and seven iterations, we can obtain and .
With and seven iterations, we can obtain and .
By comparing the response curve in
Figure 4 with that in
Figure 1b, the numerical noise is disappeared by using the IDM in the affine Krylov subspace.
Example 3. Let and , . In Equation (10), we take [21]: Other elements are all zeros, where we take . The eigenvalues are for .
We take
and plot
vs.
in
Figure 5a, where five peaks are obtained by the IDM, which correspond to the first five eigenvalues. Within a finer interval
in
Figure 5b, one peak appears at a more precise position.
Starting from
and with one iteration under a convergence criterion
, the eigenvalue obtained is very close to 0.01 with an error
.
is obtained, where
is solved from Equations (
32)–(
35) with
.
With and , after two iterations the eigenvalue obtained is very close to 0.02 with an error . is obtained.
For this problem, if we take and to compute the eigenvalue in the full space, the computational time is increased. The CPU time spent to construct the response curve is 6.09 s. For the eigenvalue 0.01 the error is the same with , but the CPU time increases to 5.83 s. In contrast, the subspace method spent 1.05 s to construct the response curve, and the CPU time for the eigenvalue 0.01 is 1.05 s. When n is increased to , the subspace method with spent 1.05 s to construct the response curve and the CPU time for the eigenvalue 0.01 is 6.53 s; however, by using the full space method with the CPU time is 44.91 s and and CPU time for the eigenvalue 0.01 is 44.67 s. Therefore, the computational efficiency of the m-dimensional affine Krylov subspace method is better than that by using the full space method with dimension n.
To further test the efficiency of the proposed method, we consider the eigenvalue problem of Hilbert matrix [
31]. In Equation (
10), we take
and
Due to highly ill-conditioned nature of the Hilbert matrix, it is a quite difficult eigenvalue problem. For this problem, we take , and to compute the largest eigenvalue, which is given as 2.182696097757424. The affine Krylov subspace method is convergent very fast with six iterations, and the CPU time is 1.52 s. The error of eigen-equation is . However, finding the largest eigenvalue in the full space with dimension , it does not converge within 100 iterations, and the CPU time is increased to 33.12 s. By using the MATLAB, we obtain 2.182696097757423 which is close to that obtained by the Krylov subspace method. However, the MATLAB leads to a large error of Det, which indicates that the characteristic equation for the Hilbert matrix is highly ill-posed. Notice that the smallest eigenvalue is very difficult to be computed, since it is very close to zero. However, we can obtain the smallest eigenvalue with one iteration and the error is obtained. For the eigenvalue problem of the Hilbert matrix, the MATLAB leads to a wrong eigenvalue , which is negative and contradicts to the positive eigenvalues of the Hilbert matrix. The eig function in Matlab cannot guarantee to obtain a positive eigenvalue for the positive definite Hilbert matrix. The first 41 eigenvalues are all negative. The first 73 eigenvalues are less than . So most of these eigenvalues computed by Matlab should be spurious. The MATLAB is effective for general purpose eigenvalue problem with normal matrices, but for the highly ill-conditioned matrices the effectiveness of the MATLAB might be lost.
In addition to the computational efficiency, the Krylov subspace method has several advantages including easy-implementation, the ability to detect all eigenvalues and computing all the corresponding eigenfunctions simultaneously. One can roughly locate the eigenvalues from the peaks in the response curve and then determine precise value by using the IDM. Although for the eigenvalue problem of the highly ill-conditioned Hilbert matrix, the Krylov subspace method is reliable.
6. An Iterative Detection Method for Complex Eigenvalue
Instead linearizing Equation (
3) to a linear generalized eigenvalue problem in Equation (
10), we directly treat the quadratic eigenvalue problem (
3). Now, we consider the detection of complex eigenvalue of the quadratic eigenvalue problem (
3). Because
,
and
are real matrices, the complex eigenvalue is written by Equation (
14). When we take
inserting Equations (
14) and (
39) into Equation (
3) yields
Equation (
40) becomes
which is an
dimensional homogeneous linear system.
Taking
where
is a given exciting vector and by Equation (
42), we have
The numerical procedures for determining the complex eigenvalue of the quadratic eigenvalue problem (
3) are summarized as follows. (i) Give
q,
,
a and
b. (ii) For each collocating point
solving Equation (
44) to obtain
and taking the optimal
to satisfy
where the size of
must be large enough to include at least one complex eigenvalue.
By Equation (
45), we gradually reduce the size of the rectangle centered at the previous peak point by renewing the range to a finer one. Give an initial interval
, and we fix the collocating points by
,
and pick up the maximal point denoted by
. Then, a finer rectangle is given by
, and
, and in that new rectangle we pick up the new maximum point denoted by
. Continuing this process until
, we can obtain the complex eigenvalue with high accuracy. This algorithm is shortened as an iterative detection method (IDM).
When the complex eigenvalue is computed, we can apply the techniques in Equations (
32)–(
35) with
replaced by
and
by
in Equation (
42) to compute the complex eigen-mode. The numerical procedures to detect the complex eigenvalue and to compute the complex eigenvector in Equations (
14)–(
18) for the generalized eigenvalue problem (
10) are the same to those in the above.