1. Introduction
Matrices have an important role in current engineering problems like artificial intelligence [
1], biomedicine [
2], machine learning [
3], neural networks [
4], etc. It could be said that they are, in fact, an interface between mathematics and the real world. Spectral matrix theory is closely related to big data and statistics, graph theory, networking, cryptography, search engines, optimization, convexity and, of course, the resolution of all kinds of linear equations.
An open problem of spectral theory is the sought-for eigenvalues and eigenvectors of a matrix. For this purpose, iterative methods are especially appropriate when the matrices have a large size and they are sparse, because the iteration matrix does not change during the process.
The classical power method of R. von Mises and H. Pollaczek-Geiringer [
5] assumes the existence of a dominant eigenvalue. One of the main contributions of this paper is the proposal of an algorithm to compute the eigenvalues of matrices and operators that do not have a dominant eigenvalue. The algorithm is based on the iterative averaged methods for fixed-point approximation proposed initially by Mann [
6] and Krasnoselskii [
7] and, subsequently, by Ishikawa [
8], Noor [
9], Sahu [
10] and others. There are recent articles on this topic (references [
11,
12,
13,
14,
15,
16,
17,
18,
19], for instance). In general, these papers deal with the problem of approximation of fixed points of non-expansive, asymptotically non-expansive and nearly asymptotically non-expansive mappings that are particular cases of Lipschitz and near-Lipschitz maps [
20,
21]. However, this article aims to prove that the usefulness of iterative averaged methods goes beyond fixed-point approximation.
In reference [
22], the author proposed a new algorithm of this type called N-iteration. The method has proved useful for the iterative resolution of Fredholm integral equations of the second kind and of the Hammerstein type [
23] (see, for instance, references [
24,
25]).
The above procedure is used in this article to compute eigenvalues and eigenvectors of matrices and linear operators (
Section 2 and 
Section 3) and for the resolution of systems of linear equations (
Section 4). All these techniques are applied to solving Fredholm integral equations [
26,
27,
28,
29,
30] of the first kind by means of orthogonal polynomials (
Section 5). In 
Section 6, we review some properties of compact and Fredholm integral operators, while 
Section 7 presents the solution of a Fredholm integral equation of the first kind in a particular case where the kernel of the operator is separable. 
Section 8 deals with a particular case of a Fredholm integral equation of the second kind. Several examples with tables and figures illustrate the performance and convergence of the algorithms proposed.
  2. Averaged Power Method for Finding Eigenvalues and Eigenvectors of a Matrix
Given a matrix , where  denotes the set of real or complex matrices of size , the power method is a well-known procedure for finding the dominant eigenvalue of matrix A. Let  represent the spectrum of A, that is to say, the set of eigenvalues of A.
Definition 1.  A matrix A has a dominant eigenvalue  if  for any ,  If  (or ) is non-null and  then v is a dominant eigenvector of 
 It is well known that if the matrix 
A is diagonalizable and it has a dominant eigenvalue, then the sequence defined by 
 and the recurrence 
 for 
 converges with high probability to a dominant eigenvector (i.e., if the component of 
 with respect to the dominant eigenvector is non-null). The dominant eigenvalue can be approximated by means of the so-called Rayleigh quotient:
      when 
k is sufficiently large. The formula comes, obviously, from the definition of an eigenvector associated with the eigenvalue 
 We shall see, with an example, that if the matrix has no dominant eigenvalue then the power method may not converge.
Example 1.  The matrix  has the eigenvalue  with eigenvectors  and  with eigenvectors  The power method does not work in general sincefor k being odd andfor k being even. The sequence () does not converge except for the starting point   The present section is devoted to proposing an algorithm to find eigenvalues and eigenvectors of matrices that do not have a dominant eigenvalue.
In reference [
22], an algorithm for the approximation of a fixed point of 
 where 
C, is a nonempty, closed and convex subset of a proposed normed space 
E. This is given by the following recurrence:
      for 
 and 
 This iterative scheme was called N-algorithm. The procedure generalizes the Krasnoselskii–Mann method, defined in references [
6,
7], obtained in the case where 
. If, additionally, 
 one has the usual Picard–Banach iteration.
Averaged power method
If the matrix 
 has no dominant eigenvalue, that is to say, 
 for any 
, the power method may not converge. The algorithm proposed here is the use of the N-iteration of the matrix 
A instead of computing the powers of 
A. The sequence of vectors obtained by the power or the averaged power method may produce a divergent sequence (even though the iterates may be good approximations of an eigenvector). To avoid this fact, a scaling of the output vector is sometimes performed at every step of the algorithm. Thus, the procedure is described by the following recurrence:
      for 
 such that 
 and 
  The notation 
 represents any norm in 
 and 
 is the inner product in the 
m-dimensional Euclidean space.
A useful criterion to stop the algorithm is to check that  for some tolerance  or .
The last value  is an approximation of an eigenvalue of A and  is the corresponding eigenvector.
In order to justify the suitability of the algorithm, we will consider the simplest case of the Krasnoselskii iteration:
      for 
 Let us denote 
 It is easy to check that
      and the eigenvectors of 
B and 
A agree, that is to say, 
X is an eigenvector of 
A with respect to an eigenvalue 
, if and only if 
X is an eigenvector of 
B with eigenvalue 
Let us assume that A has positive and negative eigenvalues with the same absolute value. Then,  for the positive eigenvalue 
For a negative eigenvalue  and, consequently,  is a dominant eigenvalue of B. Hence, the power method applied to the matrix B will converge to an eigenvector of B (and thus of A), if the component of  with respect to this eigenvector is non-null.
The power method is based on the product of a matrix  and a vector. This operation has a complexity  If the number of iterations is , the complexity is  For a large matrix, the averaging operations of the N-algorithm proposed are negligible, and the complexity is  Though the operations involved are about twice as many as in the original power method, the type of convergence (polynomial of degree 2) is not modified. If one wishes to reduce the total number of operations, the use of constant weight coefficients is advisable, and the algorithm is able to be conveniently optimized.
The speed of convergence of the power method (and consequently the number of iterations required to give a good approximation) depends on the ratio  where  are the greatest eigenvalues in magnitude. The smaller is this quotient, while the faster is the convergence. In the averaged Mann–Krasnoselskii method, this “spectral gap” is  where  is the positive eigenvalue. For instance, if , minimum quotients are reached for values of  close to 
Example 2.  Considering the matrix of Example 1,we obtain that for  the Krasnoselskii method with  gives as first approximationand the rest of iterations agree with the eigenvector  The positive eigenvalue is computed as   Example 3.  The spectrum of the matrixis  The N-iteration, with  starting at , with Euclidean scaling produces the sequence of approximations collected in Table 1. The eigenvalue computed is  The stopping criterion was 
 Example 4.  Let us consider the following matrix:The set of eigenvalues of A is  The algorithm described with  starting at  with Euclidean scaling produces the sequence of approximations and distances as between those collected in Table 2. The eigenvalue computed is  The stopping criterion was 
 Remark 1.  Let us notice that the matrices of these examples do not have a dominant eigenvalue but, nevertheless, the N-iterated power algorithm is convergent.
   Averaged Power Method for Finding Eigenvalues and Eigenfunctions of a Linear Operator
The algorithm proposed in this section can be applied to find eigenvalues and eigenvectors/eigenfunctions of a linear operator defined in an infinite-dimensional normed linear space 
 in case of existence (see, for instance, Theorem 2). Starting at 
, we would follow the steps
        for 
 such that 
 and 
Let us note that a structure of inner product space in E for computing the eigenvalue by means of the Rayleigh quotient is required. Otherwise, we should solve the equation 
The stop criteria of the algorithm are similar to the given for matrices, using the norm of the space E or the modulus in the case of employing the approximations of the eigenvalue.
Example 5.  Let us compute an eigenvalue of a Fredhom integral operator on  defined, for  asFor  the sequence of approximate dominant eigenvalues () generated by the N-algorithm with  is collected in Table 3. The norm  is computed as follows:and the inner product is given byTable 3 collects the number of iterations (n), the approximate eigenvalue (), the -distance between  and  (where  is the n-th approximation of the eigenfunction) and the successive distances between the eigenvalue approximations (). The tenth outcome of the eigenfunction iswith eigenvalue  The stopping criterion was  Figure 1 represents the number of iterations (x-axis) and corresponding magnitude  (y-axis).    3. Computation of an Eigenvector of a Known Eigenvalue of a Linear Operator
This section is devoted to studying the convenience of using the N-iteration to find an eigenvector/eigenfunction corresponding to a non-null known eigenvalue of a linear operator in finite or infinite dimensions.
Let 
 be a linear operator defined on a normed space 
E. If 
 where 
 is the point spectrum of 
T, the search for an eigenvector (or eigenfunction if 
E is a functional space) associated with 
 consists in finding an element 
 non-null such that 
. The solution of this equation can be reduced to a fixed-point problem considering that this equality is equivalent to
      assuming that 
 is non-null. The following result gives sufficient conditions for the convergence of the N-iteration applied to 
 to find an eigenvector associated to a non-null eigenvalue.
The next result was given in reference [
22].
Proposition 1.  Let E be a normed space and  be nonempty, closed and convex. Let  be a nonexpansive operator such that the set of fixed points is nonempty; then, the sequence of N-iterates () for  is such that the sequence () is convergent for any fixed point  of T. If E is a uniformly convex Banach space and the scalars are chosen such that  and , then the sequence () tends to zero for any .
 Theorem 1.  Let E be a uniformly convex Banach space, and  be a linear and compact operator. If  is non-null, such that , then the N-iteration with the scalars described in the previous proposition applied to the map  converges strongly to an eigenvector associated with 
 Proof.  Since  there exists ,  such that . Then,  is a fixed point of .
The latter operator is nonexpansive since
        for any 
 due to the hypotheses on 
 Regarding invariance, if 
 then
        and 
Let 
, and (
)
 be the sequence of N-iterates of 
. Since 
 is compact, there exists a convergent subsequence (
). Let 
f be the limit of this subsequence. Then,
        because 
 due to the properties of the N-iteration (Proposition 1). The continuity of 
 implies that 
 and 
 The fact that 
 exists (see Proposition 1) implies that the sequence of N-iterates (
) converges strongly to the fixed point 
f (that is to say, to an eigenvector of 
K associated with 
).    □
 Corollary 1.  Let λ be a non-null eigenvalue of the real or complex matrix A. If, for some subordinate matrix norm, , then the N-iteration applied to  is convergent to an eigenvector associated with 
 Proof.  This is a trivial consequence of Theorem 1 considering that, in finite dimensional spaces, all the linear operators are compact.    □
 Example 6.  The matrixhas an eigenvalue equal to 10 and  The computation of an eigenvector associated with  by means of the N-iteration with  starting at  as described in Corollary 1, has produced the approximations collected in Table 4. The seventh approximation of the estimated eigenvalue isThe stop criterion was     4. Solution of a Linear System of Equations by an Averaged Iterative Method
If we need to solve a linear system of equations such as
      where 
, 
 and 
 we can transform (
1) into an equation of the following type:
      where 
 and 
 It is clear that 
T is a Lipschitz operator:
, and 
 are a subordinate matrix norm of 
 Let us note that we use the same notation for vector and subordinate matrix norms.
If 
 then 
T is nonexpansive. Let us assume that the system (
1) has some solution 
. Then, for any 
      and any ball 
 is invariant under 
T. If 
,
      then the restriction of 
T to the ball 
 is a continuous map on a compact set. The N-iteration is such that 
 (Proposition 1). At the same time, due to the compactness of 
, for 
 there exists a subsequence 
 such that 
 By continuity, 
 tends to 
 and consequently 
 is a fixed point of 
T. The fact that 
 is convergent (Proposition 1) implies that the whole sequence 
 converges to the fixed point 
 that is to say, it approximates a solution of the equation 
Remark 2.  If , T is contractive on a complete space, and there exists a unique solution which can be approximated by means of the Picard iteration.
If  T is nonexpansive and the N-iteration where  and  can be used to find a solution, according to the arguments exposed.
 Remark 3.  The complexity of this iterative algorithm is similar to that given for the averaged power method, since the core of the procedure is the product of a matrix (which remains constant over the process) and a vector.
 Example 7.  Let us consider the equation , whereand  The spectral radius of  is  and the typical fixed-point method does not converge. The N-iteration has been applied to the map  with the values of the parameters equal to  and starting point  Table 5 collects ten approximations of the solution along with their relative errors, computed as  The stop criterion was  Figure 2 represents the number of iteration (x-axis) along with  (y-axis)    5. Solution of a Fredholm Integral Equation of the First Kind by Means of Orthogonal Polynomials
This section is devoted to solving integral equations of the Fredholm type of the first kind, given by the following expression:
      assuming that 
, 
 and the function 
u is known. The notation 
 means the space of square-integrable real functions defined on a compact real interval 
We must notice that the solution of this equation is not unique, since if 
 is a solution of (
3) and 
h belongs to the null space of the operator 
K defined as follows:
      then 
 is a solution as well.
The map  is called the kernel of the Fredholm operator K. The following are interesting particular cases of k:
If   is the Fourier Transform of f.
For  we have the Hilbert Transform of f.
When  the Laplace Transform of f is obtained.
We consider in this section an arbitrary kernel .
Assuming that 
, we consider the expansions of 
u and 
v in terms of an orthogonal basis of the space 
 in order to find a solution of (
3). In this case we have chosen the polynomials of Legendre, which compose an orthogonal basis of 
. They are defined for 
 as
	  Legendre polynomials satisfy the following equalities:
	  We have considered here orthonormal Legendre polynomials 
, that is to say,
      in order to facilitate the calculus of the coefficients of a function with respect to the basis. In this way, for 
      where
	  The convergence of the sum holds with respect to the norm 
 in 
 If the domain of the integral of the Equation (
3) is a different interval, a simple change of variable transforms 
 into a system of orthogonal polynomials defined on it.
To solve the Equation (
3), let us consider Legendre expansions of the maps 
 up to the 
q-th term:
	  The discretization of the integral Equation (
3) becomes
	  Thus,
	  Denoting 
 we obtain a system of linear equations whose unknowns are the coefficients 
 of the sought function 
      for 
 The linear system can be written as
      where 
 and 
 The system can be transformed into the fixed-point problem
      as explained in the previous section. This equation can be solved using the described method whenever 
 for some subordinate norm of the matrix 
Example 8.  In order to solve the following Fredholm integral equation of the first kindwe used the Legendre polynomials defined in the interval :We have considered an expansion with  and obtained a system of  linear equations with matrix ,The unknowns  are the coefficients of v with respect to () and  is the vector of coefficients of u with respect to the same basis. The system  has been transformed into  Since  (close to 1), we have applied the N-algorithm with all the parameters equal to  Starting at , we obtain at the 18-th iteration the approximate solution:with an error ofand stop criterion  Figure 3 displays the 6th, 12th and 18th approximation function to the solution of the equation. For the 18th Picard iteration, , the error is as follows:Consequently, the N-iteration may be better even if the Picard method can be applied.    6. Spectrum of a Compact Operator and General Solution of Fredholm Integral Equations of the First Kind
In this section we explore the relationships between spectral theory and the resolution of a Fredholm integral equation of the first kind given by following the expression:
      where 
 and 
 is defined as
      where 
 It is well known that 
K is a linear and compact operator.
A linear map  where E is a normed space, is self-adjoint if  where  is the adjoint of K. The set of eigenvalues of a linear operator is called the point spectrum of K, , and  if there exists  such that  In this case v is an eigenvector of K (or an eigenfunction, in the case where E is a functional space). The set  is the eigenspace associated with 
The next theorems describe the spectrum of compact operators (see the references [
26,
27], for instance), and provide a kind of “diagonalization” of these maps.
Theorem 2.  Let  be a linear, compact and self-adjoint operator. Then K has an orthonormal basis of eigenvectors and
If the cardinal of  is not finite then the spectrum is countably infinite and the only cluster point of  is  that is to say,  where 
The eigenspaces  corresponding to different eigenvalues are mutually orthogonal.
For ,  has finite dimension.
 According to this theorem, the eigenvalues of a compact self-adjoint operator can be ordered as follows:
      and the action of 
K on every element 
 can be expressed as follows:
      where 
 is an orthonormal basis of eigenvectors of 
K. If 
 is finite, then 
K has a finite range. If additionally 
K is positive, then the eigenvalues also are. If 
K is strictly positive, then the inverse operator admits the following expression:
      and, in general, 
K is a densely defined unbounded operator, because the range of 
K is dense in 
E.
For a compact operator, the map  is positive and self-adjoint and the positive square roots of the eigenvalues of  are the singular values of  For a not necessarily self-adjoint operator, we have the following result, known as Singular Value Decomposition of a compact operator K.
Theorem 3.  Let  be a non-null linear and compact operator. Let  be the sequence of singular values of K, considered according to their algebraic multiplicities and in decreasing order. Then there exist orthonormal sequences  and  on E and , respectively, such that, for andFor every where , and  denotes the null space of K. Moreover  Remark 4.  According to the equalities (7) and (8), for ,andthat is to say,  is an eigenvector of  with respect to the eigenvalue  and  is an eigenvector of  with respect to the same eigenvalue.  Let us consider the equation quoted above
      for 
 where 
 is the data function and 
 is the unknown. We will assume that the kernel 
 of 
K satisfies the following:
      so that 
K is linear and compact.
Let us assume that 
g can be expanded in terms of the following system 
:
	  According to (
8) and the definition of adjoint operator,
      and
	  Let us assume also that
	  Then, using the equality (
7),
      and by (
12),
	  If 
f is expanded in terms of the system 
,
      and, by (
11), there exists a solution of 
 if
	  We can conclude that if 
g admits an expansion in terms of the system 
 the Fredholm integral equation of the first kind has a solution.
  7. Solution of a Fredholm Integral Equation of the First Kind with Separable Kernel
We consider now the case where 
K has a separable kernel 
k, that is to say, 
k can be expressed as
      for some 
Spectrum of K:
As suggested by the equality (
13), we can guess that 
K has eigenfunctions of type
	  In order to find 
, we consider that
      where
      and the problem is reduced to solving the eigenvalue issue:
      where 
 is the vector of coefficients of 
 with respect to (
).
Null space of K:
Let us consider now the null space of the operator 
K. If 
,
	  If the maps 
 are linearly independent, a characterization of the functions belonging to the null space is as follows:
      for 
 Let us consider an orthogonal system of functions 
 for 
 defined following a Gram–Schmidt process from 
 If 
 is any function,
      belongs to the null space of 
K since
      for 
. In order to define an orthonomal basis of the kernel, let us consider a system of functions 
 such that 
 is linearly independent and complete. Then, we define
	  A Gram–Schmidt process on 
 provides an orthonormal basis (
) of the null space of 
K.
General solution of :
The general solution of the equation 
 has the following form:
      where 
 are the eigenfunctions, the coefficients 
 are to be computed and the 
 are arbitrary. To find 
 let us consider that
      and we have a linear system of 
r equations with unknown 
’s, 
  Example
Consider the following Fredholm integral equation of the first kind:
		Let us study the characteristics of the operator 
K.
Spectrum of K:
The first eigenvalue of the operator was computed in Example 5, obtaining a value of 
 The second eigenvalue was found by applying the averaged power method, but subtracting at every step the component of the approximation with respect to the first eigenfunction (
. The value obtained is 
 The maps of the kernel 
k are as follows:
Let us define the matrix 
A as in (
14), 
		The spectrum of the matrix is 
 Two eigenfunctions associated with these eigenvalues were computed as follows: 
 and 
 Since the map 
, the equation has a solution.
Null space of K:
As explained previously, the maps 
h belonging to the null space of 
K are characterized by the following equations:
        for 
 An orthogonal system defined from 
 is as follows:
		In order to construct an orthonormal basis of the null space, we consider the following functions: 
 First, define
		Then, by means of a Gram–Schmidt process, we obtain an orthogonal basis of the null space:
General solution of :
Let the solution of the equation be
        where 
 are the eigenfunctions of 
K and 
 Solving the system (
15), we obtain the coefficients 
 and 
 Substituting in the previous equality we find the approximate general solution of the equation:
        where 
 are arbitrary coefficients. An exact solution is 
  8. Solving the Equation 
Let us consider now the following equation:
      where the kernel is separable: 
 It is obvious that this equation has a non trivial solution if 
 belongs to the point spectrum of 
 In this case the solution is any eigenfunction 
 associated with 
 For instance, let us solve the following equation:
      where 
 is a separable kernel with 
    In this case the eigenvalue is 
 The matrix associated with the operator, 
 is as follows:
We look for an eigenfunction of the form: 
 The system of equations to find the coefficients is as follows:
	  The spectral radius of 
A is 
 Consequently, the usual iterative methods to solve this system do not converge. We apply the N-iteration to the matrix 
A starting at 
 and we obtain a sequence of vectors for the approximate solutions of 
 in terms of the maps 
 The results are collected in 
Table 6. The last outcome is 
  providing the approximate solution:
      with stop criterion 
 where 
 is the vector of coefficients computed at the 
n-th iteration. An exact solution is 
Remark 5.  All the computations have been performed using Mathematica software.