Abstract
Solving equations in abstract spaces is important since many problems from diverse disciplines require it. The solutions of these equations cannot be obtained in a form closed. That difficulty forces us to develop ever improving iterative methods. In this paper we improve the applicability of such methods. Our technique is very general and can be used to expand the applicability of other methods. We use two methods of linear interpolation namely the Secant as well as the Kurchatov method. The investigation of Kurchatov’s method is done under rather strict conditions. In this work, using the majorant principle of Kantorovich and our new idea of the restricted convergence domain, we present an improved semilocal convergence of these methods. We determine the quadratical order of convergence of the Kurchatov method and order for the Secant method. We find improved a priori and a posteriori estimations of the method’s error.
Keywords:
nonlinear equation; iterative process; convergence order; secant method; Kurchatov method; Banach space; divided difference; local; semi-local convergence MSC:
49M15; 47J25; 65G99; 65J15; 65H10
1. Introduction
We consider solving equation
using iterative methods. Here , are Banach spaces, is an open region of .
Secant method
is a popular device for solving nonlinear equations. It is due to the following: simplicity of the method; small amount of calculations on each iteration and use of the value of an operator from only two previous iterations in the iterative formula of the method. A lot of works are dedicated to this method [,,]. In [] the Secant method is used for solving the nonlinear least squares problem. The Kurchatov’s method of linear interpolation
is less known. This method has the same order of convergence as Newton’s method but does not require the calculation of derivatives. In (2) and (3), is a divided difference of the first order for the operator F at the points u and v [,].
In this work we will investigate the Secant method and Kurchatov’s method using the Kantorovich’s principle of majorants. For the first time, this principle was used by L.V. Kantorovich for investigating the convergence of the classical and modified Newton’s method, having built for the nonlinear operator a majorizing real quadratic function []. Corresponding to this, the iterative sequence for nonlinear operator is majorized by a converging sequence for nonlinear equation with one variable. Later the nonlinear majorants for investigating other methods of solving nonlinear functional equations have been built. In work [] with the help of the majorant principle, a method with the order of convergence , which in its iterative formula uses the value of an operator from the three previous iterations, is investigated. Specifically, a real cubical polynomial, which majorizes the given nonlinear operator is built. With that, the Lipschitz conditions are put upon the divided differences’ operator of the second order [,]. We investigate the Secant method with different conditions that have been put upon the nonlinear operator. In particular, if the Lipschitz condition for the divided differences of the first order are fulfilled, the quadratic majorizing function of one variable is built, and if the Lipschitz condition for operator of divided difference of the second order are fulfilled, the cubical majorizing function is built. The cubical majorizing function for Kurchatov’s method is also built. Methods of linear interpolation applied to these functions produce a numerical sequence, which majorizes by norm the iterative sequence, produced by applying these methods to the nonlinear operator. In all cases, the a priori and a posteriori error estimations of the linear interpolation methods are also provided.
2. Divided Differences and Their Properties
Let us assume that and z are three points in region .
Definition 1
([]). Let F be a nonlinear operator defined on a subset Ω of a Banach space with values in a Banach space and let be two points of Ω. A linear operator from to which is denoted by and satisfies the conditions:
(1) for all fixed two points
(2) if exist a Fréchet derivative , then
is called a divided difference of F at the points x and y.
Note that (4) and (5) do not uniquely determine the divided difference with the exception of the case when is one-dimensional. For specific spaces, the differences are defined in Section 6.
Definition 2
([]). The operator is called divided difference of the second order of function F at the points x, y and z, if
We assume that for and the conditions of the Lipschitz type are being satisfied in the following form:
If the divided difference of F satisfies (7) or (8), then F is differentiable by Fréchet on . Moreover, if (7) and (8) are fulfilled, then the Fréchet derivative is continuous by Lipschitz on with the Lipschitz constant [].
Let us denote and . The semilocal convergence of the Secant method uses the conditions :
is nonlinear operator with denoting a first order divided difference on .
Let . Suppose that the linear operator is invertible and let be nonnegative numbers such that
Assume that the following conditions hold on
or
Moreover, assume the following Lipschitz conditions hold for all for some and
and
Set . Define
provided
The following Lipschitz conditions hold on for some and
and
Set
Suppose , and define and
Moreover, suppose
, where is the unique root in of equation .
Remark 1.
The following Lipschitz condition is used in the literature for the study of iterative methods using divided differences [,,,,,,,,,,,,] for
although it is not really needed, since tighter conditions are really needed (see conditions and proofs that follow).
By these definitions we have
The sufficient semilocal convergence criterion in the literature arrived at different ways and corresponding to [] is
where
provided that (stronger than ).
Then, we have
but not necessarily vice versa, and for each .
Hence, the applicability of the Secant method is extended and under no additional conditions, since all new Lipschitz conditions are specializations of the old condition. Then, in practice the computation of requires that of of the other as special cases. Some more advantages are reported after Proposition 1. It is also worth noticing that and help define through which and p are defined too. With the old approach p depends only on , which contains . In our approach the iterates remain in (not used in []). That is why our new p constants are at least as tight as . There is where the novelty of our paper lies and the new idea helps us extend the applicability of these methods. It is also worth noticing that the new constants are specializations of the old ones. Hence, no additional conditions are added to obtain these extensions.
It is worth noting from the proof of Theorem 1 that can be defined as or .
Theorem 1.
Suppose that the conditions hold. Then, the iterative procedure (2) is well defined and the sequence generated by it converges to a root of the equation . Moreover, the following error estimate holds:
where
The semilocal convergence of the discussed methods was based on the verification of the criterion (11). If this criterion is not satisfied there is no guarantee that the methods converge. We have now replaced (11) by (10) which is weaker (see (12)).
Proof.
Notice that the sequence is generated by applying the iterative method (2) to a real polynomial
It is easy to see that the sequence monotonically converges to zero. In addition, we have
We prove by using of induction that the iterative method is well defined and that
Using , (13), (14) and
it follows that (17) holds for . Let k be a nonnegative integer and for all the fulfills (17). If , then by , we have
In view of the Banach lemma [] is invertible, and
Next, we prove that the iterative method exist for n = k + 1. We get
By condition , we have
Hence, the iterative method is well defined for each n. Hence, it follows that
Corollary 1.
The convergence order of iterative Secant method (2) is equal to .
Proof.
Concerning the uniqueness of the solution, we have the result.
Proposition 1.
Under the conditions further suppose that for
holds for all , where , and provided , where is a solution of equation . Then, is the only solution of equation in the set .
Proof.
Let , where and . Then, we get
so follows from □
Remark 2.
The result in Proposition 1 improves the corresponding one in the literature using the old condition, since . Hence, we present a larger ball inside which we guarantee the uniqueness of the solution .
If, additionally, the second divided difference of function F exists and satisfies the Lipschitz condition with constant q, then the majorizing function for is a cubical polynomial. Then, the following theorem holds.
Theorem 2.
Under the conditions (except ) further suppose
Let us presume that and denote
Let h be a real polynomial
It the following inequality is satisfied
and the closed ball , where is the root of equation .
Conditions of Propositions 1 hold on .
Then, the iterative method (2) is well defined and the generated by it sequence converges to the solution of the equation . Moreover, the following estimate is satisfied
where
This proof is analogous to the proof of Theorem 1.
Remark 3.
The majorizing sequences are more precise than the one in [] (using instead of , respectively).
Similar advantages we report in the case of Theorem 2, see e.g., [] where instead of the following condition is used on Ω
We have
(since ).
3. A Posteriori Estimation of Error of the Secant Method
If the constants are known, then we can compute the sequence before generating the sequence by the iterative Secant algorithm. With help of inequalities (13) and (27), the a priori estimation of error of the Secant method is given. We obtain an a posteriori estimation of the method’s error, which is sharper than the a priori one.
Theorem 3.
Let the conditions of the Theorem 2 hold. Denote
Then, the estimate holds for
Proof.
By condition , we have
It is easy to see that . Then, according to the Banach lemma is invertible, and
From (4) we can write
If the second divided difference of function F exists and satisfies the Lipschitz condition with constant q, then the following theorem holds.
Theorem 4.
Let the conditions of the Theorem 2 hold. Denote
Then the following estimate holds for
Proof.
The proof of this theorem is similar to the previous theorem, but instead of inequalities (32), the following majorizing inequalities are used
□
4. Semilocal Convergence of the Kurchatov’s Method
Sufficient conditions of semilocal convergence and the speed of convergence of the Kurchatov’s method (3) are determined by the following theorem.
Theorem 5.
Suppose: Conditions , and hold, but with and , where , provided and
Let us assume that and denote
Let be a real polynomial
Proof.
The proof of the theorem is realized with help of the majorants of Kantorovich. As in Theorem 1 but we also use the crucial estimate
□
Corollary 2.
The convergence order of iterative Kurchatov’s procedure (3) is quadratic.
Proof.
As a result that according to (34) the convergence of the sequence to zero not higher than quadratic, then there are and , that for all the inequality holds
Thus, Kurchatov’s method has a quadratic convergence order as Newton’s method but does not require the calculation of derivatives.
Remark 4.
We obtain similar advantages as the ones reported earlier for Theorem 2.
5. A Posteriori Estimation of Error of the Kurchatov’s Method
If the constants , are known, then we can compute the sequence before receiving the sequence by the iterative algorithm (3). With help of inequality (33) the a priori estimation of error of the Kurchatov’s method is given. We will receive a posteriori estimation of the method’s error, which is coarser than the a priori one.
Theorem 6.
Let the conditions of the Theorem 5 be fulfilled. Denote
Then for the following estimation holds
Proof.
It is easy to see that . Then by Banach lemma has inverse and
From (4) we can write
So, it follows
and
□
Proposition 2.
Under the conditions of Theorem 6 further suppose that for
holds for all , where , provided .
Proof.
This time, we have
The rest follows as in Proposition 1. □
Remark 5.
The results reported here can immediately be extended further, if we work in instead of the set , where . The new p constants will be at least as tight as the ones presented previously in our paper, since .
6. Numerical Experiments
In this Section, we verify the conditions of the theorems on convergence of the considered methods for some nonlinear operators, and also compare the old and new radii of the convergence domains and error estimates. We first consider the representation of the first-order divided differences for specific nonlinear operators [,].
Let . We have a nonlinear system of m algebraic and transcendental equations with m variables
In this case is the matrix with entries
If , then .
Let us consider a nonlinear integral equation
where is a continuous function of its arguments and continuously differentiable by x. In this case is defined by formula
If holds for some , then .
Example 1.
Let , and . The solution of equation is .
In view of F, we can write , , .
Let us choose and . Then we get , for Secant method and , for Kurchatov’s method, , , , . For the corresponding theorems in [] , .
In Table 1, there are radii and convergence domains of considered methods. They are solutions of corresponding equations and satisfy the condition . We see that hold. Moreover, for Kurchatov’s method and . So, the assumptions of the theorems are fulfilled. Next, we show that error estimates hold, i.e., , and compare them with corresponding ones in []. Table 2 and Table 3 give results for Secant method (2), and Table 4 for Kurchatov’s method (3).
Table 1.
Radii and convergence domains.
Table 2.
New and old error estimates (13).
Table 3.
New and old error estimates (27).
Table 4.
New and old error estimates (33).
Table 2, Table 3 and Table 4 show the superiority of our results over the earlier ones, i.e., obtained error estimates are tighter in all cases. That means fewer iterates than before are needed to reach a predetermined error tolerance.
Example 2.
Let , and
The solution of equation is .
For x and , we have
For this problem we verify conditions (C) and corresponding ones from []. Let us choose and . Having made calculations, we get , , , , and . Then . Next , and . The equation has two solutions and . Only . Therefore, , and .
Analogy, an equation has two solutions and . . Therefore, , and .
So, estimates (13) are tighter than the corresponding ones in [].
Secant and Kurchatov’s methods solve this system under 5 iterations for and the specified initial approximations.
Example 3.
Let and
The solution of this equation is . In view of F, we can write
Let us choose and . Both methods give approximate solution of the integral equation under 13 iterations for . To solve a linear integral equation at each iteration was applied Nystrom method. We use a trapezoidal quadrature formula with 101 nodes. On the graphs denotes and denotes (see Figure 1). We can see that , where . This corresponds to the error estimation of the trapezoidal quadrature formula.
Figure 1.
Values of (a) and (b)
at each iteration.
7. Conclusions
The investigations conducted showed the effectiveness of applying the Kantorovich majorant principle for determining the convergence and the order of convergence of iterative difference methods. The convergence of the Secant method (2) with the order and the quadratic convergence order of the Kurchatov’s method (3) are established. According to this technique, nonlinear majorants for a nonlinear operator are constructed, taking into account the conditions imposed on it. By using our idea of restricted convergence regions, we find tighter Lipschitz constants leading to a finer local convergence analysis of these methods than in []. Our new technique can be used to extend the applicability of other methods along the same lines. More details on the extensions were given in Remark 1.
Author Contributions
Editing, I.K.A.; conceptualization S.S.; investigation I.K.A., S.S. and H.Y. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Shakhno, S.M. Nonlinear majorants for investigation of methods of linear interpolation for the solution of nonlinear equations. In Proceedings of the ECCOMAS 2004—European Congress on Computational Methods in Applied Sciences and Engineering, Jyväskylä, Finland, 24–28 July 2004; Available online: http://www.mit.jyu.fi/eccomas2004/proceedings/pdf/424.pdf (accessed on 26 May 2020).
- Amat, S. On the local convergence of secant-type methods. Int. J. Comput. Math. 2004, 81, 1153–1161. [Google Scholar] [CrossRef]
- Hernandez, M.A.; Rubio, M.J. The Secant method for nondifferentiable operators. Appl. Math. Lett. 2002, 15, 395–399. [Google Scholar] [CrossRef]
- Shakhno, S.; Gnatyshyn, O. Iterative-Difference Methods for Solving Nonlinear Least-Squares Problem. In Progress in Industrial Mathematics at ECMI 98; Arkeryd, L., Bergh, J., Brenner, P., Pettersson, R., Eds.; Verlag B. G. Teubner GMBH: Stuttgart, Germany, 1999; pp. 287–294. [Google Scholar]
- Ul’m, S. Algorithms of the generalized Steffensen method. Izv. Akad. Nauk ESSR Ser. Fiz.-Mat. 1965, 14, 433–443. (In Russian) [Google Scholar]
- Ul’m, S. On generalized divided differences I, II. Izv. Akad. Nauk ESSR Ser. Fiz.-Mat. 1967, 16, 13–26. (In Russian) [Google Scholar]
- Kantorovich, L.V.; Akilov, G.P. Functional Analysis; Pergamon Press: Oxford, UK, 1982. [Google Scholar]
- Potra, F.A. On an iterative algorithm of order 1.839... for solving nonlinear operator equations. Numer. Funct. Anal. Optim. 1985, 7, 75–106. [Google Scholar] [CrossRef]
- Ul’m, S. Iteration methods with divided differences of the second order. Doklady Akademii Nauk SSSR 1964, 158, 55–58. [Google Scholar]
- Schwetlick, H. Numerische Lösung Nichtlinearer Gleichungen; VEB Deutscher Verlag der Wissenschaften: Berlin, Germany, 1979. [Google Scholar]
- Argyros, I.K. A Kantorovich-type analysis for a fast iterative method for solving nonlinear equations. J. Math. Anal. Appl. 2007, 332, 97–108. [Google Scholar] [CrossRef]
- Argyros, I.K.; George, S. On a two-step Kurchatov-type method in Banach space. Mediterr. J. Math. 2019, 16. [Google Scholar] [CrossRef]
- Argyros, I.K.; Magreñán, A.A. Iterative Methods and Their Dynamics with Applications: A Contemporary Study; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
- Kurchatov, V.A. On a method of linear interpolation for the solution of functional equations. Dokl. Akad. Nauk SSSR 1971, 198, 524–526. [Google Scholar]
- Shakhno, S.M. Kurchatov method of linear interpolation under generalized Lipschitz conditions for divided differences of first and second order. Visnyk Lviv. Univ. Ser. Mech. Math. 2012, 77, 235–242. (In Ukrainian) [Google Scholar]
- Shakhno, S.M. On the difference method with quadratic convergence for solving nonlinear operator equations. Matematychni Studii 2006, 26, 105–110. (In Ukrainian) [Google Scholar]
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).