1. Introduction
Let
X and
Y stand for Banach spaces and let
be a convex and nonempty subset of
X. A plethora of applications from diverse disciplines may be solved, if reduced to a nonlinear equation of the form
This reduction takes place by using mathematical modeling [
1,
2,
3,
4,
5,
6]. Then, a solution denoted by
is to be found that answers the application. The solution may be a number or a vector or a matrix or in general an operator. This task is very challenging in general. Obviously, the solution
is desired in an analytical form. However, in practice, this is achievable only in rare cases. That is why researchers mostly develop iterative methods convergent to
under some conditions on the initial information.
Popular methods are the modified Newton’s method (MNM) and Newton’s method (NM) defined, respectively, for starting point
and all
by
and
Here,
is the notation for the Fréchet derivative of the operator
F [
7].
Numerous articles have been written on the convergence of these two methods [
7,
8,
9,
10,
11]. The convergence conditions are mostly sufficient and in rare cases necessary. This observation indicates that there is a possibility to weaken the conditions, especially because these methods may converge even if these conditions are not fulfilled.
That is why in this article the objective is to consider alternatives.
Let us look at the main convergence conditions for these methods.
Remark 1. Suppose that there exist parameters and , such thatandfor all . Moreover, consider a parameter , such thatThen, the corresponding sufficient convergence conditions for MNM and NM are, respectively [9,12,13]and The conditions (
7) and (
8) are due to Kantorovich [
11]. Clearly, it follows that
Thus, we deduce that
That is, the condition (
7) is weaker than (
8). However, the convergence of MNM is only linear, whereas that of NM is quadratic [
11]. Moreover, one can construct even scalar equations, where both conditions (
7) and (
8) are not fulfilled. That is, convergence is not assured by either convergence result although these methods may converge. Let us look at an elementary but motivational example.
Example 1. Let us consider the domain for and the starting point Moreover, we define the function byThe conditions (4)–(6) are fulfilled if and . By plugging these values on the conditions (7) and (8) and solving for the parameter λ, we deduce that (8) does not hold for any , whereas (7) does hold provided that . However, the convergence is only linear in the MNM case. Remark 2. In view of the Example 1, the question arises: Can we weaken the condition (8) but maintain the quadratic convergence of NM? The answer was given in [1,9,12,13] and it is positive. In those articles we looked at condition (8) and realized that a weakening can take place if in condition (8) we replace: Case 1. Parameter by a smaller one;
Case 2. Parameter η by a smaller one; and
Case 3. Parameters and η by smaller ones.
Positive results for Case 1 are reported in [1,11]. Additional benefits include more precise error estimates on the distances , and an at least as extended uniqueness ball. The novelty is that all these benefits are achieved without additional conditions. Relevant research can be found in [14,15,16,17]. In this article, we present similar contributions for cases 2 and 3.
The idea is to replace NM with a Newton-type method (NTM) for
,
and all
defined by
where
are continuous operators and
is a bounded sequence of nonzero parameters. The notation
stands for
and
with
.
Next, we present a general auxiliary result for the convergence of iterative methods.
Lemma 1. Let and be normed spaces and be a continuous operator, where the set D is open. Define the NTMwhere is a sequence of linear operators. Suppose
(i) The sequence exists and is bounded
and
(ii) the sequence is Cauchy.
Then, there exists a parameter such that for all , Proof. The sequence
exists and is bounded by the hypothesis. Thus, (
14) holds. Then, it follows by method (
13) for
that
since the sequence
is Cauchy.
□
Remark 3. - (a)
The space does not have to be complete for the sequence to converge. In the case of method (12), set . Then, solves the equation . - (b)
Possible choices for H are or .
- (c)
Special cases of NTM are MNM (if and for all ), NM (if and for all ).
Other choices lead to Newton-like methods [
1]. That is why by studying the convergence of method NTM, we also unify the convergence of its specializations. Moreover, we may weaken the convergence criteria and improve the error bounds or the information on the uniqueness of the solution
at least in some cases (see
Section 4). Notice that the smallness of
is determined by
for some
. Then, in some cases, this parameter is such that
Hence, this case is favorable to our expectations.
The semilocal convergence for method (
12) is given in
Section 2, followed by the local convergence in
Section 3. The examples and the concluding remarks appear in
Section 4 and
Section 5, respectively.
2. The Semi-Local Convergence of the Method NTM
We introduce certain parameters and real sequences that are important in the convergence of the method (
12). Let
for
be given parameters. Define the parameters
Some of these parameters can be zero (see also the Example 2). Moreover, define the sequences for all
This sequence appears often in the study of Newton-like methods [
1].
The first general convergence result for the sequence follows.
Lemma 2. Suppose that for and all Then, the sequence generated by the Formula (16) is convergent and satisfying . Proof. It follows by (
16) and (
17) that the sequence
is nondecreasing and also bounded from above by
and as such is convergent to
. □
Remark 4. We can provide some stronger alternatives to the verification of (17). It is convenient to introduce some polynomials and functions defined on the interval in order to show the second convergence result for the sequence as follows:Thus, we get by these definitions thatDefine the function on the interval byThen, we getSetThen, the function can be rewritten asBy these definitions and . It then follows by the intermediate theorem that the polynomial has zeros in the interval . Notice that by Descartes’ rule of signs there is only one such zero, which we denote by γ. Suppose
where
is the smallest positive solution
assured to exist by
;
or
Notice that the conditions
is verified in Example 2 used in Theorem 1, which follows.
Lemma 3. Suppose that any of the conditions – hold. Then, the sequence generated by the formula (16) is convergent with and Proof. Mathematical induction is utilized to show for all
the assertion
This assertion holds if
, by (
16) and the second condition in
–
. It follows that
. Suppose that the assertion (
19) holds for all integer values smaller than
k. Then, we get
and
Then, evidently (
19) holds if
Assertion (
20) motivates the introduction of the recurrent polynomials
, and shows instead of (
20) that
However,
, because
. Thus, we have
.
Consequently, the estimate
can be shown instead of (
21), which is true by any of the last conditions in
–
.
The induction for the assertion is completed. Therefore, the sequence is nondecreasing and bounded from above by . Hence, it is convergent to some . □
The notation is used for the open and closed ball in X of center x and radius , respectively. We denote by the space of bounded linear operators from Y to X.
The following set of conditions are used in the semi-local convergence of NTM.
There exist
and
such that
and
Define the region
as
The conditions of any of the last two lemmas are fulfilled
and
Next, the semilocal convergence of the method NTM follows.
Theorem 1. Suppose that the conditions – and any of the conditions , or (17) hold. Then, the sequence generated by the NTM is well defined in , remains in and is convergent to some solving the equation . Moreover, the following assertion holds: Proof. Mathematical induction is used to show
for all
The definition of (
16) and the condition
imply
so the iterate
and the assertion (
23) hold for
. Let
. It follows from the conditions
–
that
Then, the estimate (
24) and the perturbation lemma by Banach on linear invertible operators [
11,
18,
19] assert that
and
Moreover, the iterate
is well defined.
The definition of NTM implies the identity
By applying the condition
on the identity (
26), we obtain in turn the estimates
and
By summing up the preceding estimates, we get
and thus
and
Hence, the iterate
and the estimate (
23) holds for all
k. Moreover, the sequence
is complete as convergent. Thus, the sequence
is also complete in Banach space
X. Hence, it is also convergent to some
. Furthermore, the continuity of
F, the Remark 3(a), and by letting
in the estimate (
27) we obtain
. □
Next, a result is presented concerning the uniqueness of the solution for the equation .
Proposition 1. Suppose
there exists a solution of equation for some ;
the condition holds on the ball ;
and
there exists , such thatDefine the region . Then, the equation is uniquely solvable by the element z in the region . Proof. Let
be a solution of the equation
. Then, we have
. Define the linear operator
. By using the conditions
and (
28), we obtain in turn that
where we also used
and
It follows by (
29) that
Consequently, we can have
That is, we conclude that
. □
Remark 5. The assumption can be dropped if the second condition in is replaced by any of or , and , where . Then, the proof of Theorem 1 still goes through. The limit point can be replaced by or given in closed form in the condition .
Notice that only the condition is used in Proposition 1. However, if all conditions are used then, we can set .
An alternative to the majorizing sequence (16) and the convergence condition can be obtained as follows. Let , , and .
Moreover, define the sequence byThen, ifit was shown in [8] that the sequence is nondecreasing and convergent toThe parameter is the smallest of the two roots of the quadratic polynomial with the largest being given byMoreover, simple induction shows, and Hence, the sequence and the conditions (31) can replace and conditions in Theorem 1, respectively. Concerning the uniqueness of the solution for this case, we already have Proposition 1. However, the uniqueness of the solution (see [8]) can be established in the regionIn practice, we shall choose the largest region, the tighter sequence and provided that both the and the conditions (31) hold. The sequence of number can be replaced by a sequence of continuous operators from Ω into X. In this case, the proof of Proposition 1 also goes through, provided thatThat is the operators and must be commutative. A more general method than (12) is given by the Picard iterationwhere are continuous operators and operator T has the same fixed points with P. Suppose that P satisfiesor not and or not. However,andand . Then, according to the contraction mapping principle [2], the operator P has a fixed point provided also that it maps a closed ball into itself. A possible choice for P and T can be 3. Local Convergence
Let
be given parameters with
and
. Moreover, define the parameter
r by
provided that
. These parameters are connected with the operators appearing on the NTM with the conditions
.
Suppose
there exists a solution of the equation , such that and ;
for all .
Define the region
and
for each
and
, where the parameter
r is given by the formula (
33).
Next, the local convergence analysis uses the parameters “l” and the conditions .
Theorem 2. Suppose that the conditions – hold and the starting point . Then, the sequence generated by NTM is such that , andwhere Proof. It follows by the conditions
,
, the hypothesis
and the radius
r that
Thus, the operator
and
Moreover, the iterate
is well defined by NTM for
, and we can write in turn that
By composing the expression in the bracket by
and using the conditions
, we see that it is bounded above by
leading together with (
36) and (
37) to the estimate (
34) for
, where we also used that
.
Hence, the iterate .
Then, the induction for the estimate (
34) is completed if we simply replace the iterates
by
in the preceding calculations.
Therefore, we have
where
implies that the iterate
and
. □
Remark 6. The last condition in can be dropped in view of the alternative estimateBy replacing this estimate in the proof of Theorem 2, we see that the radius becomeswhere we suppose . Moreover, the new sequence is defined byEven at this generality, the Theorem 2 improves earlier results in the interesting case of NM. Indeed, we have in this case , , , , implying thatThe corresponding radius given independently by Traub [4] and Rheinboldt [2] iswhere satisfiesHowever, then the estimateholds, because and . Let us look at the function F defined by for all . Then, we have for , that , , and . Hence, we haveMoreover,Furthermore, the new sequence is tighter than given in [2,4] and defined Finally, notice that the second condition in can be replaced by . Then, it follows again that .
Proposition 2. Suppose there exists a solution of the equation with The condition holds in the set , and there exists such thatDefine the region . Then, the only solution of the equation in the region is . Proof. Define the linear operator
S by
for some
with
. It then follows by
and (
39) that
Then, by the continuity of
F, the invertibility of
S and the identity
we conclude that
. □
Notice that if all hypotheses of Theorem 2 hold, then we can choose .
4. Special Cases and Numerical Problems
The operators appearing on the method (
12) are specialized in some interesting cases. Then, a favorable comparison is given with existing methods.
Example 2. Let us consider the case of NM. We shall verify the parameters in Theorem 1. It follows from (3) and (12) that the conditions () – () are verified provided that , , and to be determined if the operator is specified. The parameters delta are: , , , , and . Then, the conditions () reduce toandrespectively. This system of inequalities can be written foraswhereNotice that , and . It follows thatbut not vice versa unless . Hence, we see that the general Theorem 1 if reduced provides a weaker convergence criterion than Kantorovich’s (8). Let us return back to Example 1 given in the introduction. Then, we have , because for each . It follows by last condition in () that for each . The condition (41) is verified provided that , which improves the convergence range for NM. Recall that the Kantorovich condition (8) does not hold for any . Application 1. Setand.
Then, NTM (12) reduces toFurther special cases of the method (43) are Newton’s method ifand the simplified Newton’s method provided that.
Other choices of the operatorsare possible [20,21,22].
An interesting choice seems to beNext, some local and semilocal convergence results are presented under these choices.
Theorem 3. Suppose
(i) the inverses are well defined;
(ii) there exists a solution of the Equation (1); (iii) there exists a parameter such that for each Then, the following assertions hold:andMoreover, if the operators exist and are uniformly bounded, then, the convergence order of the method (43) is two. Proof. In view of the choice (
44), we only need to show (
46), because then (
47) follows from it.
We can write
By applying (
44) and (
45) on (
48), we get in turn
leading to (
46) by (
43). □
Remark 7. Set in (44), then . Moreover, set . Then, the method (43) further reduces toSupposeThen, by Theorem 3, we deduceThus, the method (49) has convergence order two as Newton’s method but the ease of the simplified Newton’s method, as the operator is computed only once. Method (49) can be used provided that there exists an operator h, such thatNotice that . However, is known, because h is given. Hence, is determined. As an example, define the scalar function F to beThen, we have and . A second local convergence result follows under the condition (
52).
Proposition 3. Suppose
(i) the operator exists and .
(ii) there exists such thatwhereThen, the following assertions hold:and Proof. Mathematical induction is given immediately by (
51) and (
54). □
Example 3. Method (49) for F, given by (53), is defined byand converges to with the order two provided that for , because ς satisfies (54). Proposition 4. Suppose
(i) the operator T exists with ;
(ii) the operator h is Lipschitz continuous and for some and ;
(iii) , where .
Then, the following assertion holds:
Iteration (49) has a unique solution ,and . Proof. Define the operator
. Then, the method (
49) can be written as
. By using
and (ii), we obtain
Then, the result follows the celebrated contraction mapping principle. □
So far, we presented local convergence results. Next, we provide a semilocal convergence result.
Notice that if
h and
F are
and
(
) Lipschitz continuous, then by (
52) the operator
is Lipschitz continuous with parameter
. Moreover, we obtain
Set
and define the parameter
Furthermore, define the parameters
and
Theorem 4. SupposeThen, the Equation (1) has a solution , which is unique in . Proof. This follows immediately by the contraction mapping principle. □
Remark 8. The contraction mapping principle assures thatwhereHowever, by Theorem 3 the convergence order is two for where m is the smallest integer satisfyingHence, we have improved earlier works in this case. Application 2. Letand consider the fixed-point problemIt is know that a fixed pointsatisfies, then, clearly the method [23,24]is another specialization of the method (12). Ifwe get NM (3) andfor [
23] and , respectively. Example 4. Let us apply methods (3), (60) and (61) for solving the Equation (58) with function defined by and . Figure 1 contains graphs of the nonlinear functions , . The intersection point of the graphs and is the solution of the corresponding equation. From graphs (A) and (B) we see that and are solutions of the equations and , respectively. It is known that a condition is sufficient for the convergence of methods (60) and (61) [23,24]. It is possible to find intervals on which this condition will hold for both cases (see Figure 2). Table 1 shows the number of iterations that are needed to obtain the approximate solutions. Results are obtained under condition . The initial approximations were chosen from the intervals and . The approximations obtained at each iteration by the methods (60) and (61) are contained in specified intervals. Therefore, the condition was fulfilled. The obtained results show that the method (60) converges faster than Newton’s and Stirling’s methods. Define the scalar functionand choose . Then, the method (60) gives the exact solution after only one iteration. The method (61) converges after four iterations. But NM does not converge, provided that the function φ connects smoothly the other part of the function with . Note that in the neighborhood of the point , NM converges more slowly than methods (60) and (61). More advantages of the method (60) and (61) over Newton’s and other methods along the same lines as Application 1. Some possible choices of the function φ are given by and for each .