Next Article in Journal
Encouraging Students’ Motivation and Involvement in STEM Degrees by the Execution of Real Applications in Mathematical Subjects: The Population Migration Problem
Previous Article in Journal
Unsupervised Fault Diagnosis of Sucker Rod Pump Using Domain Adaptation with Generated Motor Power Curves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparison Study of the Classical and Modern Results of Semi-Local Convergence of Newton-Kantorovich Iterations

by
Samundra Regmi
1,
Ioannis K. Argyros
2,*,
Santhosh George
3 and
Christopher I. Argyros
4
1
Learning Commons, University of North Texas at Dallas, Dallas, TX 75201, USA
2
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505, USA
3
Department of Mathematical and Computational Sciences, National Institute of Technology Karnataka, Mangalore 575025, India
4
Department of Computing and Technology, Cameron University, Lawton, OK 73505, USA
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(8), 1225; https://doi.org/10.3390/math10081225
Submission received: 10 March 2022 / Revised: 26 March 2022 / Accepted: 7 April 2022 / Published: 8 April 2022

Abstract

:
There are a plethora of semi-local convergence results for Newton’s method (NM). These results rely on the Newton–Kantorovich criterion. However, this condition may not be satisfied even in the case of scalar equations. For this reason, we first present a comparative study of established classical and modern results. Moreover, using recurrent functions and at least as small constants or majorant functions, a finer convergence analysis for NM can be provided. The new constants and functions are specializations of earlier ones; hence, no new conditions are required to show convergence of NM. The technique is useful on other iterative methods as well. Numerical examples complement the theoretical results.
MSC:
49M15; 65J15; 65G99

1. Introduction

The concern of this study is to solve the following nonlinear equation
F ( x ) = 0 ,
using the celebrated Newton’s method (NM) in the following form
x n + 1 = x n F ( x n ) 1 F ( x n ) .
Here, F : Ω B 1 B 2 is differentiable according to Fréchet and operates between Banach spaces B 1 and B 2 , whereas subset Ω .
Kantorovich provided the semi-local convergence analysis of NM utilizing Banach’s contraction mapping principle. In particular, he presented two different proofs using majorant functions or recurrent relations [1]. The Newton-Kantorovich Theorem is that no assumption as to the solution is made, while at the same time the existence of the solution x * is established. Numerous researchers have used this theorem, both in applications and as a theoretical tool [2,3,4,5,6,7,8,9,10]. While the convergence criteria may not hold, NM may converge.
Assume that there exist constants η 0 , L 1 > 0 , and a point z Ω such that the Lipschitz condition
F ( z ) 1 ( F ( v ) F ( u ) ) L 1 v u
and
F ( z ) 1 F ( z ) η ,
hold for each u , v Ω . The main convergence criterion is provided by
2 L 1 η 1 .
There are even simple scalar equations where criterion (5) is not satisfied. Indeed, let Ω = U [ z , 1 q ] , q ( 0 , 1 2 ) , F ( t ) = t 3 q , and set z = x 0 = 1 . Then, conditions (3) and (4) are satisfied provided that L 1 = 2 ( 2 q ) and η = 1 q 3 . However, then condition (5) is not satisfied for any q ( 0 , 1 2 ) , as 2 L 1 η > 1 . Hence, there is a need to replace (5) with a weaker criterion without adding conditions. The same extension is useful in the Hölder case or when L 1 is replaced by a majorant function. The last two cases are not considered in [11]. Moreover, the Lipschitz case is extended further than in [11] (see the end of Section 4). This is the motivation for presenting new results that both extend the convergence region and provide more precise error estimates and better knowledge as to the location of the solution. In this way, the use of NM can be extended. The novelty of the present article is that these benefits require no additional conditions. The same technique used here can be applied to extend other iterative methods along the same lines.
Majorizing scalar sequences for NM are provided in Section 2. The semi-local convergence of NM appears in Section 3. A discussions is provided in Section 4, Examples can be found in Section 5, and Conclusions in Section 6.

2. Majorizing Sequences

Real functions are utilized to develop sequences that are shown to be majorizing for NM in Section 3. Let S = [ 0 , ) . Assume:
There exist functions ψ ¯ τ : S S , ψ 1 : S S non-decreasing and continuous such that function ψ 1 ( t ) 1 has a smallest zero b S { 0 } . Set S 0 = [ 0 , b ) . Consider function ψ τ : S 0 S for τ [ 0 , 1 ] continuous and non-decreasing.
Let η 0 be a given constant. Define scalar sequence { t n } for t 0 = 0 , t 1 = η by
t 2 = t 1 + 0 1 ψ ¯ τ ( τ ( t 1 t 0 ) ) d τ ( t 1 t 0 ) 1 ψ ¯ 1 ( t 1 ) , t n + 2 = t n + 1 + 0 1 ψ τ ( τ ( t n + 1 t n ) ) d τ ( t n + 1 t n ) 1 ψ 1 ( t n + 1 ) , n = 1 , 2 , .
Next, results appear on the convergence of sequence { t n } .
Lemma 1.
Assume:
ψ ¯ 1 ( η ) < 1 ,
t n b 0 f o r s o m e b 0 [ 0 , b ) , n = 2 , 3 , ,
and
ψ 1 ( b 0 ) < 1 .
Then, sequence { t n } is such that t n t n + 1 , n = 0 , 1 , 2 , and lim n t n = t * , where t * is the unique least upper bound of sequence { t n } satisfying t * [ 0 , b 0 ] .
Proof. 
It follows from definition (6), condition (7), and the definition of functions ψ ¯ 1 and ψ ¯ τ that t 1 t 2 . Similarly, the definitions of functions ψ 1 and ψ τ and condition (9) imply ψ 1 ( t n + 1 ) ψ 1 ( b 0 ) < 1 . Hence, t n + 1 t n + 2 , n = 1 , 2 , . Therefore, sequence { t n } is non-decreasing and bounded from above by b 0 . , and thus converges to t * .
Remark 1. (i) If function ψ 1 is strictly increasing, then b 0 in conditions (8) and (9) can be defined by
b : = ψ 1 1 ( 1 ) .
(ii), although the conditions of Lemma 1 can be verified by constructing sequence { t n } in advance. Next, convergence conditions which are stronger than (7)–(9) are provided, which are easier to verify. Let K 0 , K , L 0 , L 1 be non-negative constants and p [ 0 , 1 ] . Define functions
ψ ¯ 1 ( t ) = K 0 t p ,
ψ ¯ τ ( t ) = K ( τ t ) p ,
ψ 1 ( t ) = L 0 t p ,
and
ψ τ ( t ) = L ( θ t ) p .
Notice that for p = 1 the Lipschitz and for p ( 0 , 1 ) the Hölder case is obtained. It follows from these definitions that sequence { t n } is reduced for t 0 = a , t 1 = η to
t 2 = t 1 + K ( t 1 t 0 ) p ( 1 + p ) ( 1 K 0 t 1 p ) , t n + 2 = t n + 1 + L ( t n + 1 t n ) p ( 1 + p ) ( 1 L 0 t n + 1 p ) , n = 1 , 2 , .
Define parameters δ 0 , η 0 and δ 1 by
δ 0 = K ( t 2 t 1 ) p ( 1 + p ) ( 1 K 0 t 2 p ) , η 0 = t 2 t 1 ,
and
δ 1 = 1 η 0 ( 1 L 0 ) p t 1 .
Moreover, define functions on the interval [ 0 , 1 ) for k = 1 , 2 by
h k , p ( t ) = L 1 + p t k p η 0 + t L 0 ( t 1 + 1 t k + 1 1 t η 0 ) p t ,
and
g n + 1 , p ( t ) = L 1 + p t ( k + 1 ) η 0 p L 1 + p t k p η 0 p + t L 0 [ ( t 1 + ( 1 + t + + t k + 1 ) η 0 ) p ( t 1 + ( 1 + t + + t k ) η 0 p ] .
Next, the second convergence result for sequence { t n } obtained by (11) is presented, using the preceding notation.
Lemma 2.
Assume:
(H1) δ [ δ 0 , 1 ) , p ( 0 , 1 ] :
g n + 1 , p ( δ ) 0 , n = 1 , 2 , 3 , ,
or
(H2) δ [ δ 0 , 1 ) , p ( 0 , 1 ] :
g n + 1 , p ( δ ) 0 , n = 1 , 2 , 3 ,
Then, the following assertions hold under condition (H1):
h n , p ( δ ) h n + 1 , p ( δ ) ,
and the following under condition (H2):
h n , p ( δ ) h n + 1 , p ( δ ) .
Moreover, if under condition (H1)
0 δ 0 δ δ 1 ,
or under condition (H2)
h 1 , p ( δ ) 0
hold, then sequence { t n } obtained by (11) is such that 0 t n t n + 1 , n = 0 , 1 , 2 , and lim n t n = t * .
Proof. 
Induction is utilized to show
0 L ( t m + 1 t m ) p ( 1 + p ) ( 1 L 0 t m + 1 p ) δ .
Inequality (18) holds if m = 1 per the definition of δ 0 and δ . Then, it follows per (11) that
0 t 3 t 2 = δ 0 ( t 2 t 1 ) t 3 = t 2 + δ ( t 1 t 0 ) = t 2 + ( 1 + δ 0 ) ( t 2 t 1 ) ( t 2 t 1 ) = t 1 + ( 1 + δ 0 ) ( t 2 t 1 ) = t 1 + 1 δ 0 2 1 δ 0 ( t 2 t 1 )
t 1 + t 2 t 1 1 δ 0 t 1 + t 2 t 1 1 δ = t * * .
Assume estimates (19) and (20) hold for all values of m n . It follows by the induction hypotheses that
0 t m + 2 t m + 1 δ m ( t 2 t 1 ) ,
and
t m + 2 t m + 1 + δ m ( t 2 t 1 ) t m + δ m 1 ( t 2 t 1 ) + δ m ( t 2 t 1 ) t 1 + ( 1 + δ 0 ( 1 + δ + + δ m 1 ) ) ( t 2 t 1 ) t 1 + ( 1 + 1 δ m + 1 1 δ ( t 2 t 1 ) ) t * * .
Hence, per estimates (21) and (22), assertion (18) holds if
L 1 + p δ m ( t 2 t 1 ) p + δ L 0 t 1 ( 1 + δ + + δ m ) η 0 p δ 0 ,
or
h m , δ ( t ) 0 a t t = δ .
The recurrent functions are related, as
h m + 1 , p ( t ) = h m + 1 , p ( t ) h m , p ( t ) + h m , p ( t ) = L 1 + p t ( m + 1 ) p η 0 p + t L 0 ( t 1 + ( 1 + t + + t m + 1 ) η 0 ) p t + h m , p ( t ) L 1 + p t m p η 0 p + t L 0 ( t 1 + ( 1 + t + + t m ) η 0 ) p + t , = h m , p ( t ) + g m + 1 , p ( t ) .
Case of Condition (H1): Define function
h , p ( t ) = lim m h m , p ( t ) .
Per estimates (23) and (26), it follows that
h m + 1 , p ( δ ) h m , p ( δ ) ,
and
h , p ( t ) = t [ L 0 ( t 1 + η 0 1 t ) p 1 ] .
Thus, estimate (24) holds if
h , p ( δ ) 0 ,
which is true per condition (16). The induction under (H1) is terminated.
Case of Condition (H2): It follows that estimate (24) holds, as
h m , p ( δ ) h 1 , p ( δ ) 0 .
That is, the induction is completed under condition (H2) as well. Hence, under either condition (H1) or (H2), sequence { t m } is non-decreasing and bounded from above. Hence, it converges to t * .
Remark 2.
Define function f p : [ 0 , 1 ] S by
f p ( t ) = L 0 t 1 + p + L 1 + p t p L 1 + p .
It follows that f p ( 0 ) L 1 + p < 0 and f p ( 1 ) = L 0 > 0 . Hence, the intermediate value theorem ensures that the equation f p ( t ) = 0 has zeros in ( 0 , 1 ) . Denote by γ p the smallest such zero. Notice that
γ 1 = 2 L L + L 2 + 8 L 0 L .
The parameter δ is preferred in closed form. This is possible in certain cases.
Lipschitz Case: Set p = 1 . Then, the choice for δ = γ 1 .
In the case that p ( 0 , 1 ) , the zero γ p is not in closed form; instead
γ p L L + ( 1 + p ) L 0 p 1 + p .
Indeed, because γ p solves equation f p ( t ) = 0 , then
L = L 0 ( 1 + p ) γ p 1 + 1 p + L γ p ( L 0 ( 1 + p ) + L ) γ p 1 + 1 p
leading to condition (33). Under (H1), one obtains h k + 1 , 1 ( t ) = h k , 1 ( t ) + f 1 ( t ) t n 1 η , and condition (16) becomes
δ 0 γ 1 δ 1 .
By solving the left hand side inequality for η,
η 2 γ 1 ( 1 + 2 γ 1 ) L 0 .
The right hand side inequality in condition (34) is solved as follows:
Case: 2 ( 1 γ 1 ) K 0 K = 0 . Set
η 1 = 1 K 0 + L 0 .
Hence, (34) holds if
η min { 2 γ 1 ( 1 + 2 γ 1 ) L 0 , 1 K 0 + L 0 } .
To consider the rest of the cases, define function φ : S S by
φ ( t ) = L 0 ( 2 γ 1 ) K 0 K ) t 2 + 2 ( 1 γ 1 ) ( K 0 + L 0 ) t + 2 ( 1 δ ) .
Denote by Δ the discriminant of φ . If 2 ( 1 δ ) K 0 K > 0 and Δ < 0 , set
η 2 = 2 γ 1 L 0 ( 1 + 2 γ 1 ) .
Then, condition (34) holds if
η η 2 .
If 2 ( 1 δ ) K 0 K > 0 and Δ > 0 , then φ has a positive zero denoted by η 3 . Then, (34) holds if
η min { 2 γ 1 L 0 ( 1 + 2 γ 1 ) , η 3 } .
If 2 ( 1 δ ) K 0 K < 0 and Δ > 0 , then denote by η 4 the positive zero of function φ . Then, condition (34) holds if
η min { 2 γ 1 L 0 ( 1 + 2 γ 1 ) , η 4 } .
If 2 ( 1 δ ) K 0 K < 0 and Δ < 0 , the right hand side of (34) is not solvable and does not hold.
Hölder Case. Let p ( 0 , 1 ) . Per the definition of functions h m , p , g m + 1 , p , it follows that
h m + 1 , p ( t ) h m , p ( t ) + 1 1 + p t ( m + 1 ) p η 0 p L 1 + p t m η 0 p + t L 0 t m + 1 η 0 p , h m , p ( t ) + f p ( t ) f m p η 0 p .
Thus, condition (H2) becomes
h 1 , p ( γ p ) 0 a n d 0 δ 0 γ p ,
where
h 1 , p ( t ) = L 1 + p t p η 0 p + t L 0 ( t 1 + ( 1 + t ) ) p η 0 p t .
The convergence conditions in this Section are compared to existing ones in Section 4.

3. Convergence of NM

The notation U ( w , ρ ) , U [ w , ρ ] refers to the open and closed balls with radius ρ > 0 and center w X , respectively. The semi-local convergence of NM is presented under conditions (A). The functions ψ are provided in Section 2. Assume:
(A1)
There exist x 0 Ω , η 0 such that F ( x 0 ) 1 L ( B 2 , B 1 ) and F ( x 0 ) 1 F ( x 0 ) η .
(A2)
F ( x 0 ) 1 ( F ( x 1 ) F ( x 0 ) ) ψ ¯ 1 ( x 1 x 0 ) and
F ( x 0 ) 1 ( F ( x 0 + τ ( x 1 x 0 ) ) F ( x 0 ) ) ψ ¯ τ ( x 1 x 0 )
τ [ 0 , 1 ) , where x 1 = x 0 F ( x 0 ) 1 F ( x 0 ) .
(A3)
F ( x 0 ) 1 ( F ( x ) F ( x 0 ) ) ψ 1 ( x x 0 ) , x Ω . Set Ω 0 = U ( x 0 , b ) Ω .
(A4)
F ( x 0 ) 1 ( F ( x + τ ( y x ) ) F ( x ) ) ψ τ ( y x ) x , y Ω 0 , τ [ 0 , 1 ) .
(A5)
Conditions of Lemma 1 hold and
(A6)
U [ x 0 , t * ] Ω (or U [ x 0 , b 0 ] Ω ).
Next, the main semi-local convergence result for NM is developed.
Theorem 1.
Assume conditions (A) hold. Then, the sequence { x n } generated by NM exists in U [ x 0 , t * ] , stays in U [ x 0 , t * ] for all n = 0 , 1 , 2 , and converges to a solution x * U [ x 0 , t * ] of equation F ( x * ) = 0 .
Proof. 
Estimates
x i + 1 x i t i + 1 t i ,
and
U [ x i + 1 , t * t i + 1 ] U [ x i , t * t i ] ,
i = 0 , 1 , 2 , shall be shown by induction. Let u U [ x 1 , t * t i + 1 ] U [ x i , t * t i ] . It follows that
u x 0 u x 1 + x 1 x 0 t * t 1 + t 1 t 0 = t * .
Thus, u U [ x 0 , t * t 0 ] . Then, we obtain
x 1 x 0 = F ( x 0 ) 1 F ( x 0 ) η = t 1 t 0
per condition (A1), implying that estimates (43) and (44) hold for i = 0 . Assume (43) and (44) hold for all n i . It follows that
x i + 1 x 0 j = 1 i + 1 x j x j 1 , j = 1 i + 1 ( t j t j 1 ) = t i + 1 t 0 = t i + 1 t * ,
and
x i + τ ( x i + 1 x i ) t i + τ ( t i + 1 t i ) t * , τ [ 0 , 1 ] .
Let v U [ x 0 , t * ] . Using (A2), we obtain
F ( x 0 ) 1 ( F ( v ) F ( x 0 ) ) ψ ¯ 1 ( v x 0 ) , ψ ¯ 1 ( η ) ψ 1 ( t * ) < 1 .
Hence, F ( v ) 1 L ( B 2 , B 1 ) and
F ( v ) 1 F ( x 0 ) 1 1 ψ ¯ 1 ( v x 0 ) ,
are derived by Lemma on invertible linear operators due to Banach [2,9,12]. In particular, for v = x 1 , iterate x 2 exists per the second substep of NM. Moreover, one can write
F ( x 1 ) = F ( x 1 ) F ( x 0 ) F ( x 0 ) ( x 1 x 0 ) , = 0 1 ( F ( x 0 + τ ( x 1 x 0 ) ) d τ F ( x 0 ) ) ( x 1 x 0 ) .
Then, by (A2) and estimate (46), it follows that
F ( x 0 ) 1 F ( x 1 ) 0 1 ψ ¯ τ ( x 1 x 0 ) d τ x 1 x 0 , 0 1 ψ ¯ τ ( t 1 t 0 ) d τ ( t 1 t 0 ) .
It then follows from estimates (46) (for v = x 1 ), (48) and NM for n = 2 that
x 2 x 1 F ( x 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 1 ) , 0 1 ψ ¯ τ ( t 1 t 0 ) d τ ( t 1 t 0 ) 1 ψ ¯ 1 ( t 1 ) = t 2 t 1 .
Similarly, we can write
F ( x i + 1 ) = F ( x i + 1 ) F ( x i ) F ( x i ) ( x i + 1 x i ) , = 0 1 ( F ( x i + τ ( x i + 1 x i ) ) d τ F ( x i ) ) ( x i + 1 x i ) .
Thus, per condition (A4) and estimates (46) and (50) (for v = x i + 1 ),
F ( x 0 ) 1 F ( x i + 1 ) 0 1 ψ τ ( x i + 1 x i ) d τ x i + 1 x i , 0 1 ψ τ ( t i + 1 t i ) d τ ( t i + 1 t i ) .
Consequently,
x i + 2 x i + 1 F ( x i + 1 1 F ( x 0 ) F ( x 0 ) 1 F ( x i + 1 ) , 0 1 ψ τ ( t i + 1 t i ) d τ ( t i + 1 t i ) 1 ψ 1 ( t i + 1 ) t i + 2 t m + 1 .
Hence, estimate (43) holds i . Furthermore, if w U [ x i + 2 , t * t i + 2 ] , one obtains
w x i + 1 w x i + 2 + x i + 2 x i + 1 t * t i + 2 + t i + 2 t i + 1 = t * t i + 1 ;
thus, w U [ x i + 1 , t * t i + 1 ] . Induction for estimates (43) and (44) is terminated. Sequence { t i } is fundamental as convergent. Hence, per estimate (43) and (44), sequence { x n } is fundamental in a Banach space B 1 as well, and thus converges to some x * U [ x 0 , t * ] . Per the continuity of F , and letting i in estimate (48), it follows that F ( x * ) = 0 .
Next, we present a uniqueness of the solution result.
Proposition 1.
Assume:
(i) Point x * is a simple solution of equation F ( x ) = 0 in U [ x 0 , r 0 ] Ω for some r 0 > 0 , and condition (A3) holds.
(ii) There exists r r 0 such that
0 1 ψ 1 ( ( 1 τ ) r 0 + τ r ) d τ < 1 .
Let Ω 1 = U [ x 0 , r ] Ω . Then, the only solution of equation F ( x ) = 0 is x * .
Proof. 
Assume there exists x * Ω 1 such that F ( x * ) = 0 . Set J = 0 1 F ( x * + τ ( x * x * ) ) d τ . Then, it follows from conditions (A3) and (54) that
F ( x 0 ) 1 ( J F ( x 0 ) ) 0 1 ψ 1 ( ( 1 τ ) x 0 x * + τ x 0 x * ) d τ , 0 1 ψ 1 ( ( 1 τ ) r 0 + τ r ) d τ < 1 .
Thus, x * = x * is implied by the invertibility of J , and the approximation J ( x * x * ) = F ( x * ) F ( x * ) = 0 .
Remark 3.
Proposition 1 uses only condition (A3). However, under conditions (A) set r 0 = t * .

4. Discussion and Conclusions

There are a plethora of results on NM. The sufficient convergence criteria in these studies are weakened by the new technique without adding conditions (see the numerical Section). Notice that with the developed technique there is an at least as tight subset Ω 0 of Ω containing the iterates. Therefore, the Lipschitz–Hölder majorant functions are at least as tight as the ones provided in earlier studies, leading to the advantages stated in the introduction. Below, we provide several examples.
Assume [2,9,13]
F ( x 0 ) 1 ( F ( x ) F ( y ) ) ω ( x y )
for all x , y Ω . Define function
Q ¯ ( t ) = 0 r ( r t ) ω ( t ) d t t + η
and iteration { t ¯ n } per t ¯ 0 = 0 , t ¯ 1 = η :
t ¯ n + 1 = t ¯ n + Q ¯ ( t ¯ n ) Q ¯ ( t ¯ n ) .
The sufficient convergence criterion is derived supposing that scalar function Q ¯ has a smallest positive zero r Q ¯ . However, under condition (A4), define instead for t 0 = 0 , t 1 = η
t n + 1 = t n + Q ( t n ) Q ( t n ) ,
Q ( t ) = 0 r ( r t ) ψ 1 ( t ) d t t + η .
Notice that
Q ( t ) Q ¯ ( t ) .
In particular as ψ 1 ( t ) ω ( t ) .
Q ( r ¯ τ ) Q ¯ ( r ¯ Q ) = 0 ,
and Q ( 0 ) = 0 , Q has an at least as small zero r Q . Hence, the new convergence criterion is at least as weak. This is the contribution in the case of majorant functions.
In particular, in the case where ω and ψ 1 are constant functions, such as
ω ( t ) = L 1 t ,
ψ 1 ( t ) = L t ,
then
η 1 2 L 1
and
η 1 2 L ,
which is weaker, as L L 1 and L 1 is the Lipschitz constant on Ω .
(a) Lipschitz Case: Convergence criteria already in the literature are listed below in the left column, and the new criteria in the right column.
T 1 : η 1 μ 1 ,    [1]    T ˜ 1 : η 1 μ ˜ 1 ,                                          (66)
T 2 : η 1 μ 2 ,    [12]    T ˜ 2 : η 1 μ ˜ 2 ,                                         (67)
T 3 : η 1 μ 3 ,    [11,14]    T ˜ 3 : η 1 μ ˜ 3 ,                                    (68)
T 4 : η 1 μ 4 ,    [12]    T ˜ 4 : η 1 μ ˜ 4                                          (69)
where
μ 1 = 2 L 1 , μ ˜ 1 = 2 L ,
μ 2 = L 0 + L 1 , μ ˜ 2 = L 0 + L ,
μ 3 = 1 4 ( L 1 + 4 L 0 + L 2 + 8 L 0 L 1 ) , μ ˜ 3 = 1 4 ( L + 4 L 0 + L 2 + 8 L 0 L ) ,
and
μ 4 = 1 4 ( 4 L 0 + L 0 L 1 + 8 L 0 2 + L 0 L 1 ) , μ ˜ 4 = 1 4 ( 4 L 0 + L 0 L + 8 L 0 2 + L 0 L ) .
It follows from these definitions that
L 0 L 1 and L L 1 .
Consequently,
μ 4 μ 3 μ 2 μ 1 , μ ˜ 4 μ ˜ 3 μ ˜ 2 μ ˜ 1 ,
( T 1 ) ( T 2 ) ( T 3 ) ( T 4 ) , and ( T ˜ 1 ) ( T ˜ 2 ) ( T ˜ 3 ) ( T ˜ 4 ) .
although not necessarily vice versa unless L 0 = L = L 1 . Parameter L 1 is the Lipschitz constant on Ω used by Kantorovich and others [1,7]. Another comparison between convergence criteria can be made as follows. Notice that
1 μ 1 1 μ 2 = μ 2 μ 1 = 1 2 1 + L 0 L 1 2 as L 0 L 0 .
Similarly,
1 μ 2 1 μ 3 1 4 , 1 μ 2 1 μ 3 1 2 , 1 μ 1 1 μ 4 0 , 1 μ 3 1 μ 4 0 , and 1 μ 2 1 μ 4 0 as L 0 L 0 .
These limits show by how many times at most the applicability of NM can be extended. Moreover, under the new approach this applicability is extended even further. Indeed, notice that
1 μ 1 1 μ ˜ 1 = μ ˜ 1 μ 1 = 2 L 2 L 1 = L L 1 0 ,
1 μ 2 1 μ ˜ 2 = μ ˜ 2 μ 2 = L 0 L 1 + L L 1 L 0 L 1 + 1 0 as L 0 L 1 0 and L L 1 0 .
Similarly,
1 μ 3 1 μ ˜ 3 0 and 1 μ 4 1 μ ˜ 4 0 as L 0 L 1 0 and L L 1 0 .
Clearly, the new convergence criterion can be arbitrarily many times weaker than the previous ones. This technique of replacing constants with smaller ones leads to both weaker criteria and more precise majorizing sequences. Indeed, consider again iterations { t ¯ n } and { t n } with the choice of functions ω and ψ 1 obtained by (62) and (63), respectively. It follows that
0 t n t ¯ n ,
then
0 t n + 1 t n t ¯ n + 1 t ¯ n
and
s 1 = lim n t n lim n t ¯ n = s 2 ,
as
L L 1 ,
t n + 1 t n = L ( t n + 1 t n ) 2 2 ( 1 L t n ) ,
and
t ¯ n + 1 t ¯ n = L 1 ( t ¯ n + 1 t ¯ n ) 2 2 ( 1 L 1 t ¯ n )
where
r 1 = 1 1 2 L η L and r 2 = 1 1 2 L 1 η L 1 .
The constant r 2 was reported by Kantorovich. Hence, tighter error estimates on x n + 1 x n and x * x n become possible if L < L 1 .
Moreover, the information on the location of the solution is more accurate if s 1 < s 2 , which is possible for L < L 1 . Concerning the uniqueness of the solution balls, u ( x 0 , r 3 ) u ( x 0 , r 4 ) where
r 3 = 1 + 1 2 L η L 1 r 4 = 1 + 1 2 L η L .
Thus, the uniqueness of the solution ball is extended. The constant r 3 is provided by Kantorovich. Furthermore, if we choose ψ 1 ( t ) = L 0 t and assume η 1 2 L 0 , set r 5 = 1 1 2 L 0 η L 0 . Then, r 4 < r 5 if L 0 < L , and the solution x * is unique in the ball u ( x 0 , r 5 ) according to Proposition 1. This is the contribution in the Lipschitz case.
(b) Hölder Cases:
( H 1 ) : L 1 η p p 1 + p p ,
( H 2 ) : L 1 η p 2 p 1 p 1 + p p ,     [1,2,4,6,13,15]
where the new contributions are, respectively,
( H 1 ) : L η p p 1 + p p ,
and
( H 2 ) : L η p p 1 + p p .
Hence,
( H 1 ) ( H 1 ) ,
and
( H 2 ) ( H 2 ) .
Notice that the proofs of ( H 1 ) and ( H 2 ) are implied by ( H 1 ) , ( H 2 ) , respectively, if Ω 0 is used instead of Ω . The conditions of Lemma 1 are weaker than all preceding, whereas those of Lemma 1 can be weaker as well, as L 0 L 1 and L L 1 (see the numerical Section).
Notice again that the Hölder conditions of the Lemmas and Remark 2 are weaker as well. This is the contribution in the Hölder case.
  • (c) It turns out that the results can be weakened even more if Lipschitz or Hölder constants are replaced by constants at least as small in two different ways, namely:
  • (i) Noting that convergence conditions should not hold for all x , y Ω 0 , only for x Ω 0 and y = x F ( x ) 1 F ( x ) , the proofs of the results proceed in this weaker setting. However, the corresponding constant is at least as tight. Denote such a constant per L 2 ; it follows that L 2 L .
  • (ii) Consider ball Ω 2 = U ( x 1 , b 0 η ) Ω for b 0 > η and have Ω 2 replace Ω 0 in condition (A3). Then, as Ω 2 Ω 0 , the constant is again at least as small. Denote such a constant per L 3 . It follows that L 3 L . The rest of the convergence results found in the literature can be immediately weakened under this technique as well.
Finally, it is worth noticing that the new constants L 0 , L , L 2 , L 3 are special cases of L 1 . Hence, the benefits under the new technique are obtained without additional computational effort or conditions, which represents one contribution of this paper.

5. Numerical Experiments

In the first example, the new constants are smaller than those of earlier studies.
Example 1.
Define scalar function
F ( t ) = λ 0 t + λ 1 + λ 2 sin e λ 3 t ,
for t 0 = 0 where λ j and j = 0 , 1 , 2 , 3 are real parameters. It follows by this definition that for λ 3 sufficiently large and for λ 2 sufficiently small L 0 L 1 can be arbitrarily small. In particular, L 0 L 1 0 .
In the last two examples, Kantorovich criterion (66) is not satisfied; however, the new criterion (66) holds.
Example 2.
Let B 1 = B 2 = R , Ω = [ q , 2 q ] for q M = ( 0 , 0.5 ) , and t 0 = 1 . Define real function F on Ω as
F ( t ) = t 3 q .
Then, the parameters are η = 1 3 ( 1 q ) , K 0 = K = L 0 = 3 q , a n d L 1 = 2 ( 2 q ) . Moreover, one obtains Ω 0 = U ( 1 , 1 q ) U ( 1 , 1 L 0 ) = U ( 1 , 1 L 0 ) , thus L = 2 ( 1 + 1 3 q ) . Denote by S i , i = 1 , 2 , 3 , 4 the set of values q for which conditions ( T i ) are satisfied. Then, by solving inequalities (66)–(69) for q, S 1 = , S 2 = [ 0.464816242 , 0.5 ) , S 3 = [ 0.45039002 , 0.5 ) , and S 4 = [ 0.4271907643 , 0.5 ) , respectively.
Next, the new conditions ( T ˜ i ) are tested. Concerning the new criterion (66), the inequality
4 3 1 + 1 3 q ( 1 q ) 1
must be solved, obtaining q 17 177 8 = 0.461983163 ; thus,
S ˜ 1 = [ 0.461983163 , 0.5 ) .
Similarly, when solving for q in new criteria (67)–(69), the intervals S 2 , S 3 , S 4 are extended. This is due to the fact that L < L 1 for all q in M. Hence, the range of values q is extended in the new approach. Notice in particular that the Newton–Kantorovich criterion (66) is not satisfied for any q ( 0 , 0.5 ) .
Example 3.
Let B 1 = B 2 = C [ 0 , 1 ] be the domain of continuous real functions defined on the interval [ 0 , 1 ] . The max-norm is used. Set Ω = U [ x 0 , 3 ] and define operator F on Ω as
F ( v ) ( v 1 ) = v ( v 1 ) y ( v 1 ) 0 1 N ( v 1 , t ) v 3 ( t ) d t , v C [ 0 , 1 ] , v 1 [ 0 , 1 ]
where y is given in C [ 0 , 1 ] and N is a kernel obtained by Green’s function as
N ( v 1 , t ) = ( 1 v 1 ) t , t v 1 v 1 ( 1 t ) , v 1 t .
It follows per this definition that the derivative of F
[ F ( v ) ( z ) ] ( v 1 ) = z ( v 1 ) 3 0 1 N ( v 1 , t ) v 2 ( t ) z ( t ) d t
z C [ 0 , 1 ] , v 1 [ 0 , 1 ] . Pick x 0 ( v 1 ) = y ( v 1 ) = 1 . It then follows from (78)–(80) that F ( x 0 ) 1 L ( B 2 , B 1 ) ,
I F ( x 0 ) < 0.375 , F ( x 0 ) 1 1.6 ,
η = 0.2 , L 0 = 2.4 , L 1 = 3.6 ,
and Ω 0 = U ( x 0 , 3 ) U ( x 0 , 0.4167 ) = U ( x 0 , 0.4167 ) , thus L = 1.5 . Notice that L 0 < L 1 and L < L 1 . Choose K 0 = K = L 0 . The Kantorovich convergence criterion (66) is not satisfied, as 2 L 1 η = 1.44 > 1 . Hence, convergence of converge NM is not assured by the Kantorovich criterion. However, new criterion (66) is satisfied, as 2 L η = 0.6 < 1 .

6. Conclusions

A comparison study between results on the semi-local convergence of NM is provided. The technique uses recurrent functions. In this way, the new sufficient convergence criteria are weaker in the Lipschitz, Hölder, and more general cases, as they rely on smaller constants. Other benefits include more precise error bounds and uniqueness of the solution results. The new constants are special cases of earlier ones. The technique is very general, rendering it useful to extend the usage of other iterative methods.

Author Contributions

Conceptualization, S.R., C.I.A., I.K.A. and S.G.; methodology, S.R., C.I.A., I.K.A. and S.G.; software, S.R., C.I.A., I.K.A. and S.G.; validation, S.R., C.I.A., I.K.A. and S.G.; formal analysis, S.R., C.I.A., I.K.A. and S.G.; investigation, S.R., C.I.A., I.K.A. and S.G.; resources, S.R., C.I.A., I.K.A. and S.G.; data curation, S.R., C.I.A., I.K.A. and S.G.; writing—original draft preparation, S.R., C.I.A., I.K.A. and S.G.; writing—review and editing, S.R., C.I.A., I.K.A. and S.G.; visualization, S.R., C.I.A., I.K.A. and S.G.; supervision, S.R., C.I.A., I.K.A. and S.G. project administration, S.R., C.I.A., I.K.A. and S.G.; funding acquisition, S.R., C.I.A., I.K.A. and S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kantorovich, L.V.; Akilov, G.P. Functional Analysis; Pergamon Press: Oxford, UK, 1982. [Google Scholar]
  2. Appell, J.; De Pascale, E.; Lysenko, J.V.; Zabreiko, P.P. New results on Newton-Kantorovich approximations with applications to nonlinear integral equations. Numer. Funct. Anal. Optim. 1997, 18, 1–18. [Google Scholar] [CrossRef]
  3. Cianciaruso, F.; De Pascale, E. The Newton-Kantorovich approximations when the derivative is Hölder: Old and New Results. Numer. Funct. Anal. 2003, 24, 713–723. [Google Scholar] [CrossRef]
  4. Demidovich, N.T.; Zabreiko, P.P.; Lysenko, J.V. Some remarks on the Newton-Kantorovich method for nonlinear equations with Hölder continuous linearizations. Izv. Akad. Nauk Beloruss 1993, 3, 22–26. (In Russian) [Google Scholar]
  5. De Pascale, E.; Zabreiko, P.P. The convergence of the Newton-Kantorovich method under Vertgeim conditions: A new improvement. Z. Anal. Anwend. 1998, 17, 271–280. [Google Scholar] [CrossRef] [Green Version]
  6. Ezquerro, J.A.; Hernandez, M.A. Newton’s Methods: An Updated Approach of Kantorovich’s Theory; Birkhäuser: Cham, Switzerland, 2017. [Google Scholar]
  7. Ortega, L.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  8. Potra, F.A.; Pták, V. Nondiscrete Induction and Iterative Processes; Research Notes in Mathematics, 103; Pitman (Advanced Publishing Program): Boston, MA, USA, 1984. [Google Scholar]
  9. Proinov, P.D. New general convergence theory for iterative processes and its applications to Newton-Kantorovich type theorems. J. Complex. 2010, 26, 3–42. [Google Scholar] [CrossRef] [Green Version]
  10. Yamamoto, T. Historical developments in convergence analysis for Newton’s and Newton-like methods. J. Comput. Appl. Math. 2000, 124, 1–23. [Google Scholar] [CrossRef] [Green Version]
  11. Argyros, I.K. Unified Convergence Criteria for Iterative Banach space Valued Methods with Applications. Mathematics 2021, 9, 1942. [Google Scholar] [CrossRef]
  12. Argyros, I.K.; Hilout, S. On an improved convergence analysis of Newton’s scheme. Appl. Math. Comput. 2013, 225, 372–386. [Google Scholar]
  13. Zabreiko, P.P.; Nguen, D.F. The majorant method in the theory of Newton-Kantorovich approximations and the Pták error estimates. Numer. Funct. Anal. and Optim. 1987, 9, 671–684. [Google Scholar] [CrossRef]
  14. Argyros, I.K.; Hilout, S. Weaker conditions for the convergence of Newton’s method. J. Complex. 2012, 28, 364–387. [Google Scholar] [CrossRef] [Green Version]
  15. Vertgeim, B.A. On some methods of the approximate solution of nonlinear functional equations in Banach spaces. Engl. Transl. Am. Math. Soc. Transl. 1960, 16, 378–382. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Regmi, S.; Argyros, I.K.; George, S.; Argyros, C.I. A Comparison Study of the Classical and Modern Results of Semi-Local Convergence of Newton-Kantorovich Iterations. Mathematics 2022, 10, 1225. https://doi.org/10.3390/math10081225

AMA Style

Regmi S, Argyros IK, George S, Argyros CI. A Comparison Study of the Classical and Modern Results of Semi-Local Convergence of Newton-Kantorovich Iterations. Mathematics. 2022; 10(8):1225. https://doi.org/10.3390/math10081225

Chicago/Turabian Style

Regmi, Samundra, Ioannis K. Argyros, Santhosh George, and Christopher I. Argyros. 2022. "A Comparison Study of the Classical and Modern Results of Semi-Local Convergence of Newton-Kantorovich Iterations" Mathematics 10, no. 8: 1225. https://doi.org/10.3390/math10081225

APA Style

Regmi, S., Argyros, I. K., George, S., & Argyros, C. I. (2022). A Comparison Study of the Classical and Modern Results of Semi-Local Convergence of Newton-Kantorovich Iterations. Mathematics, 10(8), 1225. https://doi.org/10.3390/math10081225

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop