Next Article in Journal
Fatty Acid Vesicles as Hard UV-C Shields for Early Life
Previous Article in Journal
Investigation of the Acid/Basic Sites of Zeolite Trough Some Catalysed Nucleophilic Reactions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extended Newton-like Midpoint Method for Solving Equations in Banach Space

1
Department of Computing and Mathematical Sciences, Cameron University, Lawton, OK 73505, USA
2
Department of Mathematics, Hans Raj Mahila Mahavidyalaya, Jalandhar 144008, Punjab, India
3
Department of Mathematics, University of Houston, Houston, TX 77204, USA
*
Author to whom correspondence should be addressed.
Foundations 2023, 3(1), 82-98; https://doi.org/10.3390/foundations3010009
Submission received: 14 December 2022 / Revised: 2 February 2023 / Accepted: 2 February 2023 / Published: 21 February 2023
(This article belongs to the Section Mathematical Sciences)

Abstract

:
In this study, we present a convergence analysis of a Newton-like midpoint method for solving nonlinear equations in a Banach space setting. The semilocal convergence is analyzed in two different ways. The first one is shown by replacing the existing conditions with weaker and tighter continuity conditions, thereby enhancing its applicability. The second one uses more general ω -continuity conditions and the majorizing principle. This approach includes only the first order Fréchet derivative and is applicable for problems that were otherwise hard to solve by using approaches seen in the literature. Moreover, the local convergence is established along with the existence and uniqueness region of the solution. The method is useful for solving Engineering and Applied Science problems. The paper ends with numerical examples that show the applicability of our convergence theorems in cases not covered in earlier studies.

1. Introduction

One of the most challenging problems in Engineering and Applied Sciences is to determine a locally unique solution x * of a nonlinear equation
F ( x ) = 0 ,
where the operator F is defined on the Banach space B 1 with values in a Banach space B 2 . As an example, engineering problems reduce to solving differential or integral equations, which in turn are set up as (1). A solution x * of the Equation (1) is difficult to find in closed form. That forces researchers to develop iterative methods, which generate iterations convergent to x * , provided that certain initial conditions hold.
A popular iterative method is defined for each m = 0 , 1 , 2 , by
x 0 Ω B 1 , x m + 1 = x m F ( x m ) 1 F ( x m ) .
This is the so-called Newton’s method (NM), which is only quadratically convergent [1,2,3,4]. In order to increase the order of convergence as well as the efficiency, a plethora of iterative methods have been developed (see, e.g., [5,6,7,8,9] and references therein). Among those, special attention has been given to the Newton-like midpoint method (NLMM) defined by
x 0 Ω , y 0 = x 0 F ( x 0 ) 1 F ( x 0 ) x m + 1 = x m F x m + y m 2 1 F ( x m ) , m = 0 , 1 , 2 , y m = x m F x m 1 + y m 1 2 1 F ( x m ) , m = 1 , 2 , 3 , .
NLMM requires per iteration one operator evaluation and one of the inverse of F . The efficiency index according to Ostrowski is shown to be approximately 1.5537 [10]. This index is higher than NM ( 1.4142 ), as well as the one given in [11] ( 1.3160 ). The construction of this method was essentially given in [12] when B 1 = B 2 but with no formal proof of convergence. That is why the semilocal convergence is developed in [10] under Kantorovich’s hypotheses. Moreover, favorable comparisons are given to methods using similar information.
  • Motivation for writing this study: The following concerns arise with the applicability of this method in general.
( 1 )
The convergence region in [10] is not large.
( 2 )
The upper bounds on the distances x n x * and x n + 1 x n are not tight enough.
( 3 )
The uniqueness region of the solution x * is not large.
( 4 )
A Lipschitz condition on the second derivative is assumed (see the condition ( H 3 ) in Section 3). However, the second derivative does not appear on the method and may not exist (see the numerical example in Section 4). Additionally, the method may converge.
( 5 )
The local convergence analysis is not studied in [10].
  • Novelty: Due to the importance of this method, the items ( 1 ) ( 5 ) are positively addressed. The current study includes two procedures for analyzing the semilocal convergence of NLMM. The first analysis replaces the conditions used in [10] with weaker and tighter conditions, thereby enlarging the uniqueness region. In the second semilocal convergence, the convergence conditions used in the earlier section have been replaced by more generalized ω -continuity conditions using majorizing sequences [1,4,5,11,12,13,14,15,16,17]. The main advantage of this approach is that it uses only the first derivative, which actually appears in NLMM, for proving the convergence result instead of the second derivative used in [10], thereby enhancing its applicability. Thus, our work improves the results derived in [10] under more stringent conditions and generates finer majorizing sequences. The innovation of the study lies in the fact that extensions are achieved under weaker conditions (see also Remarks throughout the paper). The local convergence of NLMM is also established, along with the existence and uniqueness region of the solution. Moreover, the new error analysis is finer, requiring fewer iterates to achieve a predetermined error tolerance. Furthermore, more precise information is provided on the uniqueness domain of the solution. Finally, the technique can be used on other methods utilizing the inverse of an operator.
The rest of the paper is structured as follows: Section 2 includes mathematical background. In Section 3, we develop the first kind of semilocal convergence theorem based on weaker conditions. The generalized ω -continuity conditions are applied to prove the second type of semilocal convergence theorem in Section 4. The local convergence, along with the uniqueness results of NLMM, is studied in Section 5. In Section 6, numerical examples are given to illustrate the theoretical results. Concluding remarks are reported in Section 7.

2. Mathematical Background

The study of the behavior of a certain cubic polynomial and the corresponding scalar Newton function play a role in the semilocal convergence of NLMM. Let L > 0 , M 0 and d 0 be given parameters. Define the cubic polynomial
q ( L , M , d ) ( t ) = q ( t ) = L 6 t 3 + M 2 t 2 t + d ,
the Newton iteration function
N q ( t ) = t q ( t ) 1 q ( t )
and the scalar sequences { u m } , { v m } for
u 0 = 0 , v 0 = u 0 q ( u 0 ) 1 q ( u 0 ) , v m = u m q u m 1 + v m 1 2 1 q ( u m )
and
u m + 1 = u m q u m + v m 2 1 q ( u m ) .
The proof of the following auxiliary result containing some properties of q, N q , { v m } , and { u m } can be found in [10].
Lemma 1.
Suppose:
6 d M 3 + 9 d 2 L 2 + 18 d M L 3 M 2 8 L 0
or
d 4 L + M 2 M M 2 + 2 L 3 L ( M + M 2 + 2 L ) = λ ( L , M ) .
Then, the following assertions hold:
(i) 
The polynomial q given by the Formula (4) has two zeros u * , u * * with 0 < u * u * * .
(ii) 
q is decreasing in the interval [ 0 , u * ] .
(iii) 
q is increasing and q is convex in [ 0 , u * ] .
(iv) 
q is increasing in [ 0 , u * ] .
(v) 
N q is increasing in [ 0 , u * ] , N q ( u * ) = u * and N q ( u * ) = 0 .
(vi) 
The function
h q ( t ) = q ( t ) q ( t ) q ( t ) 2
is positive in [ 0 , u * ) and h q ( u * ) = 0 .
(vii) 
0 u m v m u m + 1 < u * and lim m u m = u * .

3. Semilocal Convergence I

The following conditions relating the parameters L, M, d to NLMM have been used in the semilocal convergence.
Suppose:
(H1)
There exist an initial guess x 0 D and a parameter d 0 such that F ( x 0 ) 1 L ( B 2 , B 1 ) and F ( x 0 ) 1 F ( x 0 ) d .
(H2)
F ( x 0 ) 1 F ( x 0 ) M .
(H3)
F ( x 0 ) 1 ( F ( y ) F ( x ) ) K 1 x y for some parameter K 1 > 0 and each x , y D .
(H4)
d λ ( K 1 , M ) .
and
(H5)
U [ x 0 , u * ] D .
The following semilocal convergence result was shown in [2,10].
Theorem 1.
Suppose that the conditions ( H 1 ) ( H 5 ) and
(H6)
U [ x 0 , u * ] Ω
for L = K 1 hold. Then, the following assertions hold:
(1)
F ( x 0 ) 1 F ( x 0 ) q ( 0 ) 1 q ( 0 ) .
(2)
F ( x 0 ) 1 F ( x ) q ( u ) for each x Ω such that x x 0 u u * .
(3)
F ( x 0 ) 1 ( F ( y ) F ( x ) ) | q ( v ) q ( u ) | for each u , v [ 0 , u * ] such that y x | v u | .
(4)
F ( x 0 ) 1 F ( x m ) q ( u m ) .
(5)
F ( x m ) 1 F ( x 0 ) q ( t m ) 1 .
(6)
y m x m v m u m .
(7)
F x m + y m 2 1 F ( x 0 ) q u m + s m 2 1 .
(8)
x m + 1 x m v m + 1 v m .
(9)
x m + 1 y m v m + 1 u m .
(10)
The sequence { x m } generated by NLMM is well defined in the ball U [ x 0 , u * ] , remains in U [ x 0 , u * ] and converges to the only solution x * of the equation F ( x ) = 0 in U [ x 0 , u * ] .
(11)
Moreover, the following error estimates hold:
x * x m u * u m
and
lim m + x m + 1 x * x m x * 1 + 2 = 1 2 F ( x * ) 1 F ( x * ) 2 ( 1 + 2 ) 2 + 2 .
Next, the preceding results are extended without additional conditions. Suppose:
  • ( H 3 ) F ( x 0 ) 1 ( F ( x ) F ( x 0 ) ) K 0 x x 0 for some K 0 > 0 and each x D .
Define the parameter
r = 2 M + M 2 + 2 K 0
and the region
D 0 = U ( x 0 , r ) D .
  • ( H 3 ) F ( x 0 ) 1 ( F ( y ) F ( x ) ) K y x for some parameter K > 0 and each x , y D 0 .
Clearly, we have
K 0 K 1
and
K K 1 ,
since
D 0 D .
It is assumed without loss of generality
K 0 K .
Otherwise, the results that follow hold with K 0 replacing K. Notice also that the computation of the parameter K 1 requires the computation of K 0 , K and K 1 = K 1 ( D ) , K 0 = K 0 ( D ) , but K = K ( D 0 , D ) . Hence, no additional conditions are required to develop the results that follow.
Let us consider the cubic polynomials
q 0 ( t ) = q ( K 0 , L , d ) = K 0 6 t 3 + M 2 t 2 t + d
and
q 1 ( t ) = q ( K , L , d ) = K 6 t 3 M 2 t 2 t + d .
It follows by (4), (12), (13), (16) and (17) that
q 0 ( t ) q ( t ) ,
q 0 ( t ) q ( t ) ,
q 1 ( t ) q ( t ) ,
q 1 ( t ) q ( t ) ,
q 0 ( t ) q 1 ( t ) ,
and
q 0 ( t ) q 1 ( t )
for each t 0 .
Suppose that
  • ( H 4 ) d λ ( K , M ) .
The following auxiliary result is needed.
Lemma 2.
Suppose that the condition ( H 4 ) holds. Then, the conclusions of Lemma 1 with K, q 1 replace L and q, respectively.
Proof. 
Simply replace L , q , u * , u * * by K , q 1 , u ¯ * , u ¯ * * , respectively, in the proof of Lemma 1. □
Remark 1.
It follows by ( i ) of Lemma 1 and ( H 4 ) that the polynomial q 1 has two zeros u ¯ * , u ¯ * * with 0 < u ¯ * u ¯ * * . Moreover, if ( H 4 ) holds, then by (13) and (20),
u ¯ * u *
and
u ¯ * * u * * ,
since q 1 ( u * ) q ( u * ) = 0 , q 1 ( 0 ) = d > 0 , and q 1 ( u * * ) q ( u * * ) = 0 .
Notice also that:
(a)
The parameter r is the unique positive zero of the equation
q 0 ( t ) 1 = 0
and
q 0 ( t ) = q 1 ( t ) for each t [ 0 , r ] .
(b)
( H 4 ) ( H 4 ) ,
but not necessarily vice versa unless if K 1 = K .
Hence, we arrived at the following extension of the Theorem 1.
Theorem 2.
Suppose that the conditions ( H 1 ) , ( H 2 ) , ( H 3 ) , ( H 3 ) , ( H 4 ) and
  • ( H 5 )   U [ x 0 , u ¯ * ] D
hold. Then, the assertions ( 1 ) ( 11 ) of Theorem 1 hold with K, q 1 , u ¯ * , { u ¯ m } , { v ¯ m } replacing K 1 , q, u * , { u m } , { v m } , respectively, where
u ¯ 0 = 0 , v ¯ 0 = u ¯ 0 q 1 ( u ¯ 0 ) 1 q 1 ( u 0 ) , v ¯ m = u ¯ m q 1 u m 1 + v m 1 2 1 q ( u ¯ m )
and
u ¯ m + 1 = u ¯ m q 1 u m + v m 2 1 q 1 ( u ¯ m ) .
Proof. 
The assertions ( 1 ) ( 6 ) follow with the above changes. Concerning the assertion ( 7 ) , set z 0 = x 0 + y 0 2 , μ 0 = u ¯ 0 + v ¯ 0 2 and F 1 ( x ) = F ( x 0 ) 1 F ( x ) for each x D . Then, we have in turn that
F 1 ( z 0 ) = F 1 ( x 0 ) + 0 1 F 1 ( x 0 + θ ( z 0 x 0 ) ) ( z 0 x 0 , · ) d θ ,
leading to
F 1 ( z 0 ) I 0 1 F 1 ( x 0 + θ ( z 0 x 0 ) ) ( z 0 x 0 , · ) d θ F 1 ( x 0 ) ( z 0 x 0 ) + 0 1 [ F 1 ( x 0 + θ ( z 0 x 0 ) ) F 1 ( x 0 ) ] ( z 0 x 0 , · ) d θ F 1 ( x 0 ) z 0 x 0 + 0 1 K θ ( μ 0 u ¯ 0 ) z 0 x 0 d θ .
Notice that the last inequality in (29) follows from ( H 3 ) , F 1 ( x ) = F ( x 0 ) 1 F ( x ) and q ( t ) = K t + M .
Then, from ( H 2 ) and the definition of y 0 and u ¯ 0
F 1 ( z 0 ) I q 0 ( u ¯ 0 ) ( μ 0 u ¯ 0 ) + K 2 ( μ 0 u ¯ 0 ) 2 q 0 ( μ 0 ) q 0 ( u ¯ 0 ) = 1 + q 0 ( μ 0 ) < 1 .
The condition ( H 5 ) in Theorem 2 can be replaced by
  • ( H 5 )   U [ x 0 , r ] D ,
where r is given by (10).
Moreover, the uniqueness ball can be enlarged from U ( x 0 , u * ) given in Theorem 1 to U ( x 0 , r ) . This can be seen using the weaker condition ( H 3 ) instead of ( H 3 ) used in Theorem 1 in [10] or Theorem 2 used by us. Indeed, in Theorem 1, the estimate was obtained for y * U ( x 0 , u * ) ,
0 1 F ( x 0 ) 1 F ( y * + θ ( x * y * ) ) d θ I 0 1 ( q ( y * + θ ( x * y * ) x 0 ) q ( 0 ) ) d θ < 1 ,
since q ( t ) < 0 for all t [ 0 , 1 ] , leading to
0 = F ( y * ) F ( x * ) = x * y * F ( v ) d v = 0 1 F ( y * + t ( x * y * ) ) d t ( y * x * ) ,
and consequently, y * = x * . However, the same estimate is obtained using the tighter condition ( H 3 ) with q 1 , u ¯ * replacing q, u * , respectively.
  • ( c ) The Lipschitz constant K can be replaced by an at least as small.
It follows by the Banach lemma on linear operators with inverses [3,5] and (30) that
F ( z 0 ) 1 F ( x 0 ) 1 1 F ( z 0 ) I 1 q 0 ( μ 0 ) 1 q 1 ( μ 0 ) ,
showing the assertion ( 7 ) . Notice that in Theorem 1, the less tight estimate than (31) is shown under the stronger and not actually needed condition ( H 3 ) , which is
F ( z 0 ) 1 F ( x 0 ) 1 q ( μ 0 ) .
In view of the estimate (31), the rest of the proof follows as in [10]. □
Remark 2. ( a ) In view of (24) and (28), Theorem 2 extends Theorem 1 with advantages already stated.
Define the ball U ( y 0 , r d ) for d < r . Notice that U ( y 0 , r d ) U ( x 0 , r ) . Then, suppose
  • ( H 3 )   F ( x 0 ) 1 ( F ( y ) F ( x ) ) K 2 y x for some K 2 > 0 and each y , x D 1 = U ( y 0 , r d ) D .
It follows that ( H 3 ) , K 2 can replace ( H 3 ) , K, respectively, in our results and
K 2 K .
The iterates { x m } U ( y 0 , r d ) .

4. Semilocal Convergence II

The convergence conditions of the previous section may not hold, even if the method (3) converges. As a motivational example, consider the function f defined on the interval D 2 = [ 0.5 , 1.5 ] by
f ( t ) = t 2 log t + t 4 t 3 , t 0 0 , t = 0 .
Then, clearly the function f is unbounded on D 2 . Hence, the results of the previous section cannot guarantee the convergence of method (3) to the solution x * = 1 D 2 . That is why we drop the conditions ( H 2 ) - ( H 5 ) , ( H 3 ) , ( H 3 ) , ( H 4 ) and ( H 5 ) and utilize the more general ω -continuity conditions, the first derivative that actually only appears on the method (3) and majorizing sequences to present another semilocal convergence result under weaker conditions.
Let h 0 : [ 0 , + ) R be a continuous and nondecreasing function. Suppose that the equation
h 0 ( t ) 1 = 0
has a smallest positive solution s. Moreover, suppose that there exists a function h : [ 0 , s ) R , which is continuous and nondecreasing. Let also parameters t 0 , s 0 , t 1 be such that t 0 = 0 , s 0 > 0 and s 0 < t 1 . Define the sequences { t m } , { s m } by
p m = h 0 t m + s m 2 , t m + 1 = s m + h s m s m 1 + t m t m 1 2 1 p m ( s m t m ) , α m + 1 = 0 1 h s m t m 2 + θ ( t m + 1 t m ) d θ ( t m + 1 t m ) s m + 1 = t m + 1 + α m + 1 1 p m .
This sequence shall be shown to be majorizing for the method (3). However, a convergence result is developed first.
Lemma 3.
Suppose that for each m = 0 , 1 , 2 , and some β > 0
p m < 1 and t m β
Then, the following items hold:
0 t m s m t m + 1 β
and
lim m t m = α β .
Proof. 
It follows from the definition of the sequence { t m } given by Formula (33) and the condition (34) that items (35) and (36) hold, the lim m t m exists, satisfying (36), where α is the unique least upper bound of this sequence. □
Remark 3.
A possible choice for β is any number in the interval ( 0 , s ] .
The conditions connecting the “h” functions to the operators F and F are:
(A1)
There exists an initial guess x 0 D , s 0 0 , t 1 s 0 such that F ( x 0 ) 1 L ( B 2 , B 1 ) , F ( x 0 ) 1 F ( x 0 ) s 0 , and for y 0 = x 0 F ( x 0 ) 1 F ( x 0 ) , F x 0 + y 0 2 1 L ( B 2 , B 1 ) with
F x 0 + y 0 2 1 F ( x 0 ) F ( x 0 ) 1 F ( x 0 ) t 1 s 0 .
(A2)
F ( x 0 ) 1 ( F ( x ) F ( x 0 ) ) h 0 ( x x 0 ) for each x D .
Set
D 3 = U ( x 0 , s ) D .
(A3)
F ( x 0 ) 1 ( F ( y ) F ( x ) ) h ( y x ) for each x , y D 3 .
(A4)
The condition (34) holds
and
(A5)
U [ x 0 , α ] D .
An Ostrowski-like representation [17] is needed for the iterates { x m } and { y m } .
Lemma 4.
Suppose that the iterates { x m } , { y m } exist for each m = 1 , 2 , . Then, the following items hold
x m + 1 y m = F x m + y m 2 1 F x m 1 + y m 1 2 F x m + y m 2 ( y m x m )
and
F ( x m + 1 ) = 0 1 F ( x m + θ ( x m + 1 x m ) ) d θ F x m + y m 2 ( x m + 1 x m ) .
Proof. 
By subtracting the first substep of the method from the second substep, we obtain in turn that
x m + 1 y m = F x m 1 + y m 1 2 1 F x m + y m 2 1 F ( x m ) = F x m + y m 2 1 F x m 1 + y m 1 2 1 F ( x m ) = F x m + y m 2 1 F x m 1 + y m 1 2 F x m + y m 2 F x m 1 + y m 1 2 1 F ( x m ) = F x m + y m 2 1 F x m 1 + y m 1 2 F x m + y m 2 ( y m x m )
showing the estimate (34). The estimate (38) follows from the identity
F ( x m + 1 ) = F ( x m + 1 ) F ( x m ) F x m + y m 2 ( x m + 1 x m ) ,
which is obtained by the first substep of NLMM. □
Next, the semilocal convergence is developed for the method (3).
Theorem 3.
Suppose that the conditions ( A 1 ) ( A 5 ) hold. Then, the sequence { x m } converges to a solution x * U [ x 0 , α ] such that
x m x * α t m .
Proof. 
Mathematical induction is applied to show
y m x m s m t m < α
and
x m + 1 y m t m + 1 s m .
The condition ( A 1 ) and the Formula (4) (for m = 0 ) imply that the estimate (39) holds for m = 0 . Then, we also have the iterate y 0 U ( x 0 , α ) . Moreover, x 0 + y 0 2 x 0 = 1 2 y 0 x 0 < α , so x 0 + y 0 2 U ( x 0 , α ) . Let v U ( x 0 , α ) be an arbitrary point. Then, the condition ( A 2 ) gives
F ( x 0 ) 1 ( F ( v ) F ( x 0 ) ) h 0 ( v x 0 ) h 0 ( α ) < 1 .
That is, F ( v ) 1 L ( B 2 , B 1 ) and
F ( v ) 1 F ( x 0 ) 1 1 h 0 ( v x 0 ) .
In particular, if v = x 0 + y 0 2 , then the iterate y 1 and x 1 are well defined by the method (3). Moreover, the last condition in ( A 1 ) and (33) give
x 1 y 0 = F x 0 + y 0 2 1 F ( x 0 ) F ( x 0 ) 1 F ( x 0 ) t 1 s 0 .
Thus, the assertion (40) holds for m = 0 . Suppose that the assertions (39) and (40) held for all integers smaller than n 1 . Then, we obtain from
F ( x 0 ) 1 F x m + y m 2 F ( x 0 ) h 0 x m + y m 2 x 0 h 0 x m x 0 + y m x 0 2 p m < 1 ,
thus,
F x m + y m 2 1 F ( x 0 ) 1 1 p m .
Then, by (37), we obtain in turn that
x m + 1 y m = F x m 1 + y m 1 2 1 F x m + y m 2 1 F ( x m ) = F x m + y m 2 1 F x m 1 + y m 1 2 F x m + y m 2 F x m 1 + y m 1 2 1 F ( x m ) = F x m + y m 2 1 F x m 1 + y m 1 2 F x m + y m 2 ( y m x m ) .
It follows from (33), ( A 3 ) , (42), (43) and the induction hypotheses that
x m + 1 y m h y m y m 1 + x m x m 1 2 y m x m 1 p m h s m s m 1 + t m t m 1 2 ( s m t m ) 1 p m = t m + 1 s m
and
x m + 1 x 0 x m + 1 y m + y m x 0 t m + 1 s m + s m t 0 = t m + 1 < α .
Hence, (40) holds, and the iterate x m + 1 U ( x 0 , α ) . Furthermore, by (38) and the second substep of the method (3), we have in turn that
y m + 1 x m + 1 F ( x m + 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x m + 1 ) 0 1 h x m + θ ( x m + 1 x m ) x m + y m 2 d θ x m + 1 x m 1 p m 0 1 h s m t m 2 + θ ( t m + 1 t m ) d θ ( t m + 1 t m ) 1 p m = s m + 1 t m + 1
and
y m + 1 x 0 y m + 1 x m + 1 + x m + 1 x 0 s m + 1 t m + 1 + t m + 1 t 0 = s m + 1 < α .
Thus, the induction for the estimates (39) and (40) is terminated.
However, the sequence { t m } is Cauchy as convergent by Lemma 3. Therefore, the sequence { x m } is also Cauchy in a Banach space B 1 , and as such, it converges to some x * U [ x 0 , α ] , since this set is closed. Furthermore, by using the continuity of F and letting n + in the calculation,
F ( x 0 ) 1 F ( x m + 1 ) α m + 1 ,
we conclude that F ( x * ) = 0 . □
Concerning the uniqueness of the solution in a neighborhood about the point x 0 , we have:
Proposition 1.
Suppose:
(a)
There exists a solution z * U ( x 0 , ϑ ) of the equation F ( x ) = 0 for some ϑ > 0 .
(b)
The condition ( A 2 ) holds on the ball U ( x 0 , ϑ )
and
(c)
There exists ϑ ¯ ϑ such that
0 1 h 0 ( θ ϑ ¯ ) d θ < 1 .
Set D 4 = U [ x 0 , ϑ ¯ ] D . Then, the equation F ( x ) = 0 is uniquely solvable by z * in the region D 4 .
Proof. 
Let v * D 4 with F ( v * ) = 0 . Define the linear operator T = 0 1 F ( z * + θ ( v * z * ) ) d θ . Then, it follows by condition ( A 2 ) and (47) that
F ( x 0 ) 1 ( T F ( x 0 ) ) 0 1 h 0 ( z * x 0 + θ ( z * v * ) d θ 0 1 h 0 ( ( 1 θ ) ϑ + θ ϑ ¯ ) d θ < 1 ,
which implies v * = z * . □
Remark 4.
(1) If all the conditions of Theorem 3 hold, then we can set ϑ = α .
(2)
The condition ( A 5 ) can be replaced by
  • ( A 5 )   U [ x 0 , α ] D .
(3)
Suppose that d < s . Define the set D 5 = U ( y 0 , s d ) D . Notice that D 5 U ( x 0 , s ) . Then, a tighter function h is obtained if D 5 replaces D 3 in the condition ( A 2 ) .

5. Local Convergence

We shall introduce some scalar functions and some parameters to show the local convergence analysis of NLMM.
Suppose:
(i)
There exists a function ψ 1 : [ 0 , + ) R , which is nondecreasing and continuous and a parameter M 2 0 such that the function ψ 2 ( t ) 1 has a smallest zero ρ 2 ( 0 , + ) , where ψ 2 : [ 0 , + ) R is defined by
ψ 2 ( t ) = 0 1 ψ 1 ( θ t ) d θ + M 2 t .
(ii)
The function ψ 3 : [ 0 , ρ 2 ) R is such that
ψ 3 ( t ) 1
has a smallest zero r 3 ( 0 , ρ 2 ) , where
ψ 3 ( t ) = 0 1 0 1 ψ 1 ( ( ( 1 τ ) + θ τ ) t ) 1 2 + | 1 2 θ | d τ d θ + 1 2 M 2 t 1 ψ 2 ( t )
The parameter r 3 shall be shown to be a radius of convergence for NLMM.
The convergence conditions are:
(l1)
There exists a simple solution x * Ω of the equation F ( x ) = 0 and a parameter M 2 0 such that F ( x * ) 1 F ( x * ) M 2 .
(l2)
F ( x * ) 1 ( F ( x ) F ( x * ) ) ψ 1 ( x x * ) for each x Ω .
(l3)
U [ x * , r 3 ] Ω .
Next, the local convergence is given for NLMM.
Theorem 4.
Suppose that the conditions ( l 1 ) ( l 3 ) hold. Then, the sequence { x m } generated by NLMM converges to x * provided that x 0 U ( x * , r 3 ) .
Proof. 
Let z U ( x * , r 3 ) . It follows by the conditions ( l 1 ) , ( l 2 ) and the definition of the radius r 3 in turn that
F ( x * ) 1 ( F ( z ) F ( x * ) ) = 0 1 F ( x * ) 1 F ( x * + θ ( z x * ) ) d θ ( z x * ) 0 1 F ( x * ) 1 ( F ( x * + θ ( z x * ) F ( x * ) ) d θ ( z x * ) + F ( x * ) 1 F ( x * ) ( z x * ) 0 1 ψ 1 ( θ z x * ) d θ z x * + M 2 z x * = ψ 2 ( z x * ) < 1 ;
thus, F ( z ) 1 L ( B 2 , B 1 ) and
F ( z ) 1 F ( x * ) 1 1 ψ 2 ( z x * ) .
By hypothesis x 0 U ( x * , r 3 ) and (49) for z = x 0 , we have that F ( x 0 ) 1 L ( B 2 , B 1 ) and the iterate y 0 is well defined by the first substep of NLMM.
Moreover, we can write for y θ = x * + θ ( x 0 x * )
y 0 x * = F ( x 0 ) 1 0 1 F ( y θ ) d θ F ( x 0 ) ( y θ x * ) = F ( x 0 ) 1 0 1 0 1 F ( x 0 + τ ( y θ x 0 ) ) d θ d τ ( y θ x * ) = F ( x 0 ) 1 0 1 0 1 F ( x 0 + τ ( y θ x 0 ) ) F ( x * ) d θ d τ ( y θ x * ) F ( x 0 ) 1 0 1 0 1 F ( x * ) d θ d τ ( y θ x * ) .
By applying the condition ( l 2 ) and (49) (for z = x 0 ) on (50), we obtain in turn that
y 0 x * 0 1 0 1 ψ 1 ( ( 1 τ ) + θ τ ) x 0 x * θ d θ d τ x 0 x * + 1 2 M 2 x 0 x * 1 ψ 2 ( x 0 x * ) ψ 3 ( x 0 x * ) x 0 x * x 0 x * < r 3 ,
where we used that
x 0 x * + τ ( y θ x 0 ) = ( ( 1 τ ) + θ τ ) ( x 0 x * )
and θ 1 2 + 1 2 θ . Thus, the iterate y 0 U ( x * , r 3 ) . Then, notice that
x 0 + y 0 2 x * 1 2 ( x 0 x * + y 0 x * ) < r 3 ,
so the point z 0 = x 0 + y 0 2 U ( x * , r 3 ) .
Suppose that z m = x m + y m 2 U ( x * , r 3 ) . Then, we also have
z m x * 1 2 ( x m x * + y m x * ) < r 3 ,
so the iterates x m + 1 and y m are well defined by NLMM. Then, we can write for y θ m = x * + θ ( x m x * ) and S m = F x m + y m 2
x m + 1 x * = S m 1 0 1 F ( y θ ) d θ S m ( x m x * )
However, the expression in the bracket can be written as
0 1 F ( y θ ) d θ S m = 0 1 0 1 F ( z m + τ ( y θ m z m ) ) d θ d τ ( y θ m x * ) = 0 1 0 1 F ( z m + τ ( y θ m z m ) ) F ( x * ) d θ d τ ( y θ m x * ) + 0 1 0 1 F ( x * ) d θ d τ ( y θ m x * ) .
In view of (49) (for z = z m ), (52), (53), we obtain
x m + 1 x * q m x m x * 2 1 ψ 2 ( x m x * ) ψ 3 ( x m x * ) x m x * x m x * ,
where, for the numerator, we obtain
0 1 0 1 ψ 1 ( z m x * + τ ( y θ m z m ) ) ( y θ m z m ) d θ d τ x m x * + M 2 0 1 y θ m z m d θ x m x * 0 1 0 1 ψ 1 1 2 ( 1 τ ) ( x m x * + y m x * ) + τ θ x m x * * y m x * 2 + 1 2 θ x m x * d θ d τ x m x * + 1 2 M 2 0 1 1 2 + 1 2 θ d θ x m x * 0 1 0 1 ψ 1 ( ( ( 1 τ ) + τ θ ) x m x * ) 1 2 + 1 2 θ d θ d τ x m x * 2 + 1 2 M 2 x m x * 2 = q m x m x * 2 ,
where we also used y m x * x m x * .
Similarly, the estimate
y m x * = S m 1 1 0 1 F ( y θ m 1 ) S m 1 ( x m x * ) ,
we obtain as in (54)
y m x * q m 1 x m 1 x * x m x * 1 ψ 2 ( x m 1 x * ) ψ 3 ( x m 1 x * ) x m x * x m x * .
Hence, the iterates x m + 1 , y m U ( x * , r 3 ) for each m = 0 , 1 , 2 , . Then, from the estimates
x m + 1 x * ψ 3 ( x 0 x * ) x m x *
and
y m x * ψ 3 ( x 0 x * ) x m x * ,
we deduce that lim m x m = lim m y m = x * since ψ 3 ( x 0 x * ) [ 0 , 1 ) . □
The uniqueness of the solution ball is determined in the next result.
Proposition 2.
Suppose:
(i)
The condition ( l 1 ) holds.
(ii)
There exists a solution z * U ( x * , ρ 3 ) for some ρ 3 > 0 .
(iii)
The condition ( l 3 ) holds on the ball U ( x * , ρ 3 ) .
and
(iv)
There exists ρ 4 ρ 3 such that
0 1 0 1 ψ 1 ( τ θ ρ 4 ) θ ρ 4 d θ d ρ + 1 2 M 2 ρ 4 < 1 .
Set D 6 = U [ x * , ρ 4 ] Ω . Then, the only solution of the equation F ( x ) = 0 in the region D 6 is x * .
Proof. 
Let z * D 6 with F ( z * ) = 0 . Set z θ * = x * + θ ( z * x * ) and T = 0 1 F ( z θ * ) d θ . Then, we can write
0 1 ( F ( z θ * ) F ( x * ) ) d θ = 0 1 0 1 F ( x * + τ ( z θ * x * ) ) ( z θ * x * ) d θ d τ = 0 1 0 1 [ F ( x * + τ ( z θ * x * ) ) F ( x * ) ] θ ( z θ * x * ) d θ d τ + 0 1 0 1 F ( x * ) θ ( z θ * x * ) d θ d τ .
By composing (57) with F ( x * ) 1 , using ( l 2 ) and (56), we obtain in turn that
F ( x * ) 1 ( T F ( x * ) 0 1 0 1 ψ 1 ( τ θ z * x * ) θ z * x * d θ d τ + 1 2 M 2 z * x * 0 1 0 1 ψ 1 ( τ θ ρ 4 ) θ ρ 4 d θ d ρ + 1 2 M 2 ρ 4 < 1 ;
thus, we conclude again that z * = x * . □
Remark 5.
We can certainly choose ρ 3 = r 3 .

6. Numerical Examples

In this section, some numerical examples are solved in order to corroborate the theoretical results obtained and the efficacy of our approach.
Example 1.
Let B 1 = B 2 and Ω = U ( x 0 , 1 γ ) for some parameter γ ( 0 , 1 ) . Define the polynomial F on the interval Ω by
F ( v ) = v 4 4 + φ γ v for some φ R .
Choose x 0 = 1 . Then, if we substitute F on the “h” conditions, we see that the conditions
h 0 ( t ) = ( γ 2 5 γ + 7 ) 1 γ t and h ( t ) = 3 ( 1 + s ) 2 1 γ t
are verified provided that
d = | 1 4 γ + φ | 1 γ , K 0 = 3 ( 3 γ ) 1 γ , K 1 = 6 ( 2 γ ) 1 γ , K = 6 ( 1 + r ) 1 γ
for
r = 2 ( 1 γ ) 3 + 9 + 6 ( 3 γ ) ( 1 γ ) , and M = 3 1 γ .
Notice that K 0 < K 1 and K < K 1 . Moreover, λ ( K , M ) < λ ( K 1 , M ) .
For φ = 1 10 and γ ( 0.0175194 , 0.0301594 ) ( 0.230014 , 0.233713 ) ( 0 , 1 ) ,
d > λ ( K 1 , M ) .
and
d < λ ( K , M ) .
Thus, it is clear that our new condition ( H 4 ) holds true, but the condition ( H 4 ) used in [10] does not hold. By taking γ = 1 10 and φ = 1 10 , we obtain the following
β = 0.138249 , s 0 = 0.055555 , t 1 = 0.061053 , p 0 = 0.200926 , α 1 = 0.015373 , s 1 = 0.080291 , t 2 = 0.087582 , p 1 = 0.511197 , α 2 = 0.002621 , s 2 = 0.092946 , t 3 = 0.094254 , p 2 = 0.652915 , α 3 = 0.000173 , s 3 = 0.094754 , t 4 = 0.094782 , p 3 = 0.683581 , α 4 = 1.17314 e 6 , s 4 = 0.094786 , t 5 = 0.094786 , p 4 = 0.685610 , α 5 = 5.97079 e 11 , s 5 = 0.094786 , t 6 = 0.094786 , p 5 = 0.685623 ,
which shows that conditions ( A 1 ) - ( A 4 ) are satisfied. Hence, by Theorem 3, the sequence { x m } converges to a unique solution x * U ( x 0 , α ) where, α = 0.094786 . Thus, this example can be solved using the weaker condition used in our study but not using the earlier one [10].
Example 2.
Let B 1 = B 2 = R 2 . Define the mapping F : R 2 R 2 by
F ( v ) = ( 2 ϑ 1 1 9 ϑ 1 2 ϑ 2 , ϑ 1 + 2 ϑ 2 1 9 ϑ 2 2 ) T ,
where v = ( ϑ 1 , ϑ 2 ) T . We shall find a solution to the equation F ( v ) = 0 . The first and second-order Fréchet derivatives are calculated to be
F ( v ) = 2 2 9 ϑ 1 1 1 2 2 9 ϑ 2
and
F ( v ) = 2 9 0 0 0 | 0 0 0 2 9 .
Pick x 0 = ( 11.4 , 11.4 ) T . Then, it follows that d = 1.9826 , M = 0.47618 , and K 0 can be arbitrary. By setting K = 10 3 , we see that d < λ ( K , M ) . Moreover, the solution x * = ( 9 , 9 ) T is obtained after m = 3 iterations. Notice that it takes m = 5 iterations for NM but only three for NLMM to reach x * . Thus, this method requires fewer computations than that of Newton’s method.
Example 3.
Let B 1 = B 2 = R and D = U [ 0 , 1 ] . Define a function F on D by
F ( v ) = e v 1 .
Clearly, we have x * = 0 . Then, the conditions ( l 1 ) and ( l 2 ) hold if M 2 = 1 and ψ 1 ( t ) = ( e 1 ) t . Then, ψ 2 ( t ) = t 1 t 2 + e t 2 and ψ 3 ( t ) = ( 9 ( e 1 ) t + 8 ) t 8 ( ( e 1 ) t 2 + 2 t 2 ) .
The parameter ρ 2 = 0.64385 and radius of convergence r 3 = 0.435659 .
Example 4.
Let B 1 = B 2 = R 3 and D = U [ 0 , 1 ] . Define a mapping F on D by
F ( v ) = e 1 6 ϑ 1 3 + ϑ 1 , e ϑ 2 1 , ϑ 3 3 6 + ϑ 3 T ,
where v = ( ϑ 1 , ϑ 2 , ϑ 3 ) T . Clearly, the solution is x * = ( 0 , 0 , 0 ) T . It follows by the definition of the mapping F that the first two Fréchet derivatives are
F ( v ) = e 1 2 ϑ 1 2 + 1 0 0 0 e ϑ 2 0 0 0 ϑ 3 2 2 + 1
and
F ( v ) = ( e 1 ) ϑ 1 2 0 0 0 0 0 0 0 0 | 0 0 0 0 e ϑ 2 0 0 0 0 | 0 0 0 0 0 0 0 0 ϑ 3 .
Notice that F ( x * ) = F ( x * ) 1 = d i a g ( 1 , 1 , 1 ) . Therefore, the conditions ( l 1 ) and ( l 2 ) are verified for M 2 = 1 and ψ 1 ( t ) = ( e 1 ) t . Then, ψ 2 ( t ) = t 1 t 2 + e t 2 and ψ 3 ( t ) = ( 9 ( e 1 ) t + 8 ) t 8 ( ( e 1 ) t 2 + 2 t 2 ) .
The parameter ρ 2 = 0.64385 and radius of convergence r 3 = 0.435659 .
The numerical examples were simulated by using Mathematica 8 on Intel(R) Core(TM) i5-8250U CPU @ 1.60 GHz 1.80 GHz, with 8 GB of RAM running on Windows 10 Pro version 2017. This kind of local and semilocal convergence demonstrates that the guarantee the existence and uniqueness of the solution are especially valuable in processes where it is difficult to prove the existence of solutions.

7. Conclusions

The present study deals with new local and semilocal convergence results for the Newton-like midpoint method under improved initial conditions. In the first type of semilocal convergence, the previous results are extended without using any additional postulates. The estimate obtained in [10] is less tight and uses stronger conditions in comparison to our results. The second semilocal convergence utilizes more general ω -continuity conditions and can be applied to the problems where earlier conditions fail. Notice also that the condition on F is dropped (see also the Example 1). Both semilocal convergence results are computationally verifiable and improve the previous study [10] in several directions, which are of practical importance. The local convergence theorem not given in [10] is established for the existence-uniqueness of the solution. We present varied numerical examples to show the applicability of our results. The innovation demonstrated that NLMM can also be used to extend the applicability of other methods requiring the inversion of a linear operator in an analogous way since our technique is method-free. This is the future area of research.

Author Contributions

The authors contributed equally. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Argyros, I.K.; Magreñán, Á.A. Ball convergence theorems and the convergence planes of an iterative methods for nonlinear equations. SEMA 2015, 71, 39–55. [Google Scholar]
  2. Gutiérrez, J.M. A new semilocal convergence for Newton’s method. J. Comput. Appl. Math. 1997, 79, 131–145. [Google Scholar] [CrossRef] [Green Version]
  3. Kantorovich, L.V. On Newton’s method for functional equations (russian). Dokl. Akad. Nauk. SSSR 2009, 59, 1237–1240. [Google Scholar]
  4. Wu, Q.; Zhao, Y. Convergence analysis for a deformed Newton’s method with third-order in Banach space under γ-condition. Int. J. Comput. Math. 2009, 86, 441–450. [Google Scholar]
  5. Ezquerro, J.; Hernández, M. New iterations of R-order four with reduced computational cost. BIT Numer. Math. 2009, 49, 325–342. [Google Scholar] [CrossRef]
  6. Grau-Sánchez, M.; Noguera, M.; Gutiérrez, J. On some computational orders of convergence. Appl. Math. Lett. 2010, 23, 472–478. [Google Scholar] [CrossRef] [Green Version]
  7. Herceg, D.; Herceg, D.J. Means based modifications of Newton’s method for solving nonlinear equations. Appl. Math. Lett. 2013, 219, 6126–6133. [Google Scholar] [CrossRef]
  8. Parida, P.K.; Gupta, D.K. Recurrence relations for a Newton-like method in Banach spaces. J. Comput. Appl. Math. 2007, 206, 873–887. [Google Scholar] [CrossRef] [Green Version]
  9. Sharma, J.R.; Guha, R.K.; Sharma, R. An efficient fourth order weighted-Newton method for systems of nonlinear equations. Numer. Algorithms 2013, 62, 307–323. [Google Scholar] [CrossRef]
  10. Cárdenas, E.; Castro, R.; Sierra, W. On Newton-type midpoint method with high efficiency index. J. Math. Anal. Appl. 2020, 491, 124381. [Google Scholar] [CrossRef]
  11. Argyros, I.K.; Chen, D. The midpoint method for solving nonlinear operator equations in Banach space. Appl. Math. Lett. 1992, 5, 7–9. [Google Scholar]
  12. McDougall, T.J.; Wotherspoon, S.J. A simple modification of Newton’s method to achieve convergence of order 1 + 2 . Appl. Math. Lett. 2014, 29, 20–25. [Google Scholar] [CrossRef] [Green Version]
  13. Argyros, I.K.; Chen, J. On local convergence of a Newton-type method in Banach space. Int. J. Comput. Math. 2009, 86, 1366–1374. [Google Scholar] [CrossRef]
  14. Neta, B. A sixth order family of methods for nonlinear equations. Int. J. Comp. Math. 1979, 7, 157–161. [Google Scholar] [CrossRef]
  15. Ezquerro, J.A.; Hernández, M.A. Newton’s Method: An Updated Approach of Kantorovich’s Theory; Springer: Cham, Switzerland, 2018. [Google Scholar]
  16. Kou, J. A third-order modification of Newton method for systems of non-linear equations. Appl. Math. Comput. 2007, 191, 117–121. [Google Scholar] [CrossRef]
  17. Ostrowski, A.M. Solutions of Equations and Systems of Equations; Academic Press: New York, NY, USA; London, UK, 1966. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Argyros, I.K.; Deep, G.; Regmi, S. Extended Newton-like Midpoint Method for Solving Equations in Banach Space. Foundations 2023, 3, 82-98. https://doi.org/10.3390/foundations3010009

AMA Style

Argyros IK, Deep G, Regmi S. Extended Newton-like Midpoint Method for Solving Equations in Banach Space. Foundations. 2023; 3(1):82-98. https://doi.org/10.3390/foundations3010009

Chicago/Turabian Style

Argyros, Ioannis K., Gagan Deep, and Samundra Regmi. 2023. "Extended Newton-like Midpoint Method for Solving Equations in Banach Space" Foundations 3, no. 1: 82-98. https://doi.org/10.3390/foundations3010009

APA Style

Argyros, I. K., Deep, G., & Regmi, S. (2023). Extended Newton-like Midpoint Method for Solving Equations in Banach Space. Foundations, 3(1), 82-98. https://doi.org/10.3390/foundations3010009

Article Metrics

Back to TopTop