Next Article in Journal
Generalized Bounded Turning Functions Connected with Gregory Coefficients
Next Article in Special Issue
An Accelerated Dual-Integral Structure Zeroing Neural Network Resistant to Linear Noise for Dynamic Complex Matrix Inversion
Previous Article in Journal
An Intuitionistic Fuzzy Multi-Criteria Approach for Prioritizing Failures That Cause Overproduction: A Case Study in Process Manufacturing
Previous Article in Special Issue
New Simplified High-Order Schemes for Solving SDEs with Markovian Switching Driven by Pure Jumps
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Kantorovich Theory for Nonsingular and Singular Equations

1
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505, USA
2
Department of Mathematical and Computational Sciences, National Institute of Technology Karnataka, Mangaluru 575 025, India
3
Department of Mathematics, University of Houston, Houston, TX 77205, USA
4
Department of Computer Sciences, University of Oklahoma, Norman, OK 73501, USA
*
Author to whom correspondence should be addressed.
Axioms 2024, 13(6), 358; https://doi.org/10.3390/axioms13060358
Submission received: 26 April 2024 / Revised: 21 May 2024 / Accepted: 21 May 2024 / Published: 28 May 2024
(This article belongs to the Special Issue Differential Equations and Inverse Problems)

Abstract

:
We develop a new Kantorovich-like convergence analysis of Newton-type methods to solve nonsingular and singular nonlinear equations in Banach spaces. The outer or generalized inverses are exchanged by a finite sum of linear operators making the implementation of these methods easier than in earlier studies. The analysis uses relaxed generalized continuity of the derivatives of operators involved required to control the derivative and on real majorizing sequences. The same approach can also be implemented on other iterative methods with inverses. The examples complement the theory by verifying the convergence conditions and demonstrating the performance of the methods.
MSC:
65J15; 65H10; 90C30; 90C53; 49M15

1. Introduction

Let T 1 , T 2 denote Banach spaces, and let B ( T 1 , T 2 ) be the space of linear and continuous operators from T 1 to T 2 . Newton-type methods (NTMs) [1]
x n + 1 = x n E ( x n ) # Υ ( x n ) , n = 0 , 1 , 2 ,
have been used to solve the equation
D Υ ( x ) = 0 .
Here, the operator Υ : Ω T 1 T 2 is a differentiable operator in the Fréchet sense, D B ( T 2 , T 1 ) , E ( x n ) B ( T 1 , T 2 ) approximates Υ ( x n ) . Moreover, E ( x n ) # stands for an outer inverse (OI) of E ( x n ) , i.e., E ( x n ) # E ( x n ) E ( x n ) # = E ( x n ) # .
A plethora of applications in optimization such as penalization problems, minimax problems, and goal programming are formulated as (2) using Mathematical Modelling [2,3,4,5,6,7,8,9,10,11,12,13,14,15].
Method (2) specializes to the Gauss–Newton method (GNM) for solving nonlinear least squares problems, the generalized NTM for undetermined systems, and an NTM for ill-posed Equations [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36].
As an example of (1) and (2), let T 1 and T 2 stand for Hilbert spaces. Then, consider the task of finding a local minimum u ˜ of
min H ( u ) ,
where H ( u ) = 1 2 Υ ( u ) 2 . Then, the GNM is defined by
x n + 1 = x n Υ ( x n ) Υ ( x n )
to solve
Υ ( u ) Υ ( u ) = 0 ,
where Υ ( x n ) is the Moore–Penrose inverse [5,15,31], and Υ ( u ) is the adjoint of the linear operator Υ ( u ) (see also the Remark 1).
Ben-Israel [5,17] utilized the conditions
Υ ( u ) Υ ( u ˜ ) Υ ( u ˜ ) ( u u ˜ ) c 1 u u ˜ , Υ ( u ) + Υ ( u ˜ ) + Υ ( u ˜ ) c 2 u u ˜
and
c 1 Υ ( u ˜ ) + + c 2 < 1
for all u , u ˜ in a neighborhood of x 0 Ω . He also used these conditions with Υ ( u ) # [5,17]. These results are not semilocal since they require information about Υ ( u ˜ ) + or Υ ( u ˜ ) # . Moreover, if Υ ( x 0 ) 1 B ( T 2 , T 1 ) , they require conditions not required in the Kantorovich theory [1,16,33,37,38,39,40,41]. Later Deuflhard and Heindl [39], Haussler [29], and Yamamoto [1] gave Kantorovich-type theorems for the GNM like (4) using convergence conditions involving either OI of Moore–Penrose inverses:
Υ ( u ˜ ) # I Υ ( u ) Υ ( u ) # Υ ( u ) c ( u ) u u ˜ , c ( u ) c ¯ < 1 ,
for each u , u ˜ Ω . This condition is strong and does not hold in concrete examples (see Section 4 in [31]). A Kantorovich-like result with generalized inverses can be found in [42] without (6). However, it was assumed that T 1 and T 2 are finite-dimensional and T 2 = R ( E ( x 0 ) , where R ( D ) denotes the range of a linear operator D. Other drawbacks of the earlier works are that only properties of OI are used. That is B E B = B and the projectional properties of E B and B E . However, the stability and perturbation bounds for OI are not given for method (1). However, this was accomplished through the elegant work of Nashed and Chen [31]. This work reduces to the Kantorovich theory for (2) when E ( u ) # is replaced by E ( u ) 1 without additional conditions. Later works on the convergence analysis using Lipchit-type conditions, particularly for the Newton–Gauss method (5), can be found in [10,24] and the references therein.
Next, we address the problems with the implementation of method (1) which constitutes the motivation for this paper. Let Δ B ( T 1 , T 2 ) and let Δ # B ( T 2 , T 1 ) be an OI of Δ .
Suppose Δ 1 # ( u ) = I + Δ # Δ E ( u ) 1 Δ # is an OI. A criterion for Δ 1 # to be an OI is Δ # E ( u ) Δ 1 (see Lemma 2). Then, (2) becomes
x n + 1 = x n Δ 1 # ( x n ) Υ ( x n ) .
But the main problem with the implementation of (2) of (7) still remains. This problem requires the invertibility of Δ 2 , Δ 2 ( u ) = I + Δ # Δ E ( u ) . This inversion can be avoided. Let m be a fixed natural number. Define the operators
H = H ( u ) = Δ # Δ E ( u )
and
L = L m ( u ) = I + H + H 2 + + H m .
Then, we can consider the replacement of (7) given as
x n + 1 = x n L Δ # Υ ( x n ) .
By letting m + in the definition of L, we have that
E ( x n ) # = lim m + L m ( x m ) Δ # .
Thus, it is worth studying the convergence of (8) instead of (2), since we avoid the inversion of the operator Δ 2 . Let us provide examples of possible choices for the operator Δ . First, consider the case when the operator E = Υ ( u ) is invertible. Moreover, let T 1 = T 2 = R i , i is a positive integer, and J denotes the Jacobian of the operator F. Then, choose Δ = J ( x 0 ) in the semi-local convergence case of Δ = J ( x ) in the local case, where x Ω is assumed to be a solution of the equation Υ ( u ) = 0 . The selection Δ = J ( u ¯ ) has been used in [43,44], for u ¯ Ω . In the setting of a Banach space for E = Υ ( u ) , the operator Δ can be chosen to be Δ = Υ ( x n ) (semi-local case) or Δ = Υ ( x ) (local case). Numerous selections for Δ connected to OI or generalized inverses (GI) can also be found in [5,11,13,31] and the references therein. Other selections for Δ are also possible provided that they satisfy the convergence conditions ( C 5 ) and ( C 6 ) of Section 3. The convergence analysis relies on the relaxed generalized continuity used to control the derivative Υ and majorizing sequences for the iterates { x n } (see also Section 2). The results in this article specialize immediately to solve nonsingular equations if E ( u ) # is replaced by E ( u ) 1 .
The rest of the article provides the preliminaries in Section 2; the convergence of (8) is in Section 3; and the applications are in Section 4. The article’s concluding remarks appear in Section 5.

2. Preliminaries

We reproduce standard results on OI and GI to make the article as self-contained as possible. More properties can be found in [5,11,12,13,31]. Let Δ B ( T 1 , T 2 ) . An operator B B ( T 2 , T 1 ) is said to be an inner inverse (II) of Δ if Δ B Δ = Δ , and an OI of Δ if B Δ B = B . It is well known that II and bounded OI always exist. The zero is always an OI. So, we consider only nonzero outer inverses. Suppose the operator B is either an inverse or an OI of Δ . Then, Δ B and B Δ are linear idempotents (algebraic projectors). Suppose that B is an inverse of E, then for N ( E ¯ ) , N ( B Δ ) = N ( Δ ) and R ( Δ ) = R ( Δ B ) . Consequently, the following decompositions hold T 1 = N ( Δ ) + R ( B Δ ) and T 2 = R ( Δ ) + N ( Δ B ) . Since, B is an OI of Δ if and only if Δ is an inverse of B, it follows that T 1 = R ( B ) N ( B Δ ) and T 2 = N ( B ) R ( Δ B ) . If B is an inner and an outer inverse of Δ , then B is called a GI of Δ . Moreover, there exists a unique GI B = Δ P , Q + satisfying Δ B Δ = Δ ,   B Δ B = B , B Δ = I P and Δ B = Q , where P is a given projection of T 1 into N ( Δ ) and Q a given projection of T 2 into R ( Δ ) . In the special case when T 1 and T 2 are Hilbert spaces, and P , Q are orthogonal projections. Then, Δ P , Q + is the Moore–Penrose inverse of Δ .
We need the following six auxiliary Lemmas, the proof of which can be found in Section 2 in [31].
Lemma 1. 
Let Δ B ( T 1 , T 2 ) . If Δ # is a bounded OI of Δ . Then, the following assertions hold
T 1 = R ( Δ # ) N ( Δ # Δ )
and
T 2 = N ( Δ # ) R ( Δ Δ # ) .
Lemma 2 
(Banach Perturbation-Like Lemma). Let Δ B ( T 1 , T 2 ) and Δ # B ( T 2 , T 1 ) be an OI of Δ. Let also B B ( T 1 , T 2 ) be such that Δ # ( B Δ ) < 1 . Then, B # : = I + Δ # ( B Δ ) 1 Δ # is a bounded OI of B so that N ( B # ) = N ( Δ # ) , R ( B # ) = R ( Δ # ) ,
B # Δ # Δ # ( B Δ ) Δ # 1 Δ # ( B Δ ) Δ # ( B Δ ) Δ # 1 Δ # ( B Δ )
and
B # Δ 1 1 Δ # ( B Δ ) .
Lemma 3. 
Let Δ , B B ( T 1 , T 2 ) and let Δ # and B # B ( T 2 , T 1 ) be OI of Δ, and B, respectively. Then, B # ( I Δ Δ # ) = 0 N ( Δ # ) N ( B # ) .
Lemma 4. 
Let Δ B ( T 1 , T 2 ) . Suppose that T 1 and T 2 have topological decompositions T 1 = N ( Δ ) M , T 2 = R ( Δ ) S . Let Δ + ( = Δ M , S + ) stand for the GI of Δ connected to these decompositions. Let B satisfy Δ + ( B Δ ) < 1 , and I + ( B Δ ) Δ + B sends N ( Δ ) to R ( Δ ) . Then, the following assertions hold:
B + : = B R ( Δ + ) , N ( Δ + ) + exists ,
B + = Δ + ( I + B 0 Δ + ) 1 = ( I + Δ + B 0 ) 1 Δ + ,
R ( B + ) = R ( Δ + ) , N ( B + ) = N ( Δ + )
and
B + Δ 1 Δ + ( B Δ ) ,
where B 0 = B Δ .
Lemma 5. 
Let Δ B ( T 1 , T 2 ) and Δ + be the GI as given in the Lemma 4. Let B B ( T 2 , T 1 ) satisfy Δ # ( B Δ ) 1 and R ( B ) R ( Δ ) . Then, the conclusions of the Lemma 4 hold and R ( B ) = R ( Δ ) .
Lemma 6. 
Let Δ B ( T 1 , T 2 ) and let Δ + be bounded GI of Δ. Let B B ( T 2 , T 1 ) satisfy Δ + ( B Δ ) < 1 . Let B # : = I + Δ + ( B Δ ) 1 Δ + . Then, B # is a GI of B dim N ( B ) = dim N ( Δ ) , and codim R ( B ) = codim R ( Δ ) .
Define the parameter
r = Δ # ( Δ E ( x ) ) , x Ω .
We need some estimates.
Lemma 7. 
Let Δ B ( T 1 , T 2 ) and let Δ # B ( T 2 , T 1 ) be an OI of Δ. Let E ( x ) B ( T 1 , T 2 ) with r [ 0 , 1 2 ) . Then, Δ 1 # is a bounded OI of E ( x ) . Moreover, the following estimates hold.
I L r ( 1 r m ) 1 r = b < 1 ,
the operator L 1 B ( T 2 , T 1 ) and
L 1 1 1 b = b ¯ ,
where Δ 1 # , L are as defined in the introduction.
Proof. 
The operator Δ 1 # is a bounded OI of E ( x ) by Lemma 2 for B = E ( x ) . Moreover, we have in turn by the definition of L:
I L = H + H 2 + · + H m H + H 2 + + H m = H 1 H m 1 H r 1 r m 1 r = b < 1 ,
by the choice of r. It is followed by the Lemma 2 that H 1 B ( T 2 , T 1 ) and H 1 1 1 b = b ¯ . □

3. Semi-Local Convergence

The convergence of the method (8) is shown using scalar majorizing sequences.
Definition 1. 
Let { x n } be a sequence in T 1 . Then, real sequence { p n } satisfying
x n + 1 x n p n + 1 p n , n 0
is called a majorizing sequence for { x n } . If the sequence { p n } converges, then also { x n } converges, and for x = lim n x n and p = lim n p n , we have
x x n p p n .
Therefore, the convergence of the sequence { x n } relates to that of { p n } .
Let M = [ 0 , + ) .
Some conditions are required in the convergence of (8).
Suppose that
( C 1 )
There exists parameters κ 0 , r 0 , 1 2 , a point x 0 Ω ,   Δ B ( T 1 , T 2 ) having outer inverse Δ # such that L Δ # Υ ( x 0 ) κ .
( C 2 )
There exists function ϕ 0 : M M which is continuous and nondecreasing such that the equation ϕ 0 ( t ) 1 = 0 admits a smallest positive solution r 0 . Take M 0 = [ 0 , r 0 ) .
( C 3 )
There exists functions ϕ : M 0 M , ϕ 1 : M 0 M . Define the real sequence { s n } for s 0 = 0 , s 1 = κ as
s n + 1 = s n + e n ( s n s n 1 ) ,
where e n = ( 1 + r + + r m ) 0 1 ϕ ( 1 θ ) ( s n s n 1 ) d θ + ϕ 1 ( s n 1 ) + b ¯ r m + 1 ,   n = 0 , 1 , 2 , .
The sequence { s n } is proven to be majorizing for { x n } (see Theorem 1). However, some conditions for this sequence are needed first.
( C 4 )
There exists r [ κ , r 0 ) such that for all n = 0 , 1 , 2 , , s n r .
By this condition and (12) that 0 s n s n + 1 r and there exists s [ κ , r ] such that lim n + s n = s .
The functions φ 0 , φ and φ 1 connect to the operators on the method (8).
( C 5 )
Δ # Υ ( u ) Δ ϕ 0 ( u x 0 ) for all u Ω . Set Ω 0 = Ω E ( x 0 , r 0 ) , where E ( x 0 , r 0 ) = { x X : x x 0 < r 0 } . We shall also denote by E [ x 0 , r 0 ] the closure of E ( x 0 , r 0 ) .
( C 6 )
r = Δ # Δ E ( u ) < 1 2 , Δ # Υ ( u ˜ ) Υ ( u ) ϕ ( u ˜ u ) and Δ # Υ ( u ) E ( u ) ϕ 1 ( u x 0 ) for all u , u ˜ Ω 0 .
( C 7 )
The equation e ( t ) 1 = 0 has a smallest solution in ( 0 , s ] , where e ( t ) = ( 1 + r + + r m ) 0 1 ϕ ( 1 θ ) t d θ + ϕ 1 ( t ) + b ¯ r m + 1 . Denote such solution by s ¯ and
( C 8 )
E [ x 0 , s ] Ω .
Next, the convergence is established for (8).
Theorem 1. 
Suppose that the conditions ( C 1 ) ( C 8 ) hold. Then, the sequence { x n } produced by the method (8) converges to a unique solution x E [ x 0 , s ] x 0 + R ( Δ # ) of the equation Δ # Υ ( x 0 ) = 0 . Moreover, the following assertion holds
x x n s s n .
Proof. 
Mathematical induction on n shall establish the estimates
x n + 1 x n s n + 1 s n .
Since
x 1 x 0 = E ( x 0 ) # Υ ( x 0 ) κ = s 1 = s 1 s 0 < s ,
the assertion (13) holds if n = 0 and x 1 E ( x 0 , s ) . It follows by (9) (for x = x 1 ) and Lemma 7 that Δ 1 ( x 1 ) # is an outer inverse of E ( x 1 ) and Δ 1 ( x 1 ) # A b ¯ and N ( Δ 1 ( x 1 ) # ) = N ( Δ # ) . Suppose that for i = 1 , 2 , , n ; x i x i 1 s i s i 1 and N ( Δ 1 ( x n 1 ) # ) = N ( Δ # ) . Then, we have
x n x 0 x n x n 1 + + x 1 x 0 s n s n 1 + + s 1 s 0 = s n < s ,
and N ( Δ 1 ( x n ) # ) = N ( Δ 1 ( x n 1 ) # ) = N ( Δ # ) . Hence, by the Lemma 3, it follows that
Δ 1 ( x n ) # I E ( x n 1 ) Δ 1 ( x n 1 ) # = 0 .
Then, by the method (8)
Υ ( x n ) = Υ ( x n ) Υ ( x n 1 ) Δ L 1 ( x n x n 1 ) = Υ ( x n ) Υ ( x n 1 ) Υ ( x n 1 ) ( x n x n 1 ) + Υ ( x n 1 ) A n 1 ( x n x n 1 ) + A n 1 Δ L 1 ( x n x n 1 ) = Υ ( x n ) Υ ( x n 1 ) Υ ( x n 1 ) ( x n x n 1 ) + Υ ( x n ) A n 1 A n 1 ( x n x n 1 ) + A n 1 L Δ L 1 ( x n x n 1 ) .
But, we have by the definition of L
A n 1 L Δ = A n 1 ( I + H + + H m ) Δ = A n 1 Δ + ( Δ + Δ + A n 1 ) ( I + H + + H m ) = A n 1 Δ + Δ ( H + + H m ) ( Δ H ) ( H + + H m ) = A n 1 Δ + Δ H + Δ ( H 2 + + H m ) ( Δ A n 1 ) ( H + + H m ) = Δ ( H 2 + + H m ) Δ ( Δ A n 1 ) ( H + + H m ) .
So,
Δ # ( A n 1 L Δ ) = H m + 1 ,
where we also used A n 1 Δ + Δ H = 0 by the definition of H. Using the induction hypotheses and the conditions ( C 5 ) , ( C 6 ) , (14), (15), (11), we have in turn that
Δ # Υ ( x n ) 0 1 ϕ ( 1 θ ) x n x n 1 d θ x n x n 1 + ϕ 1 ( x n 1 x 0 ) x n x n 1 + b ¯ r m + 1 x n x n 1 0 1 ( 1 θ ) ( s n s n 1 ) d θ + ϕ 1 ( s n 1 ) + b ¯ r m + 1 × ( s n s n 1 ) .
Hence, by (8), Lemma 7 and (17)
x n + 1 x n ( 1 + r + + r m ) [ 0 1 ϕ ( 1 θ ) ( s n s n 1 ) d θ + ϕ 1 ( s n 1 ) + b ¯ r m + 1 ] ( s n s n 1 ) = s n + 1 s n
and
x n + 1 x 0 x n + 1 x n + x n x 0 s n + 1 s n + s n s 0 = s n + 1 < s .
The induction is completed. Thus, we have for any n
x n + 1 x n s n + 1 s n ,
x n x 0 s n s
Δ # E ( x n + 1 ) Δ < 1 ,
and
Δ 1 ( x n + 1 ) # : = I + Δ # E ( x n + 1 ) Δ 1 Δ #
is an OI of E ( x n + 1 ) . The sequence { s n } is complete as convergent and majorizes { x n } . So, the sequence { x n } is also complete in T 1 . Then, it is convergent to a x E [ x 0 , s ] . By the definition
Δ 1 ( x n ) # = I + Δ # E ( x n ) A 1 Δ , for all n
and
0 = lim n I + Δ # E ( x n ) A ( x n x n 1 ) = lim n Δ # Υ ( x n ) = Δ # Υ ( x ) .
Thus, x solves Δ # Υ ( x ) = 0 . Using Lemma 2, we have R ( Δ 1 ( x n ) # ) = R ( Δ # ) for all n = 0 , 1 , 2 , . So, x n + 1 x n = Δ 1 ( x n ) # Υ ( x n ) R ( Δ # ) and from Lemma 1, we obtain R ( Δ # ) = R ( Δ # Δ ) , so x n + 1 x n + R ( Δ # ) . Thus, x n x 0 + R ( Δ # ) for all n. Suppose that w E [ x 0 , s ] { x 0 + R ( Δ # ) } solves the equation Δ # Υ ( x ) = 0 . Then, we have w x R ( Δ # ) and Δ # E ( w x n ) = Δ # E ( w x 0 ) + Δ # E ( x n x 0 ) = w x n , for all n = 0 , 1 , 2 , . Then, (11), as in (16) and using ( C 7 )
w x n + 1 = w x n + Δ 1 ( x n ) # Υ ( x n ) Δ 1 ( x n ) # Υ ( w ) ( 1 + r + + r m ) [ 0 1 ϕ ( 1 θ ) w x n d θ + ϕ 1 ( x n x 0 ) + b ¯ r m + 1 ] w x n
d w x n d n + 1 x 0 w < b ¯ s ,
where d = ( 1 + r + + r m ) 0 1 ϕ ( 1 θ ) w x 0 d θ + ϕ 1 ( s ) + b ¯ r m + 1 [ 0 , 1 ) . Therefore, we conclude x = lim n + x n = w . Finally, from (14) and the triangle inequality, we obtain for j = 0 , 1 , 2 ,
x n + j x n x n + j x n + j 1 + x n + j 1 x n + j 2 + + x n + 1 x n s n + j s n + j 1 + s n + j 1 s n + j 2 + + s n + 1 s n = s n + j s n .
By letting j + in (19), we show the assertion (13). □
Remark 1.
(i) 
The results of the Theorem 1 specialize for the Newton method with OI defined by x n + 1 = x n Υ ( x n ) # Υ ( x n ) , for solving Equation (5). Simply, take E ( x ) = Υ ( x ) and ϕ 1 = 0 .
(ii) 
Under the conditions ( C 1 ) ( C 8 ) , further suppose that the operator I + Δ + Δ E ( x ) 1 E ( x ) sends N ( A ) to R ( A ) provided that for x Ω , the inverse of
I + Δ + E ( x ) Δ e x i s t s .
Then, by Lemma 4, E ( x n ) # : = I + Δ + E ( x n ) Δ 1 Δ + is a GI. Thus, the proof of Theorem 1 establishes the convergence of method (8) for GI.
(iii) 
By the Lemma 6, the condition (20) can be exchanged by rank ( E ( x ) ) r a n k ( E ( x 0 ) ) for Δ = E ( x 0 ) and if T 1 and T 2 are finite dimensional. In general Banach spaces, the condition (20) can be switched by the stronger R ( E ( x ) ) R ( E ( x 0 ) ) (for Δ = E ( x 0 ) ) (see Lemma 5) or by the conditions of the Lemma 6.

4. Examples

The example considers method (4) with Υ ( x ) + = Υ ( x ) 1 for the case Δ = I , which is independent of x 0 . It is also compared with method (8) for Δ = Υ ( x 0 ) . In this case, the methods (4) and (8) become Newton’s method
x n + 1 = x n Υ ( x n ) 1 Υ ( x ) ( 4 )
and
x n + 1 = x n L Υ ( x 0 ) 1 Υ ( x ) ( 8 ) ,
respectively.
We shall solve the system
F 1 ( u , v ) = u 0.1 sin u 0.3 cos v + 0.4 F 2 ( u , v ) = v 0.2 cos u + 0.1 sin v + 0.3 .
If F = ( F 1 , F 2 ) . Then, we can write
Υ ( z ) = 0 f o r z = ( u , v ) T .
Consequently, we obtain
Υ ( ( u , v ) ) = 1 0.1 cos ( u ) 0.3 sin ( v ) 0.2 sin ( u ) 0.1 cos ( v ) + 1 .
Example 1. 
Method (8′): Set m = 1 and Δ = I .  We have that
L 1 ( x ) = I + ( I Υ ( x ) ) , P 1 ( x ) = x ( I + ( I Υ ( x ) ) ) Υ ( x ) , x j + 1 = P 1 ( x j ) .
Example 2. 
Method (8′): Set, m = 2 and Δ = I .  It follows that
L 2 ( x ) = I + ( I Υ ( x ) ) + ( I Υ ( x ) ) 2 , P 2 ( x ) = x L 2 ( x ) Υ ( x ) , x j + 1 = P 2 ( x j ) .
Example 3. 
Method (8′): Set m = 3 and Δ = I .  Then, we have
L 3 ( x ) = I + ( I Υ ( x ) ) + ( I Υ ( x ) ) 2 + ( I Υ ( x ) ) 3 , P 3 ( x ) = x L 3 ( x ) Υ ( x ) , x j + 1 = P 3 ( x j ) .
Example 4. 
Method (8′): Set m = 4 and Δ = I .  Then, we have
L 4 ( x ) = I + ( I Υ ( x ) ) + ( I Υ ( x ) ) 2 + ( I Υ ( x ) ) 3 + ( I Υ ( x ) ) 4 , P 4 ( x ) = x L 4 ( x ) Υ ( x ) , x j + 1 = P 4 ( x j ) .
Example 5. 
Method (8′): Set m = 5 and Δ = I .  Then, we have
L 5 ( x ) = I + ( I Υ ( x ) ) + ( I Υ ( x ) ) 2 + ( I Υ ( x ) ) 3 + ( I Υ ( x ) ) 4 + ( I Υ ( x ) ) 5 , P 5 ( x ) = x L 5 ( x ) Υ ( x ) , x j + 1 = P 5 ( x j ) .
Example 6. 
Method (8′): Set m = 1 , 5 ¯ and Δ = Υ ( x 0 ) .  It follows that
x n + 1 = x n L Δ 1 Υ ( x n ) , H = Δ 1 ( Δ Υ ( x ) ) , L = I + j = 1 m H j .
Definition 2. 
Let { x n } be a sequence. Then, the computational order of convergence (COC) is for θ n = x n x [4]
h ¯ n = ln | θ n + 1 / θ n | ln | θ n / θ n 1 | .
Definition 3. 
Let { x n } be a sequence. Then, the approximate computational order of convergence (ACOC) is for θ ^ n = x n x n 1
h ^ n = ln | θ ^ n + 1 / θ ^ n | ln | θ ^ n / θ ^ n 1 | .
The Table 1, Table 2, Table 3, Table 4 and Table 5 demonstrate that the cheaper-to-implement method (8) is behaving the same as Newton’s method for a large enough m .

5. Conclusions

We developed a semi-local Kantorovich-like analysis of Newton-type methods for solving singular nonlinear operator equations using outer or generalized inverses. These methods do not use inverses as in earlier studies but a sum of operators. This sum converges to the inverse and makes the implementation of these methods easier than the ones using inverses. The analysis of the methods relies on the concept of generalized continuity for the operators involved and majorizing sequences. Examples complement the theory. Due to its generality, this article’s technique can be applied on other method with inverses along the same lines [6,14,19,32,39,43,45,46,47,48,49]. It is worth noting that the method (8) should be used for sufficiently small m . Otherwise, if m is very large, it may be as expensive to implement as method (4).

Author Contributions

Conceptualization, I.K.A., S.G., S.R. and M.I.A.; Methodology, I.K.A., S.G., S.R. and M.I.A.; Software, I.K.A., S.G., S.R. and M.I.A.; Validation, I.K.A., S.G., S.R. and M.I.A.; Formal analysis, I.K.A., S.G., S.R. and M.I.A.; Investigation, I.K.A., S.G., S.R. and M.I.A.; Resources, I.K.A., S.G., S.R. and M.I.A.; Data curation, I.K.A., S.G., S.R. and M.I.A.; Writing—original draft, I.K.A., S.G., S.R. and M.I.A.; Writing—review & editing, I.K.A., S.G., S.R. and M.I.A.; Visualization, I.K.A., S.G., S.R. and M.I.A.; Supervision, I.K.A., S.G., S.R. and M.I.A.; Project administration, I.K.A., S.G., S.R. and M.I.A.; Funding acquisition, I.K.A., S.G., S.R. and M.I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data is contained within the article.

Acknowledgments

We would like to express our sincere graduate to Mykhailo Havdiak.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yamamoto, T. A convergence theorem for Newton-like algorithms in Banach spaces. Numer. Math. 1987, 51, 545–557. [Google Scholar] [CrossRef]
  2. Adly, S.; Ngai, H.V.; Nguyen, V.V. Newton’s method for solving generalized equations: Kantorovich’s and Smale’s approaches. J. Math. Anal. Appl. 2016, 439, 396–418. [Google Scholar] [CrossRef]
  3. Artacho, F.J.A.; Dontchev, M.; Gaydu, A.L.; Geoffroy, M.H.; Veliov, V.M. Metric regularity of Newton’s iteration. SIAM J. Control Optim. 2011, 49, 339–362. [Google Scholar] [CrossRef]
  4. Argyros, I.K. The Theory and Applications of Iteration Methods with Applications, Engineering Series, 2nd ed.; CRC Press, Taylor and Francis Publ.: Boca Raton, FL, USA, 2022. [Google Scholar]
  5. Ben-Israel, A. On applications of generalized inverses in nonlinear analysis. In Theory and Application of Generalized Inverses of Matrices; Boullion, T.L., Odell, P.L., Eds.; Mathematics Series 4; Texas Technological College: Lubbock, TX, USA, 1968; pp. 183–202. [Google Scholar]
  6. Cibulka, R.; Dontchev, A.; Preininger, J.; Roubal, T.; Veliov, V. Kantorovich-Type Theorems for Generalized Equations, Research Report 2015–16; Vienna University of Technology: Vienna, Austria, 2015. [Google Scholar]
  7. Dokov, S.P.; Dontchev, A.L. Robinson’s Strong Regularity Implies Robust Local Convergence of Newton’s Method. In Optimal Control (Gainesville, FL, 1997), Applied Optimization; Kluwer: Dordrecht, The Netherlands, 1998; Volume 15, pp. 116–129. [Google Scholar]
  8. Deuflhard, P. Newton Algorithms for Nonlinear Problems: Affine Invariance and Adaptive Algorithms; Springer Series in Computational Mathematics; Springer: Berlin/Heidelberg, Germany, 2004; Volume 35. [Google Scholar]
  9. Kelley, C.T. Solving Nonlinear Equations with Iterative Methods, Solvers and Examples in Julia, Fundamentals of Algorithms; SIAM: Philadelphia, PA, USA, 2023. [Google Scholar]
  10. Li, C.; Ng, K.F. Convergence analysis of the Gauss–Newton method for convex inclusion and convex-composite optimization problems. J. Math. Anal. Appl. 2012, 389, 469–485. [Google Scholar] [CrossRef]
  11. Nashed, M.Z. On the perturbation theory for generalized inverse operators in Banach spaces. In Functional Analysis Methods in Numerical Analysis; Nashed, M.Z., Ed.; Lecture Notes in Mathematics; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 1979; Volume 701, pp. 180–195. [Google Scholar]
  12. Nashed, M.Z. Inner, outer, and generalized inverses in Banach and Hilbert spaces. Nemer. Funct. Anal. Optim. 1987, 9, 261–325. [Google Scholar] [CrossRef]
  13. Nashed, M.Z. A new approach to classification and regularization of ill-posed operator equations. In Inverse and Ill-Posed Problems; Engl, H.W., Groetsch, C.W., Eds.; Notes and reports in mathematics in science and engineering; Academic Press: New York, NY, USA, 1987; Volume 4, pp. 53–75. [Google Scholar]
  14. Pollock, S.; Rebholz, L.G. Anderson acceleration for contraction and noncontractive operators. IMA Numer. Anal. 2021, 41, 2841–2872. [Google Scholar] [CrossRef]
  15. Rockafellar, R.T. Convex Analysis; Princeton Mathematics Series; Princeton University Press: Princeton, NJ, USA, 1970; Volume 28. [Google Scholar]
  16. Argyros, I.K.; George, S. On a unified convergence analysis for Newton-type methods solving generalized equations with the Aubin property. J. Complex. 2024, 81, 101817. [Google Scholar] [CrossRef]
  17. Ben-Israel, A.; Greville, T.N.E. Generalized Inverses: Theory and Applications; John Wiley and Sons: New York, NY, USA, 1974. [Google Scholar]
  18. Bian, W.; Chen, X.; Kelley, C.T. Anderson acceleration for a class of nonsmooth fixed-point problems. SIAM J. Sci. Comp. 2021, 43, S1–S20. [Google Scholar] [CrossRef]
  19. Canes, E.; Kremlin, G.; Levitt, A. Convergence analysis of direct minimization and self-consistent iterations. SIAM J. Sci. Comp. 2021, 42, 243–274. [Google Scholar]
  20. Dontchev, A.L. Local analysis of a Newton-type method based on partial linearization. In The Mathematics of Numerical Analysis (Park City, UT, 1995), Lectures in Applied Mathematics; AMS: Providence, RI, USA, 1996; Volume 32, pp. 295–306. [Google Scholar]
  21. Dontchev, A.L.; Rockafellar, R.T. Newton’s method for generalized equations: Asequential implicit function theorem. Math. Program. 2010, 123, 139–159. [Google Scholar] [CrossRef]
  22. Dontchev, A.L.; Rockafellar, R.T. Convergence of inexact Newton methods for generalized equations. Math. Program. 2013, 139, 115–137. [Google Scholar] [CrossRef]
  23. Ferreira, O.P. A robust semi-local convergence analysis of Newtons method for cone inclusionproblems in Banach spaces under affine invariant majorant condition. J. Comput. Appl. Math. 2015, 279, 318–335. [Google Scholar] [CrossRef]
  24. Ferreira, O.P.; Goncalves, M.L.N.; Oliveira, P.R. Local convergence analysis of the Gauss–Newton method under a majorant condition. J. Complex. 2011, 27, 111–125. [Google Scholar] [CrossRef]
  25. Ferreira, O.P.; Goncalves, M.L.N.; Oliveira, P.R. Convergence of the Gauss–Newton method for convex composite optimization under a majorant condition. SIAM J. Optim. 2013, 23, 1757–1783. [Google Scholar] [CrossRef]
  26. Ferreira, O.P.; Silva, G.N. Inexact Newton’s Method to Nonlinear Functions with Valuesin a Cone. arXiv 2015, arXiv:1510.01947. [Google Scholar]
  27. Ferreira, O.P.; Silva, G.N. Local Convergence Analysis of Newton’s Method for Solving Strongly Regular Generalized Equations. arXiv 2016, arXiv:1604.04568. [Google Scholar] [CrossRef]
  28. Ferreira, O.P.; Svaiter, B.F. Kantorovich’s majorants principle for Newton’s method. Comput. Optim. Appl. 2009, 42, 213–229. [Google Scholar] [CrossRef]
  29. Háussler, W.M. A Kantorovich-type convergence analysis for the Gauss–Newton–Method. Numer. Math. 1986, 48, 119–125. [Google Scholar] [CrossRef]
  30. Josephy, N. Newton’s Method for Generalized Equations and the PIES Energy Model; University of Wisconsin-Madison: Madison, WI, USA, 1979. [Google Scholar]
  31. Nashed, M.Z.; Chen, X. Convergence of Newton-like methods for singular operators using outer inverses. Numer. Math. 1993, 66, 235–257. [Google Scholar] [CrossRef]
  32. Proinov, P.D. New general convergence theory for iterative processes and its applications to Newton–Kantarovich type theorems. J. Complex. 2010, 25, 3–42. [Google Scholar] [CrossRef]
  33. Rheinboldt, W.C. A unified convergence theory for a class of iterative process. SIAM J. Numer. Anal. 1968, 5, 42–63. [Google Scholar] [CrossRef]
  34. Robinson, S.M. Extension of Newton’s method to nonlinear functions with values in a cone. Numer. Math. 1972, 19, 341–347. [Google Scholar] [CrossRef]
  35. Robinson, S.M. Normed convex processes. Trans. Amer. Math. Soc. 1972, 174, 127–140. [Google Scholar] [CrossRef]
  36. Robinson, S.M. Strongly regular generalized equations. Math. Oper. Res. 1980, 5, 43–62. [Google Scholar] [CrossRef]
  37. Argyros, I.K.; George, S.; Shakhno, S.; Regmi, S.; Havdiak, M.; Argyros, M.I. Asymptotically Newton–Type Methods without Inverses for Solving Equations. Mathematics 2024, 12, 1069. [Google Scholar] [CrossRef]
  38. Argyros, I.K.; George, S.; Regmi, S.; Argyros, C.I. Hybrid Newton-like Inverse Free Algorithms for Solving Nonlinear Equations. Algorithms 2024, 17, 154. [Google Scholar] [CrossRef]
  39. Deuflhard, P.; Heindl, G. Affine invariant convergence theorems for Newton’s method and extensions to related methods. SIAM J. Numer. Anal. 1979, 16, 1–10. [Google Scholar] [CrossRef]
  40. Kantorovich, L.V.; Akilov, G. Functional Analysis in Normed Spaces; Fizmatgiz: Moscow, Russia, 1959. [Google Scholar]
  41. Silva, G.N. Kantorovich’s theorem on Newton’s method for solving generalized equationsunder the majorant condition. Appl. Math. Comput. 2016, 286, 178–188. [Google Scholar]
  42. Yamamoto, T. Uniqueness of the solution in a Kantorovich-type theorem of Haussler for the Gauss–Newton Method. Jpn. J. Appl. Math. 1989, 6, 77–81. [Google Scholar] [CrossRef]
  43. Ezquerro, J.A.; Gutierrez, J.M.; Hernandez, M.A.; Romero, N.; Rubio, M.J. The Newton algorithm: From Newton to Kantorovich. Gac. R. Soc. Mat. Esp. 2010, 13, 53–76. (In Spanish) [Google Scholar]
  44. Ezquerro, J.A.; Hernandez-Veron, M.A. Domain of global convergence for Newton’s algorithm from auxiliary points. Appl. Math. Lett. 2018, 85, 48–56. [Google Scholar] [CrossRef]
  45. Catinas, E. The inexact, inexact perturbed, and quasi-Newton algorithms are equivalent models. Math. Comp. 2005, 74, 291–301. [Google Scholar] [CrossRef]
  46. Candelarion, G.; Cordero, A.; Torregrosa, J.R.; Vassileva, M.P. Generalized conformable fractional Newton-type method for solving nonlinear systems. Numer. Algorithms 2023, 93, 1171–1208. [Google Scholar] [CrossRef]
  47. Erfanifar, R.; Hajariah, M.A. A new multi-step method for solving nonlinear systems with high efficiency indices. Numer. Algorithms 2024. [Google Scholar] [CrossRef]
  48. Gutierrez, J.M.; Hernandez, M.A. Newtons method under weak Kantorovich conditions. IMA J. Numer. Anal. 2000, 20, 521–532. [Google Scholar] [CrossRef]
  49. Singh, S. A third order iterative algorithm for inversion of cumulative beta distribution. Numer. Algor. 2023, 94, 1331–1353. [Google Scholar]
Table 1. Iterations to obtain error tolerance ε = 10 9 for initial point x 0 = ( 1 , 1 ) , where I Υ ( x 0 ) = 0.3129 < 1 .
Table 1. Iterations to obtain error tolerance ε = 10 9 for initial point x 0 = ( 1 , 1 ) , where I Υ ( x 0 ) = 0.3129 < 1 .
MethodIterationsCPU TimeMethodIterationsCPU Time
(4)’ Newton4 10.82 × 10 6 (4)’ Newton4 10.82 × 10 6
(21),  m = 1 6 6.94 × 10 6 (26),  m = 1 8 13.063 × 10 6
(22),  m = 2 5 6.922 × 10 6 (26),  m = 2 6 12.255 × 10 6
(23),  m = 3 4 5.946 × 10 6 (26),  m = 3 5 9.089 × 10 6
(24),  m = 4 4 6.324 × 10 6 (26),  m = 4 5 9.198 × 10 6
(25),  m = 5 4 6.836 × 10 6 (26),  m = 5 4 8.743 × 10 6
Table 2. Iterations to obtain error tolerance of ε = 10 9 for initial point x 0 = ( 0 , 0 ) , where I Υ ( x 0 ) = 0.1414 < 1 .
Table 2. Iterations to obtain error tolerance of ε = 10 9 for initial point x 0 = ( 0 , 0 ) , where I Υ ( x 0 ) = 0.1414 < 1 .
MethodIterationsCPU TimeMethodIterationsCPU Time
(4)’ Newton3 7.215 × 10 6 (4)’ Newton3 7.215 × 10 6
(21),  m = 1 5 5.122 × 10 6 (26),  m = 1 3 4.256 × 10 6
(22),  m = 2 4 5.8 × 10 6 (26),  m = 2 3 5.862 × 10 6
(23),  m = 3 3 5.503 × 10 6 (26),  m = 3 3 5.575 × 10 6
(24),  m = 4 3 5.771 × 10 6 (26),  m = 4 3 5.825 × 10 6
(25),  m = 5 3 5.895 × 10 6 (26),  m = 5 3 5.973 × 10 6
Table 3. Iterations to obtain error tolerance of ε = 10 9 , for x 0 = ( 15 , 15 ) , where I Υ ( x 0 ) = 0.257 < 1 .
Table 3. Iterations to obtain error tolerance of ε = 10 9 , for x 0 = ( 15 , 15 ) , where I Υ ( x 0 ) = 0.257 < 1 .
MethodIterationsCPU TimeMethodIterationsCPU Time
(4)’ Newton5 13.454 × 10 6 (4)’ Newton5 13.454 × 10 6
(21),  m = 1 7 7.184 × 10 6 (26),  m = 1 9 12.015 × 10 6
(22),  m = 2 5 10.195 × 10 6 (26),  m = 2 7 13.245 × 10 6
(23),  m = 3 5 7.352 × 10 6 (26),  m = 3 6 9.378 × 10 6
(24),  m = 4 5 7.883 × 10 6 (26),  m = 4 6 9.829 × 10 6
(25),  m = 5 5 8.636 × 10 6 (26),  m = 5 5 8.936 × 10 6
Table 4. Iterations to obtain error tolerance of ε = 10 12 , for x 0 = ( 15 , 15 ) , where I Υ ( x 0 ) = 0.257 < 1 .
Table 4. Iterations to obtain error tolerance of ε = 10 12 , for x 0 = ( 15 , 15 ) , where I Υ ( x 0 ) = 0.257 < 1 .
MethodIterationsCPU TimeMethodIterationsCPU Time
(4)’ Newton7 16.719 × 10 6 (4)’ Newton7 16.719 × 10 6
(21),  m = 1 8 15.677 × 10 6 (26),  m = 1 12 24.742 × 10 6
(22),  m = 2 7 20.052 × 10 6 (26),  m = 2 8 27.178 × 10 6
(23),  m = 3 7 14.711 × 10 6 (26),  m = 3 8 20.43 × 10 6
(24),  m = 4 7 15.357 × 10 6 (26),  m = 4 7 19.2 × 10 6
(25),  m = 5 7 16.49 × 10 6 (26),  m = 5 7 17.538 × 10 6
Table 5. COC versus ACOC with x 0 = ( 15 , 15 ) , ε = 10 12 .
Table 5. COC versus ACOC with x 0 = ( 15 , 15 ) , ε = 10 12 .
MethodCOCACOC
(2)   Newton1.86241.9697
(21),    m = 1 0.8631
(22),    m = 2 0.26951.0438
(23),    m = 3 1.97142.3569
(24),    m = 4 1.83541.9453
(25),    m = 5 1.86421.9661
(26),    m = 1 0.90651.0118
(26),    m = 2 0.59120.999
(26),    m = 3 0.73210.9926
(26),    m = 4 1.9332.0151
(26),    m = 5 1.86791.9578
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Argyros, I.K.; George, S.; Regmi, S.; Argyros, M.I. On the Kantorovich Theory for Nonsingular and Singular Equations. Axioms 2024, 13, 358. https://doi.org/10.3390/axioms13060358

AMA Style

Argyros IK, George S, Regmi S, Argyros MI. On the Kantorovich Theory for Nonsingular and Singular Equations. Axioms. 2024; 13(6):358. https://doi.org/10.3390/axioms13060358

Chicago/Turabian Style

Argyros, Ioannis K., Santhosh George, Samundra Regmi, and Michael I. Argyros. 2024. "On the Kantorovich Theory for Nonsingular and Singular Equations" Axioms 13, no. 6: 358. https://doi.org/10.3390/axioms13060358

APA Style

Argyros, I. K., George, S., Regmi, S., & Argyros, M. I. (2024). On the Kantorovich Theory for Nonsingular and Singular Equations. Axioms, 13(6), 358. https://doi.org/10.3390/axioms13060358

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop