Next Article in Journal
A Parallel Search Strategy Based on Sparse Representation for Infrared Target Tracking
Next Article in Special Issue
Some Improvements to a Third Order Variant of Newton’s Method from Simpson’s Rule
Previous Article in Journal
Multi-Feedback Interference Cancellation Algorithms for OFDM Systems over Doubly-Selective Channels
Previous Article in Special Issue
A Quartically Convergent Jarratt-Type Method for Nonlinear System of Equations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Accessibility of Newton’s Method under a Hölder Condition on the First Derivative

by
José Antonio Ezquerro
* and
Miguel Ángel Hernández-Verón
Department of Mathematics and Computation, University of La Rioja, Calle Luis de Ulloa s/n, 26004 Logroño, Spain
*
Author to whom correspondence should be addressed.
Algorithms 2015, 8(3), 514-528; https://doi.org/10.3390/a8030514
Submission received: 22 June 2015 / Revised: 15 July 2015 / Accepted: 17 July 2015 / Published: 23 July 2015
(This article belongs to the Special Issue Numerical Algorithms for Solving Nonlinear Equations and Systems)

Abstract

:
We see how we can improve the accessibility of Newton’s method for approximating a solution of a nonlinear equation in Banach spaces when a center Hölder condition on the first derivative is used to prove its semi-local convergence.

1. Introduction

It is often necessary in many scientific and engineering problems to find a root of a nonlinear equation. To solve this equation, we can use iterative methods. To give sufficient generality to the problem of approximating a solution of a nonlinear equation by iterative methods, we are concerned, in this work, with the problem of approximating a locally-unique solution x * of the equation F ( x ) = 0 , where F is a nonlinear operator defined on a nonempty open convex subset Ω of a Banach space X with values in a Banach space Y, so that many scientific and engineering problems can be written as a nonlinear equation in Banach spaces.
Newton’s method is probably the best-known iterative method, and it is well known that it converges quadratically. Newton’s method for solving scalar equations was extended to nonlinear equations in Banach spaces by Kantorovich in 1948 [1] in the following way:
x 0 given in Ω , x n = x n 1 [ F ( x n 1 ) ] 1 F ( x n 1 ) , n N ,
where F is the Fréchet derivative of the operator F. For this reason, many authors refer to it as the Newton–Kantorovich method.
Kantorovich proved the semi-local convergence of Newton’s method under the hypothesis that the operator involved F is twice differentiable Fréchet with bounded second derivative:
There exists a constant C 0 , such that F ( x ) C for x Ω .
It is well known that this condition can be replaced by a Lipschitz condition on the first derivative F of the operator involved [2]:
There exists a constant L 0 , such that F ( x ) F ( y ) L x y for x , y Ω .
In view of the applications, numerous papers have recently appeared where the convergence of the method is proven under different conditions for the derivatives of the operator F. A known variant is the study of the convergence of the method under a Hölder condition on the first derivative [3,4,5]:
There exist two constants K 0 and p ( 0 , 1 ] such that F ( x ) F ( y ) K x y p for x , y Ω .
In order to improve the accessibility of Newton’s method under the last condition, we can use different strategies that change the condition. In this work, we use a similar one as that used by Argyros in [6] for the Lipschitz condition on the first derivative F , which consists of noticing that, as a consequence of the last condition being satisfied in Ω, we have, for the starting point x 0 , that:
  • There exist two constants K 0 0 and p ( 0 , 1 ] , such that F ( x ) F ( x 0 ) K 0 x x 0 p for x Ω ,
with K 0 K . We then say that F is center Hölder in x 0 .
In this paper, we focus our attention on the analysis of the semi-local convergence of Sequence (1), which is based on demanding the last condition to the initial approximation x 0 and provides the so-called domain of parameters corresponding to the conditions required for the initial approximation that guarantee the convergence of Sequence (1) to the solution x * .
In this work, we carry out an analysis of the domain of parameters for Newton’s method under the last two conditions for F and use a technique based on recurrence relations. As a consequence, we improve the domains of parameters associated with the semi-local convergence result given for Newton’s method by Hérnández in [4].
We prove in this paper that center conditions on the first derivative of the operator involved in the solution of nonlinear equations play an important role in the semi-local convergence of Newton’s method, since we can improve the accessibility of Newton’s method from them.
Throughout the paper, we denote B ( x , ϱ ) ¯ = { y X ; y x ϱ } and B ( x , ϱ ) = { y X ; y x < ϱ } .

2. Preliminaries

The best-known semi-local convergence result for Newton’s method under a Hölder condition on the first derivative of the operator involved, when the technique of recurrence relations is used to prove it, is that given by Hernández in [4], which is established under the following conditions:
(A1)
There exists Γ 0 = [ F ( x 0 ) ] 1 L ( Y , X ) , for some x 0 Ω , with Γ 0 β and Γ 0 F ( x 0 ) η , where L ( Y , X ) is the set of bounded linear operators from Y to X.
(A2)
There exist two constants K 0 and p ( 0 , 1 ] , such that F ( x ) F ( y ) K x y p for x , y Ω .
(A3)
If ξ ( p ) is the unique zero of the auxiliary function:
ϕ ( x ; p ) = ( 1 + p ) p ( 1 x ) 1 + p x p , p ( 0 , 1 ] ,
in the interval 0 , 1 2 , it is satisfied that h = K β η p ξ ( p ) and B ( x 0 , R ) Ω , where R = ( 1 + p ) ( 1 h ) ( 1 + p ) ( 2 + p ) h η .
Theorem 1. (Theorem 2.1 of [4]) Let F : Ω X Y be a continuously-differentiable operator defined on a nonempty open convex domain Ω of a Banach space X with values in a Banach space Y. Suppose that Conditions (A1)(A3) are satisfied. Then, Newton’s sequence, given by (1), converges to a solution x * of the equation F ( x ) = 0 , starting at x 0 , and x n , x * B ( x 0 , R ) ¯ , for all n = 0 , 1 , 2 ,
Note that Condition (A1), required for the initial approximation x 0 , defines the parameters β and η, and Condition (A2), required for the operator F, defines the parameter K fixed for all point of Ω. Observe that, every point z Ω , such that the operator [ F ( z ) ] 1 exists with [ F ( z ) ] 1 β and [ F ( z ) ] 1 F ( z ) η , has associated the pair ( K , β η p ) of the x y -plane, where x = K and y = β η p . In addition, fixed p ( 0 , 1 ] , if we consider the set:
D T = ( x , y ) R 2 : x y ξ ( p ) ,
we can observe that every point z, such that its associated pair ( K , β η p ) belongs to D T , can be chosen as the starting point for Newton’s method, so that the method converges to a solution x * of the equation F ( x ) = 0 when it starts at z. The set D T is then called the domain of parameters associated with Theorem 1 and can be drawn by choosing x = K , y = β η p and coloring the region of the x y -plane whose points satisfy the condition h ξ ( p ) (namely, x y ξ ( p ) ) of the Theorem 1. In Figure 1, we see the domain of parameters D T (blue region).
In relation to the above, we can think that the larger the size of the domain of parameters is, the more possibilities we have for choosing good starting points for Newton’s method. As a consequence, we are interested in D T being as big as possible, since this fact allows us to find a greater number of good starting points for Newton’s method.

3. Semi-Local Convergence of Newton’s Method

To improve the semi-local convergence of Newton’s method from increasing the domain of parameters D T , we consider a procedure that consists of observing that, as a consequence of Condition (A2), once x 0 Ω is fixed, the condition:
F ( x ) F ( x 0 ) K 0 x x 0 p , x Ω ,
is satisfied with K 0 K . Then, by considering jointly the parameters K and K 0 , we can relax the semi-local convergence conditions of Newton’s method given in Theorem 1 and obtain a larger domain of parameters.
Now, we present a semi-local convergence result for Newton’s method under Condition (A1) for the starting point x 0 and Condition (A2) for the first derivative F . Note that Condition (3) follows from Condition (A2) for the starting point x 0 . In addition, we obtain a semi-local convergence result by combining Conditions (A2) and (3), which allows increasing the domain of parameters D T , so that the possibility of choosing good starting points for Newton’s method is increased, as we can see later in the applications. In particular, we study the convergence of Newton’s method to a solution of the equation F ( x ) = 0 under certain conditions for the pair ( F , x 0 ) . From some real parameters, a system of four recurrence relations is constructed in which two sequences of positive real numbers are involved. The convergence of Newton’s method is then guaranteed from them.

3.1. Auxiliary Scalar Sequences

From Conditions (A1)–(A2), we define γ = ε h , b 0 = h and δ = γ ( 1 + p ) ( 1 γ ) , where ε = K 0 K [ 0 , 1 ] . Now, we define b 1 = δ p 1 γ b 0 and:
a n = b n ( 1 + p ) ( 1 b n ) , n 1 ,
b n = b n 1 a n 1 p 1 b n 1 , n 2 .
Observe that we consider the case b 0 > 0 , since if b 0 = 0 , a trivial problem results, as the solution of the equation F ( x ) = 0 is x 0 .
Next, we prove the following four recurrence relations for Sequences (1), (4) and (5):
Γ 1 = [ F ( x 1 ) ] 1 Γ 0 1 γ ,
x 2 x 1 δ x 1 x 0 ,
K Γ 1 x 2 x 1 p b 1 ,
x 2 x 0 ( 1 + δ ) x 1 x 0 ,
provided that:
x 1 Ω a n d γ < 1 .
If x 1 Ω , then:
I Γ 0 F ( x 1 ) Γ 0 F ( x 0 ) F ( x 1 ) K β η p = h < 1 .
Then, by the Banach lemma on invertible operators, it follows that the operator Γ 1 exists and:
Γ 1 Γ 0 1 I Γ 0 F ( x 1 ) Γ 0 1 γ .
After that, from Taylor’s series and Sequence (1), we have:
F ( x 1 ) = 0 1 ( F ( x 0 + τ ( x 1 x 0 ) ) F ( x 0 ) ) ( x 1 x 0 ) d τ K 0 η p 1 + p x 1 x 0 .
Thus,
x 2 x 1 Γ 1 F ( x 1 ) δ x 1 x 0 Γ 1 x 2 x 1 p K δ p 1 γ Γ 0 x 1 x 0 p b 1 , x 2 x 0 x 2 x 1 + x 1 x 0 ( 1 + δ ) x 1 x 0 < 1 + δ 1 a 1 η = R ,
provided that δ 1 and a 1 < 1 .
Now, we prove in the same way as above the following four recurrence relations for Sequences (1), (4) and (5):
Γ 2 = [ F ( x 2 ) ] 1 Γ 1 1 b 1 ,
x 3 x 2 a 1 x 2 x 1 ,
K Γ 2 x 3 x 2 p b 2 ,
x 3 x 0 1 + δ 1 + a 1 x 1 x 0 ,
provided that:
x 2 Ω a n d b 1 < 1 .
In addition, we generalize the last recurrence relations to every point of Sequence (1), so that we can guarantee that (1) is a Cauchy sequence from them. For this, we analyze the scalar sequences defined in (4) and (5) in order to prove later the semi-local convergence of Sequence (1). For this, it suffices to see that (1) is a Cauchy sequence and (10) and (15) are true for all x n and b n 1 with n 3 . We begin by presenting a technical lemma.
Lemma 2 If γ 1 + p 2 + p and b 1 is such that
b 1 < 1 + p 2 + p and b 1 + a 1 p < 1 ,
then:
(a)
the sequences { b n } and { a n } are strictly decreasing,
(b)
a n < 1 and b n < 1 , for all n 1 .
If b 1 = 1 a 1 p < 1 + p 2 + p , then a n = a 1 < 1 and b n = b 1 < 1 for all n 2 .
Proof. We first consider the case in which b 1 satisfies (16). Item (a) is proven by mathematical induction on n. As b 1 + a 1 p < 1 , then b 2 < b 1 and a 2 < a 1 . If we now suppose that b j < b j 1 and a j < a j 1 , for all i = 2 , 3 , , n , then:
b n + 1 = b n a n p 1 b n < b n a 1 p 1 b 1 < b n and a n + 1 = b n + 1 ( 1 + p ) ( 1 b n + 1 ) < a n .
As a result, the sequences { a n } and { b n } are strictly decreasing for all n 2 .
To see Item (b), we have a n < a 1 < 1 and b n < b 1 < 1 , for all n 2 , by Item (a) and the conditions given in (16).
Second, if b 1 = 1 a 1 p , then b n = b 1 = 1 a 1 p < 1 , for all n 2 . Moreover, if b 1 < 1 + p 2 + p , then we have a n = a 1 < 1 , for all n 2 .   □

3.2. Main Result

We now give a semi-local convergence result for Newton’s method from a modification of the convergence conditions required in Theorem 1. Therefore, we consider Conditions (A1), (A2) and the modification of Condition (A3) given by:
  • (A3b) γ 1 + p 2 + p , b 1 satisfies (16) and B ( x 0 , R ) Ω , where R = 1 + δ 1 a 1 η .
Remember that Condition (3) follows from Condition (A2) for the starting point x 0 .
Theorem 3. Let F : Ω X Y be a continuously-differentiable operator defined on a nonempty open convex domain Ω of a Banach space X with values in a Banach space Y. Suppose that Conditions (A1), (A2) and (A3b) are satisfied. Then, Newton’s sequence, given by (1), converges to a solution x * of the equation F ( x ) = 0 , starting at x 0 , and x n , x * B ( x 0 , R ) ¯ , for all n = 0 , 1 , 2 , .
proof. We begin by proving the following four items for Sequences (1), (4) and (5) and n 3 :
(I)
There exists Γ n 1 = [ F ( x n 1 ) ] 1 and Γ n 1 Γ n 2 1 b n 2 ,
(II)
x n x n 1 a n 2 x n 1 x n 2 ,
(III)
K Γ n 1 x n x n 1 p b n 1 ,
(IV)
x n Ω .
Observe that x 1 Ω , since η < R . Moreover, from (6), (7), (8) and (9), it follows that x 2 Ω . Furthermore, from (11), (12), (13) and (14), we have that Items (I), (II), (III) and (IV) hold for n = 3 . If we now suppose that Items (I), (II) and (III) are true for some n 1 , it follows, by analogy to the case where n = 3 and induction, that Items (I), (II) and (III) also hold for n. Notice that b n < 1 for all n 1 . Now, we prove (IV). Therefore,
x n x 0 x n x n 1 + x n 1 x n 2 + + x 1 x 0 ( II ) 1 + i = 1 n 2 j = 1 i a j x 2 x 1 + x 1 x 0 < 1 + i = 1 n 2 a 1 i x 2 x 1 + x 1 x 0 < 1 1 a 1 x 2 x 1 + x 1 x 0 1 + δ 1 a 1 x 1 x 0 R
and x n B ( x 0 , R ) . As B ( x 0 , R ) Ω , then x n Ω for all n 0 . Note that the conditions given in (15) are satisfied for all x n and b n 1 with n 3 .
Next, we prove that { x n } is a Cauchy sequence. For this, we follow an analogous procedure to the latter. Therefore, for m 2 and n 2 , we have:
x n + m x n x n + m x n + m 1 + x n + m 1 x n + m 2 + + x n + 1 x n ( II ) i = n 1 n + m 2 j = 1 i a j x 2 x 1 < L e m m a ( a ) i = n 1 n + m 2 a 1 i x 2 x 1 = i = 0 m 1 a 1 n + i 1 x 2 x 1 = 1 a 1 m 1 a 1 a 1 n 1 x 2 x 1 .
Thus, { x n } is a Cauchy sequence.
After that, we prove that x * is a solution of equation F ( x ) = 0 . As Γ n F ( x n ) 0 when n , if we take into account that:
F ( x n ) F ( x n ) Γ n F ( x n )
and { F ( x n ) } is bounded, since:
F ( x n ) F ( x 0 ) + K 0 x n x 0 p F ( x 0 ) + K 0 R p ,
it follows that F ( x n ) 0 when n . As a consequence, we obtain F ( x * ) = 0 by the continuity of F in B ( x 0 , R ) ¯ .   □

4. Accessibility of Newton’s Method

The accessibility of an iterative method is analyzed from the set of possible starting points that guarantee the convergence of the iterative method when it starts at them. As we have indicated, the set of starting points that guarantees the convergence of the iterative method is related to the domain of parameters associated with a result of semi-local convergence of the iterative method.
Next, we study the domain of parameters associated with Theorem 1 and compare it with that associated with Theorem 3. To guarantee the convergence of Newton’s method from Theorem 3, the following three conditions must be satisfied:
γ 1 + p 2 + p , b 1 < 1 + p 2 + p and b 1 + b 1 p ( 1 + p ) p ( 1 b 1 ) p < 1 ,
where p ( 0 , 1 ] .
From the auxiliary function given in (2), the third condition of (17) can be written as:
ϕ ( b 1 ; p ) > 0 .
Observe that ϕ ( x ; p ) is a non-increasing and convex function and ϕ ( 0 ; p ) = ( 1 + p ) p > 0 and ϕ 1 2 ; p 0 , for all p ( 0 , 1 ] . Besides, if p = 1 , the unique zero of ϕ ( x ; 1 ) in the interval 0 , 1 2 is 1 2 . If we now denote, for a fixed p ( 0 , 1 ] , the unique zero of ϕ ( x ; p ) in 0 , 1 2 by ξ ( p ) and demand b 1 < ξ ( p ) , then Condition (18) holds. Moreover, since ξ ( p ) 1 2 , the second condition of (17) is satisfied. Now, as:
b 1 = δ p b 0 1 γ = γ p b 0 ( 1 + p ) p ( 1 γ ) 1 + p ,
the second and third conditions of (17) are satisfied, provided that:
b 0 < ξ ( p ) ( 1 + p ) p ( 1 γ ) 1 + p γ p .
After that, we write the first condition of (17) as:
h 1 + p ε ( 2 + p )
and take into account that the second and third conditions of (17) are satisfied if:
h < ξ ( p ) ( 1 + p ) p ( 1 ε h ) 1 + p ε p h p ,
since b 0 = h and γ = ε h . In addition, Condition (20) is equivalent to:
ϖ ( h ) = ε p h 1 + p ξ ( 1 + p ) p ( 1 ε h ) 1 + p < 0 .
Furthermore, ϖ ( h ) 0 , so that ϖ ( h ) is a nondecreasing function for all h 0 , and ϖ ( 0 ) 0 and ϖ 1 + p ε ( 2 + p ) > 0 . Therefore, Condition (19) is satisfied if Condition (20) holds, and we can then give the following result.
Corollary 4 Let F : Ω X Y be a continuously-differentiable operator defined on a nonempty open convex domain Ω of a Banach space X with values in a Banach space Y. Suppose that (A1)–(A2) are satisfied. If Condition (20), where ξ ( p ) is the unique zero of function (2) in the interval 0 , 1 2 , is satisfied and B ( x 0 , R ) Ω , where R = 1 + υ 1 ν η , υ = ε h ( 1 + p ) ( 1 ε h ) , ν = ϑ ( 1 + p ) ( 1 ϑ ) and ϑ = h υ p 1 ε h , then Newton’s sequence, given by (1), converges to a solution x * of the equation F ( x ) = 0 , starting at x 0 , and x n , x * B ( x 0 , R ) ¯ , for all n = 0 , 1 , 2 ,
From the last result, we can define the following domains of parameters:
D C ε = ( x , y ) R 2 : x y ξ ( p ) ( 1 + p ) p ( 1 ε x y ) 1 + p ε p x p y p ,
where ε = K 0 K [ 0 , 1 ] , p ( 0 , 1 ] and ξ ( p ) is the unique zero of Function (2) in the interval 0 , 1 2 .
Next, we compare the conditions required for the semi-local convergence of Newton’s method in Theorem 1 and Corollary 4. In Figure 1, we see that D T D C ε for p = 1 10 , 1 5 , 2 5 , 4 5 , so that we can guess that the smaller the quantity ε = K 0 K [ 0 , 1 ] is, the larger the domain of parameters is: orange for ε = 0 . 1 , green for ε = 0 . 2 , red for ε = 0 . 4 and yellow for ε = 0 . 8 . Note that, if ε 1 , the domain of parameters associated with Corollary 4, D C ε , tends to be that obtained by Theorem 1 (blue region). As a consequence,
D T = D C 1 = D C ε j D C ε j 1 D C ε 0 for ε 0 < < ε j 1 < ε j = 1 .
Figure 1. Domains of parameters of Newton’s method associated with Theorem 1 (blue region) and Corollary 4 (orange for ε = 0 . 1 , green for ε = 0 . 2 , red for ε = 0 . 4 and yellow for ε = 0 . 8 ) for p = 1 10 , 1 5 , 2 5 , 4 5 .
Figure 1. Domains of parameters of Newton’s method associated with Theorem 1 (blue region) and Corollary 4 (orange for ε = 0 . 1 , green for ε = 0 . 2 , red for ε = 0 . 4 and yellow for ε = 0 . 8 ) for p = 1 10 , 1 5 , 2 5 , 4 5 .
Algorithms 08 00514 g001
On the other hand, in Figure 2, we observe the relationship between the domains of parameters associated with Theorem 1 (gray region) and Corollary 4 (magenta region) from the variability of ε and for four different values of p: p = 1 10 , 1 5 , 2 5 , 4 5 . As we can see, the domain associated with Corollary 4 is always larger, for all ε [ 0 , 1 ] , than that associated with Theorem 1.
Figure 2. Domains of parameters of Newton’s method associated with Corollary 4 (magenta region) and Theorem 1 (gray region) for p = 1 10 , 1 5 , 2 5 , 4 5 .
Figure 2. Domains of parameters of Newton’s method associated with Corollary 4 (magenta region) and Theorem 1 (gray region) for p = 1 10 , 1 5 , 2 5 , 4 5 .
Algorithms 08 00514 g002
In addition, we prove analytically in the following what we have just seen graphically. First, we prove that D T D C ε , for each p ( 0 , 1 ] and ε [ 0 , 1 ] . For this, if ( K , β η p ) D T , then h ξ ( p ) 1 2 1 2 ε 1 + p ε ( 2 + p ) , since p ( 0 , 1 ] and ε [ 0 , 1 ] . Moreover, h ξ ( p ) ξ ( p ) ( 1 + p ) p ( 1 ε h ) 1 + p ε p h p , since ϕ ( ε h ; p ) 0 when h ξ ( p ) and ε h ξ ( p ) . Therefore, ( K , β η p ) D C ε for each p ( 0 , 1 ] and ε [ 0 , 1 ] .
Finally, we see that D C ε 2 D C ε 1 if ε 1 < ε 2 with ε 1 , ε 2 [ 0 , 1 ] . Indeed, from ψ ( ε ) = 1 + p ( 2 + p ) ε and φ ( ε ) = ξ ( p ) ( 1 + p ) p ( 1 ε h ) 1 + p ε p h p , we have ψ ( ε ) = ( 1 + p ) ( 2 + p ) ε 2 0 and φ ( ε ) = ( 1 + p ) p ξ ( p ) ( 1 ε h ) p ( p + ε h ) ε 1 + p h p 0 , since 1 ε h 1 2 + p 0 , so that ψ ( ε 2 ) ψ ( ε 1 ) and φ ( ε 2 ) φ ( ε 1 ) , and therefore D C ε 2 D C ε 1 .
To that end, we have proven the improvement obtained for the domain of parameters of Newton’s method with the help of the conditions of type (3) that we have just shown in Figure 2.

5. Application

Now, we illustrate all of that mentioned above with the following mildly nonlinear elliptic equation:
u x x + u y y = u 5 / 3 .
This type of equation is of interest in the theory of gas dynamics [7]. An associated Dirichlet problem can be formulated as follows. Suppose that the equation is satisfied in the interior of the square 0 x , y 1 in R 2 and that u ( x , y ) > 0 is given and continuous on the boundary of the square ([8]):
u ( x , 0 ) = 2 x 2 x + 1 , u ( x , 1 ) = 2 , 0 x 1 , u ( 0 , y ) = 2 y 2 y + 1 , u ( 1 , y ) = 2 , 0 y 1 .
Our discussion focuses on the formulation of the finite difference equation for the elliptic boundary value Problems (21)–(22). The method of finite differences applied to this problem yields a finite system of equations. For general use, iterative techniques often represent the best approach to the solution of such finite systems of equations.
Specifically, central difference approximations for (21) are used, so that Problems (21)–(22) are reduced to the problem of finding a real zero of a function F : Ω R m R m , namely a real solution x * of a nonlinear system F ( x ) = 0 with m equations and m unknowns. The common technique used to approximate x * is the application of iterative methods. In this case, Newton’s method goes on being the most used iterative method for approximating x * , since this method is one of the most efficient.
For Problems (21)–(22) in R 2 , Equation (21) can be approximated using central difference approximations for the spacial derivatives. Consider a grid with step size h = 1 N + 1 in x and k = 1 M + 1 in y defined over the region D, so that D is partitioned into a grid consisting of ( N + 1 ) × ( M + 1 ) rectangles with sides h = 1 N + 1 and k = 1 M + 1 . The mesh points ( x i , y j ) are given by:
x i = i h , y i = j k , i = 0 , 1 , , N + 1 , j = 0 , 1 , , M + 1 .
Considering the following finite difference expressions to approximate the partial differentials:
u x x ( x i , y j ) = u ( x i 1 , y j ) 2 u ( x i , y j ) + u ( x i + 1 , y j ) h 2 + O ( h 2 ) ,
u y y ( x i , y j ) = u ( x i , y j 1 ) 2 u ( x i , y j ) + u ( x i , y j + 1 ) k 2 + O ( k 2 ) ,
Equation (21) is approximated at each interior grid point ( x i , y j ) by the difference equation:
u ( x i + 1 , y j ) 2 u ( x i , y j ) + u ( x i 1 , y j ) h 2 + u ( x i , y j + 1 ) 2 u ( x i , y j ) + u ( x i , y j 1 ) k 2 = u ( x i , y j ) 5 / 3 ,
for i = 1 , 2 , , N and j = 1 , 2 , , M . The boundary conditions are:
u ( x i , y 0 ) = 2 x i 2 x i + 1 , u ( x i , y M ) = 2 , i = 1 , 2 , , N , u ( x 0 , y j ) = 2 y j 2 y j + 1 , u ( x N , y j ) = 2 , j = 0 , 1 , , M + 1 .
If we now denote the approximate value of u ( x i , y j ) as u i , j , we obtain the difference equation:
2 h k 2 + 1 u i , j ( u i 1 , j + u i + 1 , j ) h k 2 ( u i , j 1 + u i , j + 1 ) = h 2 u i , j 5 / 3 ,
for i = 1 , 2 , , N and j = 1 , 2 , , M , with:
u i , 0 = 2 x i 2 x i + 1 , u i , M + 1 = 2 , i = 1 , 2 , , N , u 0 , j = 2 y j 2 y j + 1 , u N + 1 , j = 2 , j = 0 , 1 , , M + 1 .
Equation (21) with the boundary conditions given in (22) forms an N M × N M nonlinear system of equations. To set up the nonlinear system, the N M = m interior grid points are labeled row-by-row from x 1 to x m starting from the left-bottom corner point. The resulting system is:
A x + h 2 q ( x ) = v ,
where:
A = B C 0 0 C B C 0 C B C 0 C 0 0 C B M × M ,
B = 2 ( λ + 1 ) 1 0 0 1 2 ( λ + 1 ) 1 0 1 2 ( λ + 1 ) 1 0 1 0 0 1 2 ( λ + 1 ) N × N ,
C = λ I , λ = h k 2 , I is the identity matrix in R N , x = ( x 1 , x 2 , , x m ) t , q ( x ) = x 1 5 / 3 , x 2 5 / 3 , , x m 5 / 3 t and v is a vector formed from the boundary conditions (systems of this type are so-called mildly nonlinear systems).
If we denote the previous system by F ( x ) = 0 , where:
F ( x ) = A x + h 2 q ( x ) v and F : R m R m , m = N M ,
then F ( x ) is a linear operator, which is given by the matrix:
F ( x ) = A + h 2 Q ( x ) , Q ( x ) = 5 3 diag x 1 2 / 3 , x 2 2 / 3 , , x m 2 / 3 .
We now choose N = M = 4 and the infinity norm. In addition, the number of equations is m = 16 , h = k = 1 5 and λ = 1 . Besides, q ( x ) = x 1 5 / 3 , x 2 5 / 3 , , x 16 5 / 3 t and:
v = 44 25 , 23 25 , 28 25 , 87 25 , 23 25 , 0 , 0 , 2 , 28 25 , 0 , 0 , 2 , 87 25 , 2 , 2 , 4 t .
In this case, we observe that a solution x * of the system F ( x ) = 0 with F defined in (23) satisfies:
x * A 1 v + h 2 q ( x ) x * 5 3 4 + 1 25 x * 5 / 3 0 ,
where A 1 = 5 3 and v = 4 , so that x * [ 0 , ϱ 1 ] [ ϱ 2 , + ] , where ϱ 1 = 9 . 5149 and ϱ 2 = 45 . 9125 are the two positive real roots of the scalar equation t 5 3 4 + t 5 / 3 25 = 0 . Then, we can consider:
F : Ω R 16 R 16 with Ω = x R 16 : x < 10 ,
since ϱ 1 < 10 < ϱ 2 .
Moreover, F ( x ) is the linear operator given by the matrix:
A + 1 15 diag x 1 2 / 3 , x 2 2 / 3 , , x 16 2 / 3
and:
F ( x ) F ( y ) = 1 15 diag x 1 2 / 3 y 1 2 / 3 , x 2 2 / 3 y 2 2 / 3 , , x 16 2 / 3 y 16 2 / 3 ,
where y = ( y 1 , y 2 , , y 16 ) t . In addition,
F ( x ) F ( y ) 1 15 x 1 / 3 + y 1 / 3 x 1 / 3 y 1 / 3 2 15 10 3 x y 1 / 3 , F ( x ) F ( x 0 ) 1 15 x 1 / 3 + x 0 1 / 3 x x 0 1 / 3 1 15 10 3 + x 0 1 / 3 x x 0 1 / 3 .
Thus, K = 0 . 2872 , K 0 = 0 . 2199 and p = 1 3 .
If we choose the starting point x 0 = 3 2 , 3 2 , , 3 2 t , we obtain β = 1 . 4862 and η = 0 . 5185 , so that the condition h = K β η p ξ ( p ) of Theorem 1 is not satisfied, since h = 0 . 3429 > ξ ( p ) = ξ 1 3 = 0 . 3071 , where ξ 1 3 is the unique zero of the corresponding auxiliary function given by (2), ϕ x ; 1 3 = 1 3 6 2 / 3 ( 1 t ) 4 / 3 3 t 3 , in the interval 0 , 1 2 . As a consequence, we cannot use Theorem 1 to guarantee the convergence of Newton’s method for approximating a solution of the system F ( x ) = 0 , where F is defined in (23).
However, we can guarantee the convergence of Newton’s method from Corollary 4, since Condition (20) is satisfied: h = 0 . 3429 < ξ ( p ) ( 1 + p ) p ( 1 ε h ) 1 + p ε p h p = 0 . 3517 with ε = 0 . 7656 . Therefore, we can then apply Newton’s method for approximating a solution of the system F ( x ) = 0 with F defined in (23) and obtain the approximation given by the vector x * = ( x 1 * , x 2 * , , x 16 * ) t that is shown in Table 1, reached after four iterations with a tolerance 10 16 . In Table 2, we show the errors x n x * using the stopping criterion x n x n 1 < 10 16 . Notice that the vector shown in Table 1 is a good approximation of the solution of the system, since F ( x * ) C × 10 16 . See the sequence { F ( x n ) } in Table 2.
Table 1. Approximation of the solution x * of F ( x ) = 0 with F given in (23).
Table 1. Approximation of the solution x * of F ( x ) = 0 with F given in (23).
i x i * i x i * i x i * i x i *
1 0 . 979069 5 1 . 097445 9 1 . 291651 13 1 . 587503
2 1 . 097445 6 1 . 245767 10 1 . 422935 14 1 . 664776
3 1 . 291651 7 1 . 422935 11 1 . 561551 15 1 . 742203
4 1 . 587503 8 1 . 664776 12 1 . 742203 16 1 . 843388
Table 2. Absolute errors obtained by Newton’s method and { F ( x n ) } .
Table 2. Absolute errors obtained by Newton’s method and { F ( x n ) } .
n x n x * F ( x n )
0 5 . 2093 × 10 1 1 . 318622
1 2 . 4028 × 10 3 5 . 4477 × 10 3
2 8 . 4485 × 10 8 1 . 2913 × 10 7
3 1 . 2409 × 10 16 1 . 4741 × 10 16
Finally, we note that if we interpolate the points of Table 1 and take into account that the solution satisfies the boundary conditions given in (22), we obtain the approximation of the numerical solution shown in Figure 3.
Figure 3. Approximated solution of Problems (21)–(22).
Figure 3. Approximated solution of Problems (21)–(22).
Algorithms 08 00514 g003

Acknowledgments

This work has been partially supported by the project MTM2014-52016-C2-1-P of the Spanish Ministry of Economy and Competitiveness.

Author Contributions

The contributions of the two authors have been similar. Both authors have worked together to develop the present manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kantorovich, L.V.; Akilov, G.P. Functional Analysis; Pergamon Press: Oxford, UK, 1982. [Google Scholar]
  2. Fenyö, I. Über die lösung der in Banachschen raume definierten nichtlinearen gleichungen. Acta Math. Acad. Sci. Hung. 1954, 5, 85–93. [Google Scholar] [CrossRef]
  3. Argyros, I.K. Remarks on the convergence of Newton’s method under Hölder continuity conditions. Tamkang J. Math. 1992, 23, 269–277. [Google Scholar]
  4. Hernández, M.A. The Newton method for operators with Hölder continuous first derivative. J. Optim. Theory Appl. 2001, 109, 631–648. [Google Scholar] [CrossRef]
  5. Rokne, J. Newton’s method under mild differentiability conditions with error analysis. Numer. Math. 1972, 18, 401–412. [Google Scholar] [CrossRef]
  6. Argyros, I.K. On the Newton-Kantorovich hypothesis for solving equations. J. Comput. Appl. Math. 2004, 169, 315–332. [Google Scholar] [CrossRef]
  7. Greenspan, D. Introductory Numerical Analysis of Elliptic Boundary Value Problems; Harper and Row: New York, NY, USA, 1965. [Google Scholar]
  8. Rall, L.B. Computational Solution of Nonlinear Operator Equations; Robert E. Krieger Publishing Company: Malabar, FL, USA, 1979. [Google Scholar]

Share and Cite

MDPI and ACS Style

Ezquerro, J.A.; Hernández-Verón, M.Á. On the Accessibility of Newton’s Method under a Hölder Condition on the First Derivative. Algorithms 2015, 8, 514-528. https://doi.org/10.3390/a8030514

AMA Style

Ezquerro JA, Hernández-Verón MÁ. On the Accessibility of Newton’s Method under a Hölder Condition on the First Derivative. Algorithms. 2015; 8(3):514-528. https://doi.org/10.3390/a8030514

Chicago/Turabian Style

Ezquerro, José Antonio, and Miguel Ángel Hernández-Verón. 2015. "On the Accessibility of Newton’s Method under a Hölder Condition on the First Derivative" Algorithms 8, no. 3: 514-528. https://doi.org/10.3390/a8030514

APA Style

Ezquerro, J. A., & Hernández-Verón, M. Á. (2015). On the Accessibility of Newton’s Method under a Hölder Condition on the First Derivative. Algorithms, 8(3), 514-528. https://doi.org/10.3390/a8030514

Article Metrics

Back to TopTop