You are currently viewing a new version of our website. To view the old version click .
Symmetry
  • Article
  • Open Access

5 March 2020

Convergence and Dynamics of a Higher-Order Method

,
,
,
,
and
1
Universidad de la Rioja, Av. de la Paz, 93-103, 26006 Logroño, La Rioja, Spain
2
Department of Mathematical Sciences, Cameron University, Lawton, OK 73505, USA
3
Escuela de Ciencias Físicas y Matemáticas, Universidad de las Americas, Avda. de los Granados y Colimes, Quito 170125, Ecuador
4
Universidad Internacional de la Rioja (UNIR), Av. de la Paz 137, 26006 Logroño, La Rioja, Spain
This article belongs to the Special Issue Iterative Numerical Functional Analysis with Applications

Abstract

Solving problems in various disciplines such as biology, chemistry, economics, medicine, physics, and engineering, to mention a few, reduces to solving an equation. Its solution is one of the greatest challenges. It involves some iterative method generating a sequence approximating the solution. That is why, in this work, we analyze the convergence in a local form for an iterative method with a high order to find the solution of a nonlinear equation. We extend the applicability of previous results using only the first derivative that actually appears in the method. This is in contrast to either works using a derivative higher than one, or ones not in this method. Moreover, we consider the dynamics of some members of the family in order to see the existing differences between them.

1. Introduction

Mathematics is always changing and the way we teach it also changes, as can be seen in the literature. Moreover, in advanced mathematics we need to use different alternatives since we all know the different problems that students encounter. In this paper, we present a study on iterative methods that can be used in postgraduate studies in order to teach them.
In the present work, we are focused on the problem of solving the equation
g ( x ) = 0 ,
giving an approximating solution x * , where g : Ω S S is differentiable and S = R or C . There exist several studies related to this problem, since we need to use iterative methods to find the solution. We refer the reader to the book by Petkovic et al. [] for a collection of relevant methods. The method of interest is in this case:
y n = x n g ( x n ) g ( x n ) , t n = y n g ( y n ) ( g ( x n ) + ρ g ( y n ) ) g ( x n ) ( g ( x n ) + ( ρ 2 ) g ( y n ) ) , x n + 1 = t n δ K ( t n ) g ( x n ) ,
where a starting point x 0 is chosen, parameters ρ , δ S , and
K ( t n ) = g ( x n ) + g ( x n ) ( t n y n ) 2 ( t n x n ) ( y n x n ) ( x n + 2 y n 3 t n ) + g ( t n ) ( t n y n ) ( x n t n ) x n + 2 y n 3 t n g [ x n ; y n ] ( t n x n ) 3 ( y n x n ) ( x n + 2 y n 3 t n ) .
If we consider only the first two steps of the method in Equation (2), we obtain the King’s class of methods. This method has an order of 4 []. However, Equation (2) has limited usage, since its convergence assumes the existence of fifth order derivatives not appearing in the method. Moreover, no computable error bounds on | | x n x * | | or uniqueness results are given. Furthermore, the initial point x 0 is a shot in the dark. As an example consider function
g ( x ) = x 3 ln x 2 + x 5 x 4 , x 0 x = 0 .
Then g ( x ) is unbounded on Ω = [ 1 2 , 3 2 ] . Hence, there is no guarantee that the method in Equation (2) converges to x * = 1 under the results in [].
Our technique can also be used to extend the applicability of other methods defined in [,,]. The novelty of our work, compared to other such as [,,,,,,,,,,,,,,], is that we give weaker conditions, only in the first derivative, to guarantee the convergence of the described method. Those conditions are given in Section 2 and the dynamical study appears in the Section 3.

2. Local Convergence Analysis

In this section we study the local convergence analysis of the method in Equation (2). If v I R and μ > 0 , we can define U ( v , μ ) and U ¯ ( v , μ ) , respectively, the open and closed balls in R. Besides, we require the parameters L 0 > 0 , L > 0 , M 0 > 0 , M > 0 , γ 0 > 0 , μ , and δ I R . We need to define some parameters and functions to analyze the local convergence. Consider functions on the interval 0 , 1 L 0 by
g 1 ( t ) = L t 2 ( 1 L 0 t )
g 1 ¯ ( t ) = L 0 2 t + M | ρ 2 | g 1 ( t )
h 1 ( t ) = g 1 ¯ ( t ) 1
Then, h 1 ( 0 ) = 1 < 0 and h 1 ( t ) + as t 1 L 0 . Function h 1 has zeros in the interval 0 , 1 L 0 by the intermediate value theorem. Let r 1 be the smallest such zero. Define functions g 2 and h 2 on the interval [ 0 , r 1 ) by
g 2 ( t ) = 1 + M 2 ( 1 + | ρ | g 1 ( t ) ) ( 1 L 0 t ) ( 1 g ( t ) ) g 1 ( t )
and
h 2 ( t ) = g 2 ( t ) 1 .
By these definitions, h 2 ( 0 ) = 1 < 0 and h 2 ( t ) + as t r 1 . For this reason, function h 2 has a smallest zero r 2 ( 0 , r 1 ) . Moreover, define functions on [ 0 , r 2 ) by
g 3 ( t ) = g 2 ( t ) + M | δ | [ 1 1 L 0 t + β γ M 3 g 1 2 ( t ) ( 1 + | ρ | g 1 ( t ) ) ( 1 L 0 t ) ( 1 L 0 2 t ) ( 1 g 1 ¯ ( t ) ) 2 + β γ M 2 g 1 ( t ) ( 1 + | ρ | g 1 ( t ) ) ( 1 L 0 t ) 2 ( 1 g 1 ¯ ( t ) ) + β M 0 M 3 ( 1 + | ρ 1 | g 1 ( t ) + | ρ | g 1 2 ( t ) ) 2 ( 1 L 0 t ) 2 ( 1 L 0 2 t ) ( 1 g 1 ¯ ( t ) ) 2 ] ,
and
h 3 ( t ) = g 3 ( t ) 1 .
Suppose that
M | δ | ( 1 + β M 0 M 3 ) < 1 .
We can see that h 3 ( 0 ) = M | δ | ( 1 + β M 0 M 3 ) 1 < 0 and h 3 ( t ) + as t r 1 .
Characterize by r 3 the smallest such zero of h 3 ( t ) = 0 in ( 0 , r 1 1 ) . Set r = min { r 1 , r 2 , r 3 } . Then, we have that
0 g 1 ( t ) < 1
0 g 1 ¯ ( t ) < 1
0 g 2 ( t ) < 1
and
0 g 3 ( t ) < 1 for each t [ 0 , r ) .
We can express the method in Equation (2) in a different way as
y n = x n g ( x n ) g ( x n ) t n = y n g ( y n ) g ( x n ) ( g ( x n ) + ρ g ( y n ) ) ( g ( x n ) + ( ρ 2 ) g ( y n ) ) x n + 1 = t n δ ( A n + B n + C n + D n ) ,
provided
A n = g ( x n ) g ( x n ) , B n = ( t n y n ) 2 ( t n x n ) ( y n x n ) ( x n + 2 y n 3 t n ) , C n = g ( t n ) ( t n y n ) ( x n t n ) g ( x n ) ( x n + 2 y n 3 t n )
and
D n = g [ x n ; y n ] ( t n x n ) 3 g ( x n ) ( y n x n ) ( x n + 2 y n 3 t n ) .
Moreover, by simple algebraic manipulations we can write the previous B n , C n and D n in view of the definition of x n , y n , t n , B n , and C n as
B n = g ( y n ) 2 ( g ( x n ) + ρ g ( y n ) ) 2 ( f 2 ( x n ) + ( ρ 1 ) g ( x n ) g ( y n ) + ρ f 2 ( y n ) ) g ( x n ) ( g ( x n ) + ( ρ 2 ) g ( y n ) ) 2 ( f 2 ( x n ) + ( ρ + 1 ) g ( x n ) g ( y n ) + 3 ρ f 2 ( y n ) )
C n = g ( t n ) g ( y n ) ( g ( x n ) + ρ g ( y n ) ) ( f 2 ( x n ) + ( ρ 1 ) g ( x n ) g ( y n ) + ρ f 2 ( y n ) ) g ( x n ) ( g ( x n ) + ( ρ 2 ) g ( y n ) ) ( f 2 ( x n ) + ( ρ + 1 ) g ( x n ) g ( y n ) + 3 ρ f 2 ( y n ) )
and
D n = g [ x n ; y n ] ( f 2 ( x n ) + ( ρ 1 ) g ( x n ) g ( y n ) + ρ f 2 ( y n ) ) 3 g ( x n ) g ( x n ) 2 ( g ( x n ) + ( ρ 2 ) g ( y n ) ) 2 ( f 2 ( x n ) + ( ρ + 1 ) g ( x n ) g ( y n ) + 3 ρ f 2 ( y n ) ) ·
Next, we can give the local convergence result for the method in Equation (2) using the preceding notation.
Theorem 1. 
Let D I R be a convex subset and f : D I R a differentiable function. Consider the divided difference of order one g [ . ; . ] : D × D L ( D ) , x * D and for each x , y D the constants L 0 > 0 , L > 0 , M 0 > 0 , M > 0 , γ > 0 , ρ, δ I R , and β > 1 such that
M | δ | ( 1 + β M 0 M 3 ) < 1 ,
0 < ρ 0 ρ , max { ρ 0 , 3 2 2 } ρ 3 + 2 2 ,
g ( x * ) = 0 , g ( x * ) 0 , | | g ( x * ) 1 | | γ ,
| | g ( x * ) 1 ( g ( x ) g ( x * ) ) | | L 0 | | x x * | | ,
| | g ( x * ) 1 ( g ( x ) g ( y ) ) | | L | | x y | | ,
| | g ( x * ) 1 g ( x ) | | M
| | g ( x * ) 1 g [ x ; y ] | | M 0
and
U ¯ ( x * , r ) D ,
where the radius r is given previously
ρ 0 = ( 1 + β ) 2 ( β 1 ) ( 2 2 ( β 1 ) ( 3 β 1 ) + ( 5 β 3 ) ) .
Then, for x 0 U ( x * , r ) { x * } the method in Equation (2) generates a well defined sequence { x n } , all its terms are in U ( x * , r ) ( n = 0 , 1 , 2 , ), and the sequence converges to x * . Furthermore, the following estimates are verified
| | y n x * | | g 1 ( | | x n x * | | ) | | x n x * | | < | | x n x * | | < r ,
| | t n x * | | g 2 ( | | x n x * | | ) | | x n x * | | < | | x n x * | |
and
| | x n + 1 x * | | g 3 ( | | x n x * | | ) | | x n x * | | < | | x n x * | | ,
using the functions g 1 , g 2 and g 3 is defined above Theorem 1. Besides, x * is the only solution of F ( x ) = 0 in U ¯ ( x * , T ) for T r , 2 L 0 such that U ¯ ( x * , T ) D .
Proof. 
We shall show estimates of Equations (21)–(23) using mathematical induction. We get, through hypothesis x 0 U ( x * , r ) { x * } , the definition of r and Equation (15) that
| | g ( x * ) 1 ( g ( x 0 ) g ( x * ) ) | | L 0 | | x 0 x * | | < L 0 r < 1 .
From the Banach Lemma on invertible operators and Equation (24) it follows that g ( x 0 ) is invertible and
| | g ( x 0 ) 1 g ( x * ) | | 1 1 L 0 | | x 0 x * | | ·
Then, y 0 is well defined by the method in Equation (2) for n = 0 . Then, we can write
y 0 x * = x 0 x * g ( x 0 ) 1 g ( x 0 )
So, we get, using Equations (4), (16), (25), and (26) that
| | y 0 x * | | | | g ( x 0 ) 1 g ( x * ) | | 0 1 g ( x * ) 1 ( g ( x * + θ ( x 0 x * ) ) g ( x 0 ) ) ( x 0 x * ) d θ L | | x 0 x * | | 2 2 ( 1 L 0 | | x 0 x * | | ) = g 1 ( | | x 0 x * | | ) | | x 0 x * | | < | | x 0 x * | | < r ,
which shows y 0 U ( x * , r ) and Equation (21) for n = 0 .
Using Equation (14), we can express
g ( x 0 ) = g ( x 0 ) g ( x * ) = 0 1 g ( x * + θ ( x 0 x * ) ) ( x 0 x * ) d θ .
We get, in view of Equations (17) and (28), that
| | g ( x * ) 1 g ( x 0 ) | | = 0 1 g ( x * ) 1 g ( x * + θ ( x 0 x * ) ) ( x 0 x * ) d θ M | | x 0 x * | | ,
and similarly
| | g ( x * ) 1 g ( y 0 ) | | M | | y 0 x * | | .
At this point, we have that, by Equation (18),
| | g ( x * ) 1 g [ x 0 ; y 0 ] | | M 0 .
Hence, g ( x 0 ) + ( ρ 2 ) g ( y 0 ) is invertible. We get, using Equations (5), (14), (15), (27), and (30), that
| | ( g ( x * ) ( x 0 x * ) ) 1 ( g ( x 0 ) g ( x * ) g ( x * ) ( x 0 x * ) + ( ρ 2 ) g ( y 0 ) ) | | | | x 0 x * | | 1 0 1 g ( x * ) 1 g ( x * + θ ( x 0 x * ) g ( x * ) ) ( x 0 x * ) d θ | | + | | g ( x * ) 1 g ( y 0 ) | | x 0 x * | | 1 L 0 2 | | x 0 x * | | 2 + | ρ 2 | M | | y 0 x * | | | | x 0 x * | | 1 L 0 2 | | x 0 x * | | 2 + | ρ 2 | M g 1 ( | | x 0 x * | | ) | | x 0 x * | | = g 1 ¯ ( | | x 0 x * | | ) < 1 .
Then, by Equation (32) the function g ( x 0 ) + ( ρ 2 ) g ( y 0 ) is invertible and
| | ( g ( x 0 ) + ( ρ 2 ) g ( y 0 ) ) 1 g ( x * ) | | 1 | | x 0 x * | | ( 1 g 1 ¯ ( | | x 0 x * | | ) ) ·
It also follows that t 0 is well defined from the method in Equation (2) for n = 0 . Then, using the method in Equation (2) for n = 0 , Equations (6), (25), (29), (30), and (33), we have that
| | t 0 x * | | | | y 0 x * | | + g ( x * ) 1 g ( y 0 ) g ( x * ) 1 ( g ( x 0 ) + ρ g ( y 0 ) ) g ( x * ) 1 g ( x 0 ) g ( x * ) 1 ( g ( x 0 ) + ( ρ 2 ) g ( y 0 ) ) | | y 0 x * | | + M 2 | | y 0 x * | | ( | | x 0 x * | | + | ρ | | | y 0 x * | | ) ( 1 L 0 | | x 0 x * | | ) | | x 0 x * | | ( 1 g 1 ¯ ( | | x 0 x * | | ) ) 1 + M 2 ( | | x 0 x * | | + | ρ | g 1 ( | | x 0 x * | | ) | | x 0 x * | | ) ( 1 L 0 | | x 0 x * | | ) | | x 0 x * | | ( 1 g 1 ¯ ( | | x 0 x * | | ) ) | | y 0 x * | | = g 2 ( | | x 0 x * | | ) | | x 0 x * | | < | | x 0 x * | | < r ,
which shows Equation (22) for t 0 U ( x * , r ) and n = 0 . Next, we define estimates on | | A 0 | | , | | B 0 | | , | | C 0 | | , and | | D 0 | | . Assume that g ( x 0 ) 0 . We take into account the expressions f 2 ( x 0 ) + ( ρ + 1 ) g ( x 0 ) g ( y 0 ) + 3 ρ f 2 ( y 0 ) and f 2 ( x 0 ) + ( ρ 1 ) g ( x 0 ) g ( y 0 ) + ρ f 2 ( y 0 ) as quadratic polynomials in g ( y 0 ) (or g ( x 0 ) ). At this point, the discriminants are formed, respectively, by ( ρ 2 10 ρ + 1 ) f 2 ( x 0 ) and ( ρ 2 6 ρ + 1 ) f 2 ( x 0 ) , which are less than zero by Equation (13). Consequently,
f 2 ( x 0 ) + ( ρ + 1 ) g ( x 0 ) g ( y 0 ) + 3 ρ f 2 ( y 0 ) > 0
and
f 2 ( x 0 ) + ( ρ 1 ) g ( x 0 ) g ( y 0 ) + ρ f 2 ( y 0 ) > 0 .
Then, x 1 is well defined. Besides, we have, by Equations (35) and (36), that
f 2 ( x 0 ) + ( ρ 1 ) g ( x 0 ) g ( y 0 ) + ρ f 2 ( y 0 ) f 2 ( x 0 ) + ( ρ + 1 ) g ( x 0 ) g ( y 0 ) + 3 ρ f 2 ( y 0 ) = f 2 ( x 0 ) + ( ρ 1 ) g ( x 0 ) g ( y 0 ) + ρ f 2 ( y 0 ) f 2 ( x 0 ) + ( ρ + 1 ) g ( x 0 ) g ( y 0 ) + 3 ρ f 2 ( y 0 ) β ,
so Equation (37) reduces to exhibiting that for λ = g ( y 0 ) g ( x 0 )
φ ( λ ) 0 ,
where
φ ( t ) = ( 1 3 β ) ρ t 2 + ( ρ 1 β ( ρ + 1 ) ) t + 1 β .
The inequality of Equation (38) is satisfied for all t I R , if 1 3 β < 0 (i.e., β > 1 3 ), and the discriminant Δ of φ is
Δ 0
or
ψ ( ρ ) 0 ,
where,
ψ ( t ) = ( β 1 ) 2 t 2 2 ( β 1 ) ( 5 β 3 ) t + ( β + 1 ) 2 .
But the discriminant Δ 1 of ψ is given by
Δ 1 = 32 ( β 1 ) 3 ( 3 β 1 ) > 0 , if β > 1 .
Moreover, we have ( β 1 ) ( 5 β 3 ) > 0 for β > 1 . Then, by the Descartes rule of signs ψ has two positive zeros. Denote by ρ 0 the smallest such zero, which can be given in closed form using the quadratic formula to arrive at the definition of ρ 0 given in Equation (20). Hence, Equation (38) holds for all λ I R provided that Equation (13) is satisfied. Then, x 1 is well defined by the method in Equation (2) for n = 0 and we get, using Equations (8), (25), and (29), that
| | A 0 | | M | | x 0 x * | | 1 L 0 | | x 0 x * | | .
Next, using inequalities from Equations (9), (14), (25), (27), (30), (33), and (37), we have that
| | B 0 | | γ β M 4 | | y 0 x * | | ( | | x 0 x * | | + | ρ | | | y 0 x * | | ) 2 | | x 0 x * | | ( 1 L 0 | | x 0 x * | | 2 ) | | x 0 x * | | 2 ( 1 g 1 ¯ ( | | x 0 x * | | ) ) 2 ( 1 L 0 | | x 0 x * | | ) γ β M 4 g 1 2 ( | | x 0 x * | | ) ( 1 + | ρ | g 1 ( | | x 0 x * | | ) ) | | x 0 x * | | ( 1 L 0 2 | | x 0 x * | | ) ( 1 g 1 ¯ ( | | x 0 x * | | ) ) 2 ( 1 L 0 | | x 0 x * | | ) .
Through Equations (10), (14), (25), (27), (30), (33), and (37), we obtain that
| | C 0 | | γ β M 3 | | y 0 x * | | ( | | x 0 x * | | + | ρ | | | y 0 x * | | ) ( 1 L 0 | | x 0 x * | | ) 2 | | x 0 x * | | ( 1 g 1 ¯ ( | | x 0 x * | | ) ) γ β M 3 g 1 ( | | x 0 x * | | ) ( 1 + | ρ | g 1 ( | | x 0 x * | | ) ) | | x 0 x * | | ( 1 L 0 | | x 0 x * | | ) 2 ( 1 g 1 ¯ ( | | x 0 x * | | ) )
Then, we obtain, by Equations (11), (14), (18), (25), (27), (30), (33), and (37) that
| | D 0 | | β M 0 M 4 ( | | x 0 x * | | 2 + | ρ 1 | | | x 0 x * | | | | y 0 x * | | + | ρ | | | y 0 x * | | 2 ) 2 ( 1 L 0 | | x 0 x * | | ) 2 | | x 0 x * | | 3 ( 1 L 0 | | x 0 x * | | 2 ) ( 1 g 1 ¯ ( | | x 0 x * | | ) ) 2 β M 0 M 4 ( 1 + | ρ 1 | g 1 ( | | x 0 x * | | ) + | ρ | g 1 2 ( | | x 0 x * | | ) ) 2 | | x 0 x * | | ( 1 L 0 | | x 0 x * | | ) 2 ( 1 L 0 | | x 0 x * | | 2 ) ( 1 g 1 ¯ ( | | x 0 x * | | ) ) 2
Then, using Equations (7), (34) and (39)–(42), and the method in Equation (2) for n = 0 , we obtain that
| | x 1 x * | | | | t 0 x * | | + | δ | ( | | A 0 | | + | | B 0 | | + | | C 0 | | + | | D 0 | | ) g 3 ( | | x 0 x * | | ) | | x 0 x * | | < | | x 0 x * | | < r ,
which shows the inequality in Equation (23) for n = 0 . By simply replacing x 0 , y 0 , t 0 , x 1 by x k , y k , t k , x k + 1 in the previous estimates we have that the estimates in Equations (21)–(23) hold. Then, from | | x k + 1 x * | | a , a = g 3 ( | | x 0 x * | | ) [ 0 , 1 ) , | | x k x * | | < r , we get that lim k x k = x * and x k + 1 U ( x * , r ) . At last, to show the uniqueness part, let us suppose that there exists y * U ¯ ( x * , T ) with g ( y * ) = 0 . Define Q = 0 1 g ( y * + θ ( x * y * ) ) d θ . We get, using Equation (15), that
| | g ( x * ) 1 ( Q g ( x * ) ) | | L 0 0 1 | | y * + θ ( x * y * ) x * | | d θ L 0 0 1 ( 1 θ ) | | x * y * | | d θ = L 0 2 T < 1 .
It follows from Equation (44) that Q is invertible. Then, we conclude that x * = y * from the estimate 0 = g ( x * ) g ( y * ) = Q ( x * y * ) . □
As an example, consider g ( x ) = e * 1 , Ω = U ( 0 , 1 ) with x * = 0 and g [ x ; y ] = 0 1 g ( y + θ ( x y ) ) d θ . Then, we have γ = 1 , L 0 = e 1 , L = e , M = e , M 0 = e 2 . Then choosing ρ = 2.0 , δ = 0.001 , β = 1.01 conditions Equations (12) and (13) are satisfied. The radii obtained are: r 1 = 1.16395 , r 2 = 0.0585945 , and r 3 = 0.0530484 , so as a consequence we obtain r = r 3 = 0.0530484 .

3. Dynamical Analysis

In this section, the method in Equation (2) is applied to three different families of functions and its behavior has been analyzed by changing the δ parameter using techniques that appear in [,,,].

3.1. Exponential Family

The method has been applied to the function g ( x ) = e x 1 by considering the corresponding equation g ( x ) = 0 . This equation has a solution at the point x = 0 , which is the only one attractive fixed point of the method. In Figure 1 we observe how the method changes with the δ parameter. Dynamical planes represent the behavior of the method in the complex domain.
Figure 1. Method Representation. (a) ρ = 0.01 δ = 0. (b) ρ = 0.01 δ = 0.01. (c) ρ = 0.01 δ = 0.1. (d) ρ = 0.01 δ = 1.
In Figure 2 the symmetry of the region of convergence to the solution x = 0 with respect to the imaginary axis is observed. Small islands of convergence appear out of the main region. It is necessary to increase the maximum number of iterations to achieve convergence with high values of δ .
Figure 2. Dynamical planes associated with the method. (a) ρ = 0.01 δ = 0.01 maxiter = 10. (b) ρ = 0.01 δ = 0.1 maxiter = 10. (c) ρ = 0.01 δ = 1 maxiter = 10. (d) ρ = 0.01 δ = 1 maxiter = 20.

3.2. Sinus Family

The method can be applied to the function g ( x ) = sin ( x ) with equation g ( x ) = 0 . In this case, the equation has a periodical solutions x = π , x = 0 , x = π , and , coinciding with the fixed points of the method. In Figure 3 how the method changes with the δ parameter is shown. Dynamical planes represent the behavior of the method in the complex domain as it appears in Figure 4, where with high values of δ the region of convergence is reduced.
Figure 3. Method Representation. (a) ρ = 0.01 δ = 0. (b) ρ = 0.01 δ = 0.01. (c) ρ = 0.01 δ = 0.1. (d) ρ = 0.01 δ = 1.
Figure 4. Dynamical planes associated with the method. (a) ρ = 0.01 δ = 0.01 maxiter = 10. (b) ρ = 0.01 δ = 0.1 maxiter = 10. (c) ρ = 0.01 δ = 1 maxiter = 10. (d) ρ = 0.01 δ = 10 maxiter = 10.

3.3. Polynomial Family

The method was applied to the function g ( x ) = ( x 1 ) ( x + 1 ) with equation g ( x ) = 0 . The sink fixed points obtained in this case using the method are x = 1 , x = 1 , the solutions of the previous equation. In Figure 5 how the method changes with the δ parameter is shown. Dynamical planes represent the behavior of the method in the complex domain as it appears in Figure 6, where with larger values of δ the region of convergence is more complex.
Figure 5. Method Representation. (a) ρ = 0.01 δ = 0. (b) ρ = 0.01 δ = 0.01. (c) ρ = 0.01 δ = 0.1. (d) ρ = 0.01 δ = 1.
Figure 6. Dynamical planes associated with the method. (a) ρ = 0.01 δ = 0.01 maxiter = 10. (b) ρ = 0.01 δ = 0.1 maxiter = 10. (c) ρ = 0.01 δ = 1 maxiter = 10. (d) ρ = 0.01 δ = 10 maxiter = 25.

4. Conclusions

The study of high-order iterative methods is very important, since problems from all disciplines require the solution of some equations. This solution is found as the limit of sequences generated by such methods, since closed form solutions can rarely be found in general. The convergence order is usually found in the literature using expensive Taylor expansions, high-order derivatives, and without computable error estimates on | | x n x * | | or uniqueness results. It is worth noticing that the high-order derivatives do not appear in these methods. Moreover, the initial point is a “shot in the dark”. Hence, the applicability of these methods is very limited. To address all these problems, we have developed a technique using hypotheses only on the first derivative that actually appears on the method and Lipschitz-type conditions. This allows us to: extend the applicability of the method; find a radius of convergence as well as computable error estimates and uniqueness results based on Lipschitz constants. Although we demonstrated our technique on the method in Equation (2), clearly it can be used to extend the applicability of other methods along the same lines. In view of the involvement of parameters on the method, the dynamics of it have also been explored in many interesting cases.

Author Contributions

All authors have equally contributed to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

Research supported in part by Programa de Apoyo a la investigación de la fundación Séneca–Agencia de Ciencia y Tecnología de la Región de Murcia 20928/PI/18 and by Spanish MINECO project PGC2018-095896-B-C21.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Petković, M.S.; Neta, B.; Petković, L.D.; Dźunić, J. Multipoint Methods for Solving Nonlinear Equations: A Survey; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  2. Hueso, J.L.; Martinez, E.; Teruel, C. Convergence, efficiency and dynamics of new fourth and sixth order families of iterative methods for nonlinear systems. J. Comput. Appl. Math. 2015, 275, 412–420. [Google Scholar] [CrossRef]
  3. Behl, R.; Kanwar, V.; Kim, Y.I. Higher-order families of multiple root finding methods suitable for non-convergent cases and their dynamics. Math. Model. Anal. 2019, 24, 422–444. [Google Scholar] [CrossRef]
  4. Amat, S.; Busquier, S.; Gutiérrez, J.M. Geometric constructions of iterative functions to solve nonlinear equations. J. Comput. Appl. Math. 2003, 157, 197–205. [Google Scholar] [CrossRef]
  5. Argyros, I.K. Computational Theory of Iterative Methods. Series: Studies in Computational Mathematics, 15; Chui, C.K., Wuytack, L., Eds.; Elsevier Publ. Co.: New York, NY, USA, 2007. [Google Scholar]
  6. Argyros, I.K.; Magreñán, Á.A. Iterative Methods and Their Dynamics with Applications: A Contemporary Study; CRC Press: Boca Raton, FL, USA; Taylor & Francis Group: Abingdon, UK, 2017. [Google Scholar]
  7. Argyros, I.K.; Magreñán, Á.A. A Contemporary Study of Iterative Methods: Convergence, Dynamics and Applications; Elsevier: Amsterdam, The Netherlands, 2017. [Google Scholar]
  8. Argyros, I.K.; Hilout, S. Computational Methods in Nonlinear Analysis. Efficient Algorithms, Fixed Point Theory and Applications; World Scientific: Singapore, 2013. [Google Scholar]
  9. Argyros, I.K.; Hilout, S. Numerical Methods in Nonlinear Analysis; World Scientific Publ. Comp.: Hackensack, NJ, USA, 2013. [Google Scholar]
  10. Kantorovich, L.V.; Akilov, G.P. Functional Analysis; Pergamon Press: Oxford, UK, 1982. [Google Scholar]
  11. Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  12. Rheinboldt, W.C. An adaptive continuation process for solving systems of nonlinear equations, Polish Academy of Science. Banach Ctr. Publ. 1978, 3, 129–142. [Google Scholar] [CrossRef]
  13. Sharma, J.R. Improved Chebyshev–Halley methods with sixth and eighth order of convergence. Appl. Math. Comput. 2015, 256, 119–124. [Google Scholar] [CrossRef]
  14. Sharma, R. Some fifth and sixth order iterative methods for solving nonlinear equations. Int. J. Eng. Res. Appl. 2014, 4, 268–273. [Google Scholar]
  15. Traub, J.F. Iterative Methods for the Solution of Equations; Prentice–Hall Series in Automatic Computation: Englewood Cliffs, NJ, USA, 1964. [Google Scholar]
  16. Madhu, K.; Jayaraman, J. Higher Order Methods for Nonlinear Equations and Their Basins of Attraction. Mathematics 2016, 4, 22. [Google Scholar] [CrossRef]
  17. Sanz-Serna, J.M.; Zhu, B. Word series high-order averaging of highly oscillatory differential equations with delay. Appl. Math. Nonlinear Sci. 2019, 4, 445–454. [Google Scholar] [CrossRef]
  18. Pandey, P.K. A new computational algorithm for the solution of second order initial value problems in ordinary differential equations. Appl. Math. Nonlinear Sci. 2018, 3, 167–174. [Google Scholar] [CrossRef]
  19. Magreñán, Á.A. Different anomalies in a Jarratt family of iterative root–finding methods. Appl. Math. Comput. 2014, 233, 29–38. [Google Scholar]
  20. Magreñán, Á.A. A new tool to study real dynamics: The convergence plane. Appl. Math. Comput. 2014, 248, 215–224. [Google Scholar] [CrossRef]
  21. Magreñán, Á.A.; Argyros, I.K. On the local convergence and the dynamics of Chebyshev-Halley methods with six and eight order of convergence. J. Comput. Appl. Math. 2016, 298, 236–251. [Google Scholar] [CrossRef]
  22. Lotfi, T.; Magreñán, Á.A.; Mahdiani, K.; Rainer, J.J. A variant of Steffensen-King’s type family with accelerated sixth-order convergence and high efficiency index: Dynamic study and approach. Appl. Math. Comput. 2015, 252, 347–353. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.