1. Introduction
We consider the problem of finding a simple zero
of a function
, defined in an open interval
I. This zero can be determined as a fixed point of some function
g by means of the one-point iteration method:
where
is the starting point. The most widely-used example of these kinds of methods is the classical Newton’s method given by:
It is well known that it converges quadratically to simple zeros and linearly to multiple zeros. In the literature, many modifications of Newton’s scheme have been published in order to improve its order of convergence and stability. Interesting overviews about this area of research can be found in [
1,
2,
3]. The works of Weerakoon and Fernando [
4] and, later, Özban [
5] have inspired a whole set of variants of Newton’s method, whose main characteristic is the use of different means in the iterative expression.
It is known that if a sequence
tends to a limit
in such a way that there exist a constant
and a positive integer
such that:
for
, then
p is called the order of convergence of the sequence and
C is known as the asymptotic error constant. For
, constant
C satisfies
.
If we denote by
the exact error of the
nth iterate, then the relation:
is called the error equation for the method and
p is the order of convergence.
Let us suppose that
is a sufficiently-differentiable function and
is a simple zero of
f. It is plain that:
Weerakoon and Fernando in [
4] approximated the definite integral (
5) by using the trapezoidal rule and taking
, getting:
and therefore, a new approximation
to
is given by:
Thus, this variant of Newton’s scheme can be considered to be obtained by replacing the denominator
of Newton’s method (
2) by the arithmetic mean of
and
. Therefore, it is known as the arithmetic mean Newton method (AN).
In a similar way, the arithmetic mean can be replaced by other means. In particular, the harmonic mean
, where
x and
y are two nonnegative real numbers, from a different point of view:
where since
, then
, i.e., the harmonic mean can be seen as a convex combination between
x and
y, where every element is given the relevance of the other one in the sum. Now, let us switch the roles of
x and
y; we get:
that is the contraharmonic mean between
x and
y.
Özban in [
5] used the harmonic mean instead of the arithmetic one, which led to a new method:
being
a Newton step, which he called the harmonic mean Newton method (HN).
Ababneh in [
6] designed an iterative method associated with this mean, called the contraharmonic mean Newton method (CHN), whose iterative expression is:
with third-order of convergence for simple roots of
, as well as the methods proposed by Weerakoon and Fernando [
4] and Özban [
5].
This idea has been used by different authors for designing iterative methods applying other means, generating symmetric fixed point operators. For example, Xiaojian in [
7] employed the generalized mean of order
between two values
x and
y defined as:
to construct a third-order iterative method for solving nonlinear equations. Furthermore Singh et al. in [
8] presented a third-order iterative scheme by using the Heronian mean between two values
x and
y, defined as:
Finally, Verma in [
9], following the same procedure, designed a third-order iterative method by using the centroidal mean between two values
x and
y, defined as:
In this paper, we check that all these means are functional convex combinations means and develop a simple test to prove easily the third-order of the corresponding iterative methods, mentioned before. Moreover, we introduce a new method based on the Lehmer mean of order
, defined as:
and propose a generalization that also satisfies the previous test. Finally, all these schemes are numerically tested, and their dependence on initial estimations is studied by means of their basins of attraction. These basins are shown to be clearly symmetric.
The rest of the paper is organized as follows:
Section 2 is devoted to designing a test that allows us to characterize the third-order convergence of the iterative method defined by a mean. This characterization is used in
Section 3 for giving an alternative proof of the convergence of mean-based variants of Newton’s (MBN) methods, including some new ones. In
Section 4, we generalize the previous methods by using the concept of
-means.
Section 5 is devoted to numerical results and the use of basins of attraction in order to analyze the dependence of the iterative methods on the initial estimations used. With some conclusions, the manuscript is finished.
2. Convex Combination
In a similar way as has been stated in the Introduction for the arithmetic, harmonic, and contraharmonic means, the rest of the mentioned means can be also regarded as convex combinations. This is not coincidental: one of the most interesting properties that a mean satisfies is the averaging property:
where
is any mean function of
x and
y nonnegative. This implies that every mean that satisfies this property is a certain convex combination among its terms.
Indeed, there exists a unique
such that:
This approach suggests that it is possible to generalize every mean-based variant of Newton’s method (MBN), by studying their convex combination counterparts. As a matter of fact, every mean-based variant of Newton’s method can be rewritten as:
where
. This is a particular case of a family of iterative schemes constructed in [
10].
We are interested in studying its order of convergence as a function of
. Thus, we need to compute the approximated Taylor expansion of the convex combination at the denominator and then its inverse:
where
. Then, its inverse can be expressed as:
Now,
and by replacing it in (
18), it leads to the MBN error equation as a function of
:
Equation (
22) can be used to re-discover the results of convergence: for example, for the contraharmonic mean, we have:
where:
so that:
Thus, we can obtain the
associated with the contraharmonic mean:
Finally, by replacing the previous expression in (
22):
and we obtain again that the convergence for the contraharmonic mean Newton method is cubic.
Regarding the harmonic mean, it is straightforward that it is a functional convex combination, with:
Replacing this expression in (
22), we find the cubic convergence of the harmonic mean Newton method,
In both cases, the independent term of
was
; it was not a coincidence, but an instance of the following more general result.
Theorem 1. Let be associated with the mean-based variant of Newton’s method (MBN):where M is a mean function of the variables and . Then, MBN converges, at least, cubically if and only if the estimate:holds. Proof. We replace
in the MBN error Equation (
22), obtaining:
Now, some considerations follow.
Remark 1. Generally speaking,where are real numbers. If we put (33) in (22), we have:it follows that, in order to attain cubic convergence, the coefficient of must bezero. Therefore, . On the other hand, to achieve a higher order (i.e., at least four), we need to solve the following system:This gives us that assure at least a fourth-order convergence of the method. However, none of the MBN methods under analysis satisfy these conditions simultaneously. Remark 2. The only convex combination involving a constant θ that converges cubically is , i.e., the arithmetic mean.
The most useful aspect of Theorem 1 is synthesized in the following corollary, which we call the “-test”.
Corollary 1 (θ-test) With the same hypothesis of Theorem 1, an MBN converges at least cubically if and only if the Taylor expansion of the mean holds: Let us notice that Corollary 1 provides a test to analyze the convergence of an MBN without having to find out the inherent , therefore sensibly reducing the overall complexity of the analysis.
3. New MBN by Using the Lehmer Mean and Its Generalization
The iterative expression of the scheme based on the Lehmer mean of order
is:
where
and:
Indeed, there are suitable values of parameter p such that the associated Lehmer mean equals the arithmetic one and the geometric one, but also the harmonic and the contraharmonic ones. In what follows, we will find it again, this time in a more general context.
By analyzing the associated
-test, we conclude that the iterative scheme designed with this mean has order of convergence three.
-Means
Now, we propose a new family of means of n variables, starting again from convex combinations. The core idea in this work is that, in the end, two distinct means only differ in their corresponding weights and . In particular, we can regard the harmonic mean as an “opposite-weighted”mean, while the contraharmonic one is a “self-weighted”mean.
This behavior can be generalized to
n variables:
is the contraharmonic mean among
n numbers. Equation (
43) is just a particular case of what we call
-mean.
Definition 1 (σ-mean) Given a vector of n real numbers and a bijective map (i.e., is a permutation of ), we call the σ-mean of order the real number given by: Indeed, it is easy to see that, in an
-mean, the weight assigned to each node
is:
where the equality holds because
is a permutation of the indices. We are, therefore, still dealing with a convex combination, which implies that Definition 1 is well posed.
We remark that if we take
, i.e., the identical permutation, in (
44), we find the Lehmer mean of order
m. Actually, the Lehmer mean is a very special case of the
-mean, as the following result proves.
Proposition 1. Given , the Lehmer mean of order m is the maximum σ-mean of order m.
Proof. We recall the rearrangement inequality:
which holds for every choice of
and
regardless of the signs, assuming that both
and
are sorted in increasing order. In particular,
and
imply that the upper bound is attained only for the identical permutation.
Then, to prove the result, it is enough to replace every
with the corresponding weight defined in (
45). ☐
The Lehmer mean and
-mean are deeply related: if
, as is the case of MBN, there are only two possible permutations, the identical one and the one that swaps one and two. We have already observed that the identical permutation leads to the Lehmer mean; however, if we express
in standard cycle notation as
, we have that:
We conclude this section proving another property of -means, which is that the arithmetic mean of all possible -means of n numbers equals the arithmetic mean of the numbers themselves.
Proposition 2. Given n real numbers and denoting the set of all possible permutations of , we have:for all . Proof. Let us rewrite Equation (
48); by definition, we have:
and we claim that the last equality holds. Indeed, we notice that every term in the sum of the
-means on the left side of the last equality involves a constant denominator, so we can multiply both sides by it and also by
to get:
Now, it is just a matter of distributing the product on the right in a careful way:
If we fix
, in
, there are exactly
permutations
such that
. Therefore, the equality in (
50) follows straightforwardly.
4. Numerical Results and Dependence on Initial Estimations
Now, we present the results of some numerical computations, in which the following test functions have been used.
- (a)
,
- (b)
,
- (c)
,
- (d)
,
- (e)
.
The numerical tests were carried out by using MATLAB with double precision arithmetics in a computer with processor i7-8750H @2.20 GHz, 16 Gb of RAM, and the stopping criterion used was .
We used the harmonic mean Newton method (HN), the contraharmonic mean Newton method (CHN), the Lehmer mean Newton method (LN(m)), a variant of Newton’s method where the mean is a convex combination with
,
N, and the classic Newton method (CN). The main goals of these calculations are to confirm the theoretical results stated in the preceding sections and to compare the different methods, with CN as a control benchmark. In
Table 1, we show the number of iterations that each method needs for satisfying the stopping criterion and also the approximated computational order of convergence, defined in [
11], with the expression:
which is considered as a numerical approximation of the theoretical order of convergence
p.
Regarding the efficiency of the MBN, we used the efficiency index defined by Ostrowski in [
12] as
, where
p is the order of convergence of the method and
d is the number of functional evaluations per iteration. In this sense, all the MBN had the same
; meanwhile, Newton’s scheme had the index
. Therefore, all MBN were more efficient than the classical Newton method.
The presented numerical tests showed the performance of the different iterative methods to solve specific problems with fixed initial estimations and a stringent stopping criterion. However, it is useful to know their dependence on the initial estimation used. Although the convergence of the methods has been proven for real functions, it is usual to analyze the sets of convergent initial guesses in the complex plane (the proof would be analogous by changing the condition on the function to be differentiable by being holomorphic). To get this aim, we plotted the dynamical planes of each one of the iterative methods on the nonlinear functions , , used in the numerical tests. In them, a mesh of initial estimations was employed in the region of the complex plane .
We used the routines appearing in [
13] to plot the dynamical planes corresponding to each method. In them, each point of the mesh was an initial estimation for the analyzed method on the specific problem. If the method reached the root in less than 40 iterations (closer than
), then this point is painted in orange (green for the second, etc.) color; if the process converges to another attractor different from the roots, then the point is painted in black. The zeros of the nonlinear functions are presented in the different pictures by white stars.
In
Figure 1, we observe that Harmonic and Lehmer (for
) means showed the most stable performance, whose unique basins of attraction were those of the roots (plotted in orange, red, and green). In the rest of the cases, there existed black areas of no convergence to the zeros of the nonlinear function
. Specially unstable were the cases of Heronian, convex combination (
), and generalized means, with wide black areas and very small basins of the complex roots.
Regarding
Figure 2, again Heronian, convex combination (
), and generalized means showed convergence only to one of the roots or very narrow basins of attraction. There existed black areas of no convergence to the roots in all cases, but the widest green and orange basins (corresponding to the zeros of
) corresponded to harmonic, contra harmonic, centroidal, and Lehmer means.
Function
had only one zero at
, whose basin of attraction is painted in orange color in
Figure 3. In general, most of the methods presented good performance; however, three methods did not converge to the root in the maximum of iterations required: Heronian and generalized means with
. Moreover, the basin of attraction was reduced when the parameter
of the convex combination mean was used.
A similar performance is observed in
Figure 4, where Heronian and generalized means with
showed no convergence to only the root of
; meanwhile, the rest of the methods presented good behavior. Let us remark that in some cases, blue areas appear; this corresponded to initial estimations that, after 40 consecutive iterations, had an absolute value higher than 1000. In these cases, they and the surrounding black areas were identified as regions of divergence of the method. The best methods in this case were associated with the arithmetic and harmonic means.
In
Figure 5, the best results in terms of the wideness of the basins of the attraction of the roots were for harmonic and Lehmer means, for
. The biggest black areas corresponded to convex combination with
, where the three basins of attraction of the roots were very narrow, and for Heronian and generalized means, there was only convergence to the real root.
5. Conclusions
The proposed -test (Corollary 1) has proven to be very useful to reduce the calculations of the analysis of convergence of any MBN. Moreover, though the employment of -means in the context of mean-based variants of Newton’s method is probably not the best one to appreciate their flexibility, their use could still lead to interesting results due to their much greater capability of interpolating between numbers than already powerful means, such as the Lehmer one.
With regard to the numerical performance,
Table 1 confirms that a convex combination with a constant coefficient could converge cubically if and only if it was the arithmetic mean; otherwise, as with this case, it converged quadratically, even if it may have done so with less iterations, generally speaking, than CN. Regarding the number of iterations, there were non-linear functions for which LN(m) converged with fewer iterations than HN. In our calculations, we set
, but similar results were achieved also for different parameters. Regarding the dependence on initial estimations, the harmonic and Lehmer methods were proven to be very stable, with the widest areas of convergence in most of the nonlinear problems used in the tests.