Abstract
A new parametric class of iterative schemes for solving nonlinear systems is designed. The third- or fourth-order convergence, depending on the values of the parameter being proven. The analysis of the dynamical behavior of this class in the context of scalar nonlinear equations is presented. This study gives us important information about the stability and reliability of the members of the family. The numerical results obtained by applying different elements of the family for solving the Hammerstein integral equation and the Fisher’s equation confirm the theoretical results.
1. Design of a Parametric Family of Iterative Methods
The need to find a solution of equations or systems of nonlinear equations of the form , where , , is present in many problems of applied mathematics as a basis for solving other more complex ones. In general, it is not possible to find the exact solution to this type of equations, so iterative methods are required in order to approximate the desired solution.
The essence of these methods is to find, through an iterative process and, from an initial approximation close to a solution , a sequence of approximations such that, under different requirements, .
It is well known that one of the most used iterative methods, due to its simplicity and efficiency, is Newton’s scheme, whose iterative expression is
where denotes the derivative or the Jacobian matrix of function F evaluated in the kth iteration . In addition, this method has great importance in the study of iterative methods because it presents quadratic convergence under certain conditions and has great accessibility, that is, the region of initial estimates for which the method converges is wide, at least for polynomials or polynomial systems.
Based on Newton-type methods and by using different procedures, many iterative schemes for solving have been presented in the last years. Refs. [1,2] compile many of the methods recently designed to solve this type of problem. These books give us good overviews about this area of research.
In this paper, we use a convex combination of the methods presented by Chun et al. in [3] and Maheswari in [4]. As the mentioned schemes are designed for nonlinear equations and they have as the first step Newton’s method, we use the following algebraic manipulation in order to extend the mentioned schemes to nonlinear systems:
Therefore, the parametric family of iterative methods for solving nonlinear systems that we propose has the following iterative expression:
where is the initial estimation, and is the divided difference operator defined as
The rest of the paper is organized as follows: Section 2 is devoted to analyze the convergence of family (2) in terms of the values of parameter . In Section 3, we study the dynamical behavior of the class on quadratic polynomials in the context of scalar equations. This study allows for selecting the members that are more stable in the family. In the numerical section, (Section 4), we apply the proposed class on different examples such as the Hammerstein integral equation and the Fisher’s equation in order to confirm the theoretical results obtained in Section 2 and Section 3. We finish the work with some conclusions and the references used in it.
2. Convergence Analysis
Let us consider function , differentiable in the convex set which contains a solution of the nonlinear equation . From the Genochi–Hermite formula (see [5]) of the divided difference operator
and by performing the Taylor’s expansion of on the point x and integrating, we obtain the following development:
which we will use in the proof of the following result, when the order of convergence of family is established.
Theorem 1.
Let be a sufficiently Fréchet differentiable function in a convex neighborhood D of , being . We suppose the Jacobian matrix is continuous and non-singular in . Then, taking an initial estimate close enough to , the sequence of iterates generated with family (2) converges to with the following error equation:
where , being the set of j-linear functions of bounded functions, and . In the particular case in which , the error equation is
and so the method has an order of convergence four.
Proof.
We consider the Taylor’s expansion of around :
where , and ,
In a similar way, the derivatives of around take the form:
From the development of around , we calculate the inverse
with and satisfying .
Therefore,
- ,
- ,
- ,
- .
Then, we obtain the error equation of the first step of the parametric family (2):
Substituting this expression in the Taylor expansion of around , we get:
Furthermore,
To obtain the development of the divided difference operator of (2), we use the Taylor series expansion of (4), considering in this case and, so, . Therefore, substituting (8) and (10) in (4), we obtain
To calculate the inverse of this operator, we search
with and satisfying .
Thus,
- ,
- ,
- ,
- .
Substituting the expressions (10), (14), (17), and (18) in the scheme (2), we get the error equation of the parametric family
Finally, from the error equation, we conclude that the parametric family (2) has order 3 for all and order 4 for , being in this last case the error equation
□
In the next section, we analyze the dynamical behavior of the parametric family (2) on quadratic scalar polynomials.
3. Complex Dynamics
The dynamical analysis of (2) is performed throughout this section in terms of complex analysis. The order of convergence is not the only important criterion to study when evaluating an iterative scheme. The validity of a method also depends on other aspects such as knowing how it behaves based on the initial estimates that are taken, that is, how wide the set of initial estimations is for which the method is convergent. For this reason, it is necessary to introduce several tools that allow for a more exhaustive study.
The analysis of the dynamics of a method is becoming one of the most investigated parts within the study of iterative methods since it allows for classifying the different iterative schemes, not only from the point of view of their speed of convergence, but also analyzing its behavior based on the initial estimate taken (see, for example, [6,7,8,9,10,11,12,13]). This study allows for visualizing graphically the set of initial approximations that converge to a given root or to points that are not roots of the equation. In addition, it provides important information about the stability and reliability of the iterative method.
In this paper, we focus on studying the complex dynamic of the parametric family (2) on quadratic polynomials of the form , where . For this study, we need to present the result called the Scaling Theorem, since it allows us to conjugate the dynamical behavior of one operator with the behavior associated with another, conjugated through an affine application, that is, our operator has the same stability on all quadratic polynomials. This result will be of great use to us since we can apply the Möbius transformation on the operator associated with our parametric family acting on , assuming that the conclusions obtained will be of general application for any quadratic polynomial used.
Theorem 2
(Scaling Theorem for family (2)). Let be an analytic function in the Riemann sphere and let be an affine transformation with . We consider , . Let and be the fixed point operators of the family (2) associated with the functions f and g, respectively, that is to say,
where and . Then, is analytically conjugated to through T, that is to say,
Proof.
Taking into account that , and , so
where , and
Therefore, substituting these equalities and simplifying, we have
then , that is to say, and are analytically conjugated by . □
Now, we can apply the Möbius transformation on the operator associated with the parametric family (2) in order to obtain an operator that does not depend on the constants a and b and, thus, be able to study the dynamical behavior of this family for any quadratic polynomial. The Möbius transformation, in this case, is and has the following properties:
The fixed-point rational operator of family (2) on has the expression
We can also deduce from (23) that the order of the methods for quadratic polynomials is 3 when and 4 when .
3.1. Fixed Points
The orbit of a point is defined (see, for example, [14,15]) as the set of the successive applications of the rational operator, i.e.,
The performance of the orbit of z is deduced attending to its asymptotic behavior. A point is said to be T-periodic if and , for . For , this point is a fixed point.
Therefore, a fixed point is one that is kept invariant by the operator , that is, it is one that satisfies the equation . All the roots of the quadratic polynomial are, of course, fixed points of the operator. However, it may happen that fixed points appear that do not correspond to any root; we call these points strange fixed points. These points are not desirable from a numerical point of view because when an initial estimate is taken that is in the neighborhood of a strange fixed point, there is a possibility that the numerical method will converge to it, that is, to a point that is not a solution of the equation. Strange fixed points often appear when iterative methods are analyzed and their presence can show the instability of the method.
Fixed points can be classified according to the behavior of the derivative operator on them; thus, a fixed point can be:
- Repulsor, if ;
- Parabolic, if ;
- Attracting, if ;
- Superattracting, if .
Moreover, the basin of attraction of an attracting fixed point is the set of initial guesses whose orbits tend to . Therefore, the set of points whose orbit tends to an attracting fixed point defines the Fatou set , while its complement is the Julia set .
In what follows, we study what are the fixed points of operator and their character depending on the value of parameter . The proof of the following result is straightforward, as it only needs to solve the equation .
Proposition 1.
By analyzing the equation , one obtains the following statements:
- (i)
- and are superattracting fixed points for each value of γ.
- (ii)
- is a strange fixed point when .
- (iii)
- the roots of polynomialwhich we denote by , where , are also strange fixed points for each value of γ.
We need the expression of the differentiated operator to analyze the stability of the fixed points and to obtain the critical points:
It is clear that 0 and ∞ are always superattracting fixed points because they come from the roots of the polynomial, and the order of the iterative methods is higher than 2, but the stability of the other fixed points can change depending on the values of parameter .
Proposition 2.
The character of the strange fixed point is as follows:
- (a)
- If , then is not a strange fixed point.
- (b)
- If or , then is an attracting point.
- (c)
- If and , then is an attracting point.
- (d)
- cannot be a superattracting point.
- (e)
- If and , then is a parabolic point.
- (f)
- In another case, is the repulsor.
Proof.
We obtain that
It is not difficult to check that cannot be 0, so cannot be a superattractor, and, when , is not a fixed point.
Now, we are going to study when is an attracting point. It is easy to check that is equivalent to . Rewriting the last expression, we obtain the following inequality:
Let us see when this inequality is verified. When , that is, , is an attracting point, so we obtain that is an attracting point when or . When we have , we need to satisfy for being a superattractor.
We are going to study when is a parabolic point. will be a parabolic point when , that is, is a parabolic point when and . □
Now, we establish the stability of the strange fixed points that are roots of the polynomial (24). To do this, we calculate these roots noting that this polynomial is a sixth degree symmetric polynomial, that is, it is a polynomial that can be reduced to a third degree one, and that satisfies the following properties:
- is not the root;
- if is the root, is also the root.
Performing the reduction of (24), we obtain:
where , and . Now, we calculate the roots of this polynomial and obtain:
To calculate the roots of polynomial (24) from the , , we undo the variable change since . Therefore, we obtain the roots of the sixth degree polynomial, which are conjugated two by two
Now, we study when the roots of the polynomial (24) are superattractors. For them, we solve for all , and we get the following relevant values of :
- ,
- ,
- ,
- .
Next, we are going to study the character of the fixed points by analyzing those values of close to the values of the parameter for which some is a supertractor. To do this, we study how behaves near the four previous values, and we obtain regions where some of the roots will be attractors. These regions are represented in Figure 1.
Figure 1.
Character of the roots of polynomial : (a) , (b) , (c) , (d) .
3.2. Critical Points
The relevance of knowing that the free critical points (that is, critical points different from the roots of the polynomial) is based on this known fact: each invariant Fatou component contains, at least, one critical point. Operator has as critical points , , , and the roots of the polynomial
which we denote by , where .
Let us remark that is a preimage of the fixed point . We can see that is a symmetric polynomial, so we can obtain the roots of obtaining roots of a polynomial of degree 3. The polynomial reduced of is the following one that we obtain analogously to the polynomial (24):
In order to calculate the roots z of , we need to obtain the roots of and apply the following expression to them . Thus, the roots of are conjugated.
Now, we are going to study the asymptotic behavior of the critical points to establish if there are different convergence basins than those generated by the roots. For the free critical point , we have , who is a strange fixed point, so the parameter plane associated with this critical point is not significative, since we know the stability of .
The other free critical points are roots of a polynomial that depends on ; for that, we draw the parameter planes. As we have that the roots are conjugated, we will only draw three planes. We use as an initial estimate a free critical point that depends on . We establish a mesh in the complex plane of points. Each point of the mesh corresponds to a parameter value. In each of them, the rational function is iterated to obtain the orbit of the critical point as a function of . If that orbit converges to or to in less than 40 iterations, that point of the mesh is painted red; otherwise, the point appears in black.
As we can see, there are many values of the parameter that would result in a method in which the free critical points converge to one of the two roots. As it is observed in Figure 2, they are located in the red area on the right side of the plane. Moreover, some black areas can be identified as the regions of stability of those fixed points that can be attracting, such as Figure 1b, whose stability region appears in black on the right side of Figure 2c.
Figure 2.
Parameter planes of .
Now, we select some stable (in red in parameter planes) and unstable values of (in black) in order to show their performance.
In the case of dynamical planes, the value of the parameter is fixed. Each point in the complex plane is considered as a starting point of the iterative scheme, and it is painted in different colors depending on the point that it has converged to. In this case, we paint in blue points what converged to ∞, and in orange points what converged to 0. These dynamical planes have been generated with a mesh of points and a maximum of 40 iterations per point. We mark strange fixed points with white circles, the fixed point z = 0 with a white star, and free critical points with white squares (again, the routines used appear in [6]).
One value of the parameter that would be an interesting value is because it is the only one that obtains order 4. In that case, we obtain the dynamical plane that we can see in Figure 3a. In this case, two free critical points are in each basin of attraction, and the strange fixed points are in the boundary of both basins of attraction, so they are repulsive. In that case, the method is stable, and, as we can see, almost every point converges to 0 or ∞ (Let us notice that, in practice, any initial estimation taken in the Julia set will converge to 0 or to ∞, due to the rounding error).
Figure 3.
Dynamical plane for and .
Other value for the parameter that we study is , Figure 3b. As we can see, this dynamical plane is similar to that of , but, in this case, we obtain less free critical points and less strange fixed points, due to the simplification of the rational function for this value of .
Carrying out numerous experiments, we have realized that the simplest dynamics is that of the methods with parameter and . Next, we will see other dynamical planes associated with other values of the parameter . Some of these planes do not have a bad dynamics, although it is not as simple as the previous ones. This is the case of , Figure 4b, or the case of , Figure 4a.
Figure 4.
Dynamical plane for and .
However, values such as , or present a dynamical plane with the same number of basins of attraction but with more complicated performance. We can see some of these dynamical planes in Figure 5a,b and Figure 6a. There are also parameter values for which the number of basins of attraction increases, for example, (Figure 6b). These cases should be avoided since our method may not converge to the roots and may end up converging to other points.
Figure 5.
Dynamical planes of
and .
Figure 6.
Dynamical planes of and .
4. Numerical Experiments
In this section, we compare different iterative methods of the parametric family (2), solving two classical problems of applied mathematics: the Hammerstein integral equation and the Fisher partial derivative equation. We are going to use elements for the proposed class for which we have studied the dynamical plane because we want to verify that, although some of them have complicated dynamics, they can be methods that give good numerical results.
For the computational calculations, Matlab R2020b with variable precision arithmetics with 1000 digits of mantissa is used. From an initial estimation , the different algorithms calculate iterations until the stoping criterium is satisfied.
For the different examples and algorithms, we compare the approximation obtained, the norm of the function in the last iterate, the norm of the distance between the last two approximations, the number of iterations needed to satisfy the required tolerance, the computational time and the approximate computational convergence order (ACOC), defined by Cordero and Torregrosa in [16], which has the following expression:
4.1. Hammerstein Equation
In this example, we consider the well-known Hammerstein integral equation (see [5]), which is given as follows:
where , and the kernel F is
We transform the above equation into a finite-dimensional nonlinear problem by using the Gauss–Legendre quadrature formula given as , where the nodes and the weights are determined for by the Gauss–Legendre quadrature formula. In this case, the nodes and the weights are in Table 1.
Table 1.
Weights and nodes of the Gauss–Legendre quadrature.
By denoting the approximations of by (), one gets the system of nonlinear equations:
where and
Starting from an initial approximation and with a tolerance of , we run the parametric family for different values of the parameter . The numerical results are shown in Table 2.
Table 2.
Hammerstein results for different parameters.
In all cases, we obtain as an approximation of the solution of Equation (25) the following vector , , , , , , .
In the case of the Hammerstein integral equation, we see that the numerical results of the parametric family (2) for different values of are quite similar. The main difference observed between the methods is that the ACOC for is 4, and, for the rest of the methods, it is approximately 3. On the other hand, we note that the method with needs to perform a larger number of iterations than the rest of the methods to satisfy the required tolerance, so the time it takes to approximate the solution is also longer. Finally, taking into account the columns that measure the error of the approximation, that is, and , we see that iterative methods that get lower errors are those associated with the parameters and . These results confirm the information obtained in the dynamical section.
4.2. Fisher Equation
In this second example, we are going to study the equation proposed in [17] by Fisher to model the diffusion process in population dynamics. The analytical expression of this partial derivative equation is as follows:
where is the diffusion constant, r is the level of growth of the species, and p is the carrying capacity.
In this case, we will study the Fisher equation for the values , , and in the spatial interval and with the initial condition and null boundary conditions.
We transform the problem we just described in a set of nonlinear systems by applying an implicit method of finite differences, providing the estimated solution in the instant from the estimated one in . We denote the spatial step by and the temporal step by , where is the final instant and and are the number of subintervals in x and t, respectively. Therefore, we define a mesh of the domain , composed of points , as follows:
Our objective is to approximate the solution of problem (26) in these points of the mesh, solving as many nonlinear systems as there are temporary nodes in the mesh. For this, we use the following finite differences:
We observe that, for the time step, we use first order backward divided differences and for the spatial step they are second order centered divided differences.
By denoting as the approximation of the solution at , and, by replacing it in the Cauchy problem, we get the system
for and . The unknowns of this system are that is, the approximations of the solution in each spatial node for the fixed instant .
In this example, we are going to work with the parameters , and . As we have said, it is necessary to solve as many systems as there are temporary nodes ; for each of these systems, we use the parametric family (2) to approximate its solution. Thus, starting from an initial approximation , , with a tolerance of , we execute the parametric family for different values of so that we get Table 3.
Table 3.
Fisher results for different parameters.
In all cases, we obtain as an approximation of the solution of problem (26) the following vector ,,,,,,,,, ,.
In this case, it can seen that the results are very similar, although there are subtle differences. For example, the method when uses a smaller number of iterations than the rest to satisfy the required tolerance, although this does not make it much faster than the rest of the methods since the difference in time is seconds. On the other hand, if we look at the time column, we can see that there is a method that stands out for its slowness; this is the case of . Again, we note that the ACOC of the methods roughly match the theoretical predictions made throughout the article. Observing the columns of the errors, we find similar results as well and that, in this case, having a higher tolerance than in the first example, no great differences are observed in these results.
5. Conclusions
A parametric family of iterative methods for solving nonlinear systems is presented. The dynamical analysis of the class on quadratic polynomials is done in order to select the members of the family with better stability properties. We prove that there exist a wide set of real and complex values of the parameter for which the corresponding methods are stable. That is, the set of initial estimations converging to the roots is very wide.In particular, we have stated that those procedures with , , and are especially stable, although some other ones can also show similar dynamical properties. Two numerical examples related to Hammerstein’s equation and Fisher’s equation allow us to confirm the theoretical results corresponding to the convergence and the stability of the proposed class.
Author Contributions
The individual contributions of the authors are as follows: conceptualization, J.R.T.; writing—original draft preparation, E.G.V. and P.T.-N.; validation, A.C.; numerical experiments, E.G.V. and P.T.-N. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by Ministerio de Ciencia, Innovación y Universidades PGC2018-095896-BC22 (MCIU/AEI/FEDER, UE).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Not applicable.
Acknowledgments
The authors would like to thank the anonymous reviewers for their useful comments that have improved the final version of this manuscript.
Conflicts of Interest
The authors declare that there is no conflict of interest regarding the publication of this paper.
References
- Petković, M.S.; Neta, B.; Petković, L.D.; Dǔnić, J. Multipoint Methods for Solving Nonlinear Equations; Elserier: Amsterdam, The Netherlands, 2013. [Google Scholar]
- Amat, S.; Busquier, S. (Eds.) Advances in Iterative Methods for Nonlinear Equations; Springer: Cham, Switzerland, 2016. [Google Scholar]
- Chun, C.; Kim, Y. Several New Third-Order Iterative Methods for Solving Nonlinear Equations. Acta Appl. Math. 2010, 109, 1053–1063. [Google Scholar] [CrossRef]
- Maheshwari, A.K. A fourth order iterative method for solving nonlinear equation. Appl. Math. Comput. 2009, 211, 383–391. [Google Scholar] [CrossRef]
- Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: Cambridge, MA, USA, 1970. [Google Scholar]
- Chicharro, F.I.; Cordero, A.; Torregrosa, J.R. Drawing dynamical and parameters planes of iterative families and methods. Sci. World J. 2013, 780153. [Google Scholar] [CrossRef] [PubMed]
- Hernández-Verón, M.A.; Magre nán, Ȧ.; Rubio, M.J. Dynamics and local convergence of a family of derivative-free iterative processes. J. Comput. Appl. Math. 2019, 354, 414–430. [Google Scholar]
- Chicharro, F.I.; Cordero, A.; Garrido, N.; Torregrosa, J.R. Generating root-finder iterative methods of second order: convergence and stability. Axioms 2019, 8, 55. [Google Scholar] [CrossRef]
- Lee, M.Y.; Kim, Y.I.; Neta, B. A generic family of optimal sixteenth-order multiple-root finders and their dynamics underlying purely imaginary extraneous fixed points. Mathematics 2019, 7, 562. [Google Scholar] [CrossRef]
- Chicharro, F.I.; Cordero, A.; Garrido, N.; Torregrosa, J.R. Generalized high-order classes for solving nonlinear systems and their applications. Mathematics 2019, 7, 1194. [Google Scholar] [CrossRef]
- Chicharro, F.I.; Cordero, A.; Garrido, N.; Torregrosa, J.R. Wide stability in a new family of optimal fourth-order iterative methods. Comput. Math. Methods 2019, 1, e1023. [Google Scholar] [CrossRef]
- Sharma, D.; Parhi, S.K. Local Convergence and Complex Dynamics of a Uni-parametric Family of Iterative Schemes. Int. J. Appl. Comput. Math. 2020, 6, 1–16. [Google Scholar] [CrossRef]
- Behl, R.; Bhalla, S.; Magreñán, Á.A.; Kumar, S. An efficient high order iterative scheme for large nonlinear systems with dynamics. Comput. Appl. Math. 2020, 113249. [Google Scholar] [CrossRef]
- Blanchard, P. Complex analitic dynamics on the Riemann splere. Bull. Am. Math. Soc. 1984, 11, 85–141. [Google Scholar] [CrossRef]
- Devaney, R.L. An Introduction to Chaotic Dynamical Systems; Addison-Wesley: Bostom, MA, USA, 1989. [Google Scholar]
- Cordero, A.; Torregrosa, J.R. Variants of Newton’s method using fifth-order quadrature formulas. Appl. Math. Comput. 2007, 190, 686–698. [Google Scholar] [CrossRef]
- Fisher, R.A. The wave of advance of advantageous genes. Ann. Eugen. 1937, 7, 353–429. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).