Abstract
In this paper, a two-point step-size gradient technique is proposed by which the approximate solutions of a non-linear system are found. The two-point step-size includes two types of parameters deterministic and random. A new adaptive backtracking line search is presented and combined with the two-point step-size gradient to make it globally convergent. The idea of the suggested method depends on imitating the forward difference method by using one point to estimate the values of the gradient vector per iteration where the number of the function evaluation is at most one for each iteration. The global convergence analysis of the proposed method is established under actual and limited conditions. The performance of the proposed method is examined by solving a set of non-linear systems containing high dimensions. The results of the proposed method is compared to the results of a derivative-free three-term conjugate gradient CG method that solves the same test problems. Fair, popular, and sensible evaluation criteria are used for comparisons. The numerical results show that the proposed method has merit and is competitive in all cases and superior in terms of efficiency, reliability, and effectiveness in finding the approximate solution of the non-linear systems.
Keywords:
conjugate gradient methods; performance profiles; backtracking line search; numerical comparisons and evaluation criteria MSC:
90C26
1. Introduction
The computational knowledge revolution has brought many developments in different fields, such as technical sciences, industrial engineering, economics, operations research, networks, chemical engineering, etc. Therefore, many new problems are being generated continuously. These problems need to be solved. A mathematical modelling technique is a critical tool that is utilized to formulate these problems as mathematical problems [1].
The result is these problems are formulated as unconstrained (UCN), constrained (CON), multi-objective optimization problems (MOOP), and systems of non-linear equations (NE).
In recent years, there is a considerable advance in the study of non-linear equations.
Finding solutions to non-linear systems which include a large number of variables is a hard task. Hence, several derivative-free methods have been suggested to face this challenge [2]. A system of non-linear equations exists in many fields of science and technology. Non-linear equations represent various sorts of implementations [3,4].
The applications of systems of non-linear equations are arising in different fields such as compressive sensing, chemical equilibrium systems, optimal power flow equations, and financial forecasting problems [5,6].
The non-linear system of equations is defined by
where the function is monotone and continuously differentiable, satisfying
The solution to Problem (1) can be generated by the classical methods, i.e., the new point is obtained by
where is the step-size obtained by an appropriate line search method with the search direction .
Many ideas have been implemented to obtain new proposed methods to design different formulas by which the step-size and search direction are generated. For example, the Newton method [7], quasi-Newton method [8], and the conjugate gradient (CG) method [9,10,11,12].
The Newton method, the quasi-Newton method, and their related methods are widely used to solve Problem (1); due to their fast convergence speeds, see [13,14,15,16,17]. These methods require computing the Jacobian matrix or the approximate Jacobian matrix per iteration; hence, they are very expensive.
Thus, it is more useful to sidestep using these methods to deal with large-scale non-linear systems of equations.
The CG algorithm is a popular classical methods, due to its simplicity, less storage, efficiency, and good convergence properties. The CG method has become more popular in solving non-linear equations [18,19,20,21,22,23].
Currently, many derivative-free iterative methods have been proposed to deal with the large-scale Problem (1) [24].
Cruz [25] proposed a new spectral method to deal with large-scale systems of non-linear monotone equations. The core idea of the method, presented in [25], can be briefly described as follows.
where is the spectral parameter defined by
where and , with sufficiently large, the author of [25] set . Furthermore, the author of [25] defined the step-size as follows , where is selected by a backtracking line search algorithm to make the following inequality holds:
where for .
The author of [25] used Lemma 3.3 from [26] to prove that the sequence converges. Furthermore, the author of [25] presented a convergence analysis for their algorithm and the numerical results show that the algorithm in is computationally efficient to solve Problem (1).
Stanimirovic et al. [27] proposed a new improved gradient descent iterations to solve Problem (1). They proposed new directions and step-size gradient methods by defining various parameters. Furthermore, they presented a theoretical convergence analysis of their methods under standard assumptions. In all four cases in [27], the direction vector is computed by using the system vector, i.e., they set , with several step-size and parameters proposed by the authors.
Waziri et al. [28] proposed some hybrid derivative-free algorithms to solve Problem (1). They presented several parameters by which two methods were designed to deal with a non-linear system of equations. In both proposed methods the direction vector is computed depending on the , i.e., the values of the non-linear system of equations (vector ) are used to represent the search direction for some cases with several proposed parameters.
The performance of both methods was examined by solving a set of test problems of non-linear equations. These numerical experiments show the efficiency of both proposed methods.
Abubakar et al. [5] presented a new hybrid spectral-CG algorithm to find approximate solutions to Problem (1).
In all four cases used in [5], the vector was utilized to compute the search direction .
Kumam et al. [29] proposed a new hybrid approach to deal with systems of non-linear equations applied to signal processing. The new approach suggested by the authors of [29] depends on hybridizing several parameters of the classical CG method by using a convex combination. Under sensible hypotheses, the authors of [29] proved the global convergence of their hybrid technique. Therefore, this new hybrid approach presented incredible results for solving monotone operator equations compared with other methods.
Dai and Zhu [19] proposed a modified Hestenes–Stiefel-type derivative-free (HS-CG method) to solve large-scale non-linear monotone equations. The HS-CG method was proposed by Hestenes and Stiefel [10]. The modified method presented by the authors of [19] is similar to the Dai and Wen [30] method for unconstrained optimization.
Furthermore, the authors of [19] presented a derivative-free new line search and step-size method. They established the convergence of their proposed method if the system of non-linear equations is Lipschitz, continuous and monotone. The numerical results obtained by this method demonstrate that the proposed approach in [19] is more efficient compared to other methods.
Ibrahim et al. [31] proposed a derivative-free iterative method to deal with non-linear equations with convex constraints using the projection technique. They also established a convergence analysis of the method under a few assumptions.
Thus, it can be said that there exists a consensus on estimating the value of the gradient vector in terms of the system vector, without using approximate values of the Jacobian matrix.
The hybridization of the gradient method with the random parameters (stochastic method) has been demonstrated to exhibit high efficiency in solving unconstrained and constrained optimization problems, see, for example, [32,33].
Therefore, the above briefly discusses the different ideas form previous literature by which the different formulas of step-size and search direction were designed. All of these motivated and guided us in the current research.
This study aims to use numerical differentiation methods to estimate the gradient vector by which the step-size can be computed.
There are many studies which estimate the gradient vector, see, for example, [34,35,36,37]. The common approaches to the numerical differentiation methods are the forward difference and the central difference [38,39,40,41,42].
Sundry’s advanced methods provided fair results when numerically calculating the gradient vector values. See, for example, [34,35,36,37,43,44,45].
Obtaining a good estimation of the gradient vector depends on the optimal selection of the finite-difference interval h.
For example, the authors of [43,44] proposed a random mechanism for selecting the optimal finite-difference interval h and presented a good results when solving unconstrained minimization problems.
However, when solving Problem (1), using the numerical differentiation methods to compute the approximate values of the Jacobian matrix is extremely expensive.
Thus, we use the core idea of the forward difference method without computing the Jacobian matrix or the gradient vector, i.e., our suggested method will benefit from the essence of forward difference method where the function evaluation is at most one for each iteration.
Therefore, the primary aim of this article is to develop a new algorithm that solves Problem (1).
The contributions of this study are as follows:
- Suggesting a two -step-size gradient method.
- Design two search directions and with different trust regions that possess the sufficient descent property, i.e., and with .
- Design an iterative formula by using the -step-size and direction by which the first candidate solution can be generated.
- Hybridization of the two-step-size gradient method, and with the direction to generate the second candidate solution.
- A modified line search technique is presented to make the hybrid method a globally convergent method.
- A convergence analysis of the proposed algorithm is established.
- Numerical experiments including solving a set of test problems as a non-linear system.
2. The Suggested Method
In this section, we propose a brand new mixed hybrid two-point step-size gradient algorithm to compute the approximate solutions to Problem (1). This proposed algorithm contains two search directions, two-point step-size, two formulas to generate the candida solutions, backtracking line search technique, and several parameters. The next section describes the proposed algorithm.
2.1. Algorithm
We generate two candidate solutions by two formulas as follows. The first one is defined by
where the search direction is defined as follows
A new step-size is defined by
The parameter is defined as follows
where the and are constants with , and the parameter is randomly generated at each iteration in the range and , the parameters and are computed in Section 2.1.1 and Section 2.1.2.
The second candidate solution is generated by the following algorithm.
Step 1: Perturb (shrink or extend) the solution obtained from Formula (7) by the following action
where is the candidate solution obtained from (7).
Step 2: Design a new research direction with a wider trust region than the trust region of the research direction defined in (8). Thus, the new research direction can be defined by
where the parameter is defined by
where is a constant number.
Step 3: Generate the second point by
Note 1: The perturbation (shrink or extend) in step 1 might occasionally move point (14) next to point (7).
Therefore, the second candidate solution might move far away or remain next to the first candidate solution.
Note 2: The second candidate solution obtained by (14) may give a better or worse result compared to the first candidate solution given by (7). Therefore, the best solution is picked from the two solutions obtained from (7) and (14).
Hence, the backtracking line search algorithm is applied to the best solution obtained to ensure the following inequality holds.
The line search in (15) imitates the line search in (6), proposed by [25] with the following changes.
The sequence is defined by
where k is the number of iterations at which the two candidate solutions from both Formulas (7) and (14) are generated. Furthermore, the step-size is defined by
where , .
Therefore, our step-size is designed in a different manner than the step-size proposed in [25].
We continue to reduce the step length by a dampening factor to guarantee Inequality (15).
Note: Without the effects of the damping factor , the step-size is bounded; therefore, it converges, as shown in Corollary 2, and it is clear that term is bounded.
In the following we present two algorithms to compute the scalers and .
The following two algorithms simulate the forward difference method, defined by
where and n is the number of variables.
2.1.1. First Approach (App1)
Step 1: , where denotes the number of components of variable .
Step 2: .
2.1.2. Second Approach (App2)
Step 1: Generate a new point next to point as follows, , where is a random vector generated in the range at each iteration.
Step 2: Compute , and set and .
Note 1: Point is the optimal solution that satisfies Inequality (15).
Note 2: From the idea presented in Algorithm 2 from [43], we randomly generate the values of .
The formulas and procedures mentioned above can be expressed as a set of steps to represent the proposed algorithm as follows.
Our proposed method is called the “hybrid random two-point step-size gradient method”, denoted by “HRSG”.
Therefore, two algorithms were designed to solve Problem (1), the first depends on the first approach (App1), abbreviated by “HRSG1”; while the second depends on the second approach (App2), abbreviated by “HRSG2”.
3. Convergence Analysis of the proposed algorithm
The following assumptions, lemmas and theorems are required to establish the global convergence of Algorithm 1.
| Algorithm 1 New proposed algorithm “HRSG” |
|
Assumption 1.
(1a) there are a set of solutions to System (1).
(1b) the set of non-linear systems F satisfies the Lipschitz condition, i.e., there exists a positive constant L such that
for all .
The next lemma shows that the proposed method “HRSG” has the sufficient descent property and trust region feature.
Lemma 1.
Let sequence be generated by Algorithm 1. Then, we have
and
where and .
Proof.
Since , , and , then
Therefore, (20) is true. Since , , and , then . Therefore, the proof is complete. □
Lemma 2.
Proof.
Since , then .
From (20), we obtain .
Case 1: If , then .
We set ; hence, , with .
Case 2: If , then
In this case, since and , then . Hence, we have
Therefore, the following inequality is true.
Then,
We set ; hence, in this case, , with .
Both cases satisfy with . Then the direction satisfies the sufficient descent condition.
Therefore, (22) is true.
Since , by using (21), we obtain
, and set . Then, (23) is true, with .
Therefore, the proof is complete. □
We adopt the concept from [16] in the following lemmas.
Lemma 3.
Solodov and Svaiter [16] assumed that satisfies . Thus, let the sequence be generated by Algorithm 1. Then,
Corollary 1.
Lemma 3 implies that the sequence is non-increasing, convergent and therefore bounded. Furthermore, is bounded and
Corollary 2.
Lemma 4.
Let be generated by Algorithm 1, then
Proof.
From (8), we have for , where . Then
Since and , then .
Therefore, the following is true
By using (26), then .
Since for all k (where is bounded), then (28) is true.
Therefore, Lemma 4 holds. □
Theorem 1.
Let be generated by Algorithm 1. Then
Proof.
By using proof contradiction, we assume that (32) does not hold true, that is, there exists a positive constant such that
By combing the sufficient descent condition (20) and the Cauchy–Schwarz inequality we obtain
Accordingly, we obtain
Then,
Therefore, Theorem 1 holds. □
Note 1: Both search directions and satisfy the sufficient descent condition, i.e., and with .
Note 2: , with .
Note 3: Algorithm 1 includes two different search direction and and three step-sizes , and with a mix of random and deterministic parameters, enabling Algorithm 1 capable of approaching an approximate solution to Problem (1). The difference between the size of the trust region of both search directions helps the proposed method search the null space (where solutions are) of the entire feasible region.
Note 4: Theorem 1 confirms that sequence obtained by Algorithm 1 makes System (1) approach 0 as long as , i.e.,
4. Numerical Experiments
In this section, nine test problems are considered to illustrate the convergence of the proposed Algorithm 1.
These test problems are taken from [46]. All codes were programmed in MATLAB version 8.5.0.197613 (R2015a) and ran on a PC with Intel(R) Core(TM) i5-3230M CPU@2.60GHz 2.60 GHz with RAM 4.00GB of memory on the Windows 10 operating system.
The parameters of Algorithm 1 are listed as follows.
, , , , and .
Numerical Results for a Non-Linear System
The section gives the numerical results of a system of non-linear equations solved by Algorithm 1 that represents two approaches (App1 and App2).
The performance of the two approaches is tested by a set of test problems (a system of non-linear equations).
The number of dimensions of the test problems (systems of non-linear equations) are listed as follows: .
The HRSG method described by Algorithm 1 contains two algorithms (HRSG1 and HRSG2). The HRSG1 algorithm denotes the first approach (App1) described in Section 2.1.1.
While the HRSG2 algorithm indicates the second approach (App1) described in Section 2.1.2.
Therefore, the results of both algorithms listed in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18 are obtained by running each algorithm 51 times for all the test problems with five cases of the number dimensions and seven initial points.
Table 1.
Numeral results (Problem 1) for the terms BItr, BEFs and BTcpu of the three methods.
Table 2.
Numeral results (Problem 2) for the terms BItr, BEFs and BTcpu for the three methods.
Table 3.
Numeral results (Problem 3) for the terms BItr, BEFs and BTcpu for the three methods.
Table 4.
Numeral results (Problem 4) for the terms BItr, BEFs and BTcpu for the three methods.
Table 5.
Numeral results (Problem 5) for the terms BItr, BEFs and BTcpu for the three methods.
Table 6.
Numeral results (Problem 6) for the terms BItr, BEFs and BTcpu for the three methods.
Table 7.
Numeral results (Problem 7) for the terms BItr, BEFs and BTcpu for the three methods.
Table 8.
Numeral results (Problem 8) for the terms BItr, BEFs and BTcpu for the three methods.
Table 9.
Numeral results (Problem 9) for the terms BItr, BEFs and BTcpu for the three methods.
Table 10.
Numeral results (Problem 1) for the terms WItr, WEFs and WTcpu for the three methods.
Table 11.
Numeral results (Problem 2) for the terms WItr, WEFs and WTcpu for the three methods.
Table 12.
Numeral results (Problem 3) for the terms WItr, WEFs and WTcpu for the three methods.
Table 13.
Numeral results (Problem 4) for the terms WItr, WEFs and WTcpu for the three methods.
Table 14.
Numeral results (Problem 5) for the terms WItr, WEFs and WTcpu for the three methods.
Table 15.
Numeral results (Problem 6) for the terms WItr, WEFs and WTcpu for the three methods.
Table 16.
Numeral results (Problem 7) for the terms WItr, WEFs and WTcpu for the three methods.
Table 17.
Numeral results (Problem 8) for the terms WItr, WEFs and WTcpu for the three methods.
Table 18.
Numeral results (Problem 9) for the terms WItr, WEFs and WTcpu for the three methods.
The stopping criterion and evaluation criteria of Algorithm 1 are listed as follows. The stopping criterion is or , then if the final result of a test problem is , the algorithm is incapable of solving this test problem, with setting .
Therefore, the evaluation criteria used to test the performance of our proposed method are the number of iterations (Itr), function evaluations (EFs), and time (Tcpu). Regarding these three criteria, the results obtained by 51 runs of Algorithm 1 (HRSG1 and HRSG2) are classified into three types of the results as follows.
- The best results of the number of iterations (BItr), function evaluation (BEFs), and run time (BTcpu).
- The worst results of the number of iterations (WItr), function evaluation (WEFs), and run time (WTcpu).
- The mean results of the number of iterations (MItr), function evaluation (MEFs), and run time (MTcpu).
All three types of results are listed in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18.
Thus, we compare the performance of Algorithm 1 (HRSG1, HRSG2) with the performance of the Algo algorithm proposed by the authors of [46].
Table 1 gives the best results (BItr, BEFs, and BTcpu) of the HRSG1 algorithm versus the results of the HRSG1 and Algo algorithms. Column 1 in Table 1 contains the number dimensions of Problem 1, .
Column 1 in Table 1 presents the initial points (INP) of Problem 1. Columns 3–5 give the result of the HRSG1 algorithm according to the BItr, BEFs, and BTcpu. Columns 6–8 give the result of the HRSG2 algorithm according to the BItr, BEFs, and BTcpu. The result of the Algo algorithm according to the BItr, BEFs, and BTcpu are listed in Columns 9–12. Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9 are similar to Table 1 for Problems 2–9.
We consider the results of the Algo algorithm from [46] and present the best results (BItr, BEFs, and BTcpu), these results are taken from Columns 3–6 in Tables A1–A9 from [46]. This is because no other results of the Algo algorithm are presented in [46].
Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18 present the results of the HRSG1 and HRSG2 algorithms regarding the WItr, WEFs, WTcpu, MItr, MEFs, and MTcpu, respectively.
The performance profile is used to show the success of the proposed algorithm to solve the test problems [47,48,49,50,51].
Thus, the numerical results are shown in the form of performance profiles to simply display the performance of the proposed algorithms [49].
The most important characteristic of the performance profiles is that the results listed in many tables can be shown in one figure.
This is implemented by plotting the different solvers in a cumulative distribution function .
For example, the information listed in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9 are summarized in Figure 1 to give an obvious image of the performance of the three algorithms, HRSG1, HRSG2 and Algo. The first graph (left of Figure 1) gives the performance of the BItr for the three algorithms. At , of test problems were solved by the Algo method against and of the test problems solved by the HRSG1 and HRSG2 methods, respectively. This means that the rank of the three algorithms is as follows, Algo algorithm, the HRSG2 algorithm, and HRSG1 algorithm, for the the BItr. When , of test problems were solved by the HRSG2 method against and of the test problems solved by the HRSG1 and Algo methods, respectively. This means that the rank of the three algorithms is as follows, the HRSG2 algorithm, the HRSG1 algorithm, and the Algo algorithm, for the the BItr.
Figure 1.
The BItr, BEFs, and BTcpu obtained by the HRSG1, HRSG2 and Algo methods.
If , of the test problems were solved by the HRSG1 method against and of the test problems solved by the HRSG2 and Algo methods, respectively. This means that the rank of the three algorithms is as follows, HRSG1, HRSG2, and the Algo algorithm for the BItr.
At , all the test problems were solved by the HRSG1 method against and of the test problems solved by the HRSG2 and Algo methods, respectively. This means that the rank of the three algorithms is as follows, HRSG1, HRSG2, and the Algo for the BItr.
Finally, at , all the test problems were solved by the HRSG1 and HRSG2 methods against of the test problems solved by the Algo method. This means that the Algo method failed to solve of the test problems. It is worth noting that this failure is indicated by (-) in Table 3 (Table A3 of [46]).
The performance profiles for the BFEs and BTcpu are illustrated by the second and third graphs of Figure 1. Therefore, it is clear that the performance of the HRSG1 algorithm is the best.
Furthermore, the results listed in Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18 are presented in Figure 2 and Figure 3 for the WItr, WFEs and WTcpu, and MEItr, MEFEs and METcpu, respectively.
Figure 2.
The WItr, WFEs and WTcpu obtained by the HRSG1 and HRSG2 methods.

Figure 3.
The MEItr, MEFEs and METcpu obtained by the HRSG1 and HRSG2 methods.
In general, the performance of the HRSG1 is better than the performance of the HRSG2.
Note: Both our proposed algorithms (HRSG1 and HRSG2) are capable of solving all the test problems at the lowest cost, where the WItr did not exceed 100, the WFEs did not exceed 1000 and the WTcpu did not exceed 10 s; while the Algo algorithm failed to solve of the test problems, illustrated by (-) in Table 3 (Table A3 of [46]).
Thus, in general, the proposed algorithm (HRSG) is competitive with, and in all cases superior to, the Algo algorithm in terms of efficiency, reliability and effectiveness to find the approximate solution of the test problems of the system of non-linear of equations.
5. Conclusions and Future Work
In this paper, we proposed a new adaptive hybrid two-step-size gradient method to solve large-scale systems of non-linear equations.
The proposed algorithm includes a mix of random and deterministic parameters by which two directions are designed. By simulating the forward difference method the (, ) step-sizes are generated iteratively. The two search directions possess different trust regions and satisfy the sufficient descent condition. The convergence analysis of the HRSG method was established.
The diversity and multiplicity of the parameters give the proposed algorithm several opportunities to reach the feasible region. Therefore, this was a good action that made the proposed algorithm capable of rapidly approaching the approximate solution of all the test problems at each run.
In general, the numerical results demonstrated that the HRSG method is superior compared to the results obtained by the Algo method.
Several evaluation criteria were used to test the performance of the proposed algorithm, providing a fair assessment of the HRSG method.
The proposed HRSG method is represented by two algorithms, HRSG1 and HRSG2.
It is noteworthy that the performance of the HRSG1 algorithm is more efficiency and effective than the HRSG2 algorithm. This is because of the errors resulting from calculating the step-size (interval).
The combination of the suggested method with the new line search approach results in a new globally convergent method. This new globally convergent method is capable of dealing with local minimization problems. Therefore, when any meta-heuristic technique as a global optimization algorithm is combined with the HRSG as a globally convergent method, the result is a new hybrid semi-gradient meta-heuristic algorithm. The idea behind this hybridization is to utilize the advantages of both techniques. Therefore, the implementation of this idea will be our interest in future work.
Furthermore, the result of this hybridization is a new hybrid semi-gradient meta-heuristic algorithm, capable of dealing with unconstrained, constrained, and multi-objective optimization problems.
It is also attractive to modify and adapt the HRSG method to make it capable of dealing with image restoration problems.
Moreover, it is possible to hybridise the HRSG method with new versions of the CG method. As a result, we could obtain a new hybrid method, capable of training deep neural networks.
The idea of the HRSG method depends on imitating the forward difference method by using one point to estimate the values of the gradient vector at each iteration, where the function evaluation is, at most, one per iteration. Therefore, the HRSG method contains two approaches used to implement this task, namely HRSG1 and HRSG2. Therefore, in future work, both approaches can be used by using the forward and central differences with more than one point.
Furthermore, the approximate values of the gradient vectors can be calculated using the forward and central differences randomly in this case, where the number of function evaluations is zero.
Hence, the proposed method obtained from the above proposals can be applied to deal with systems of non-linear equations, unconstrained, constrained, and multi-objective optimization problems, image restoration, and training deep neural networks.
List of Test Problems
Problem 1: Exponential Function [27,46].
Problem 2: Exponential Function.
Problem 3: Non-Smooth Function [27,46].
Problem 4: [27,46].
Problem 5: Strictly Convex Function I [27,46].
Problem 6: Strictly Convex Function II [46].
Problem 7: Tri-Diagonal Exponential Function [27,46].
where .
Problem 8: Non-Smooth Function [27,46].
Problem 9: The Trig Exponential Function [27].
Initial Points (INP): , , , , , and .
Author Contributions
Conceptualization, E.A.; methodology, E.A. and S.M.; software, E.A. and S.M.; validation, E.A.; formal analysis, E.A.; investigation, E.A.; Funding acquisition, E.A.; data curation, E.A.; writing original draft preparation, E.A. and S.M.; writing-review and editing E.A. and S.M.; visualization, E.A. and S.M.; supervision, E.A. and S.M. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Deanship of Scientific Research at Najran University (grant number NU/DRP/SERC/12/3).
Data Availability Statement
Not applicable.
Acknowledgments
The authors are thankful to the Deanship of Scientific Research at Najran University for funding this work, under the General Research Funding program grant code (NU/DRP/SERC/12/3).
Conflicts of Interest
The authors declare no conflict of interest.
References
- Alnowibet, K.A.; Alshamrani, A.M.; Alrasheedi, A.F.; Mahdi, S.; El-Alem, M.; Aboutahoun, A.; Mohamed, A.W. Efficient Modified Meta-Heuristic Technique for Unconstrained Optimization Problems. Axioms 2022, 11, 483. [Google Scholar] [CrossRef]
- Abubakar, A.B.; Kumam, P.; Mohammad, H.; Ibrahim, A.H.; Kiri, A.I. A hybrid approach for finding approximate solutions to constrained nonlinear monotone operator equations with applications. Appl. Numer. Math. 2022, 177, 79–92. [Google Scholar] [CrossRef]
- Meintjes, K.; Morgan, A.P. Chemical equilibrium systems as numerical test problems. ACM Trans. Math. Softw. (TOMS) 1990, 16, 143–151. [Google Scholar] [CrossRef]
- Xiao, Y.; Zhu, H. A conjugate gradient method to solve convex constrained monotone equations with applications in compressive sensing. J. Math. Anal. Appl. 2013, 405, 310–319. [Google Scholar] [CrossRef]
- Abubakar, A.B.; Mohammad, H.; Kumam, P.; Rano, S.A.; Ibrahim, A.H.; Kiri, A.I. On the new spectral conjugate gradient-type method for monotone nonlinear equations and signal recovery. Math. Methods Appl. Sci. 2022. [Google Scholar] [CrossRef]
- Ibrahim, A.H.; Kumam, P.; Abubakar, A.B.; Abdullahi, M.S.; Mohammad, H. A Dai-Liao-type projection method for monotone nonlinear equations and signal processing. Demonstr. Math. 2022, 55, 978–1013. [Google Scholar] [CrossRef]
- Brown, P.N.; Saad, Y. Convergence theory of nonlinear Newton–Krylov algorithms. SIAM J. Optim. 1994, 4, 297–330. [Google Scholar] [CrossRef]
- Li, D.H.; Fukushima, M. A modified BFGS method and its global convergence in nonconvex minimization. J. Comput. Appl. Math. 2001, 129, 15–35. [Google Scholar] [CrossRef]
- Fletcher, R.; Reeves, C.M. Function minimization by conjugate gradients. Comput. J. 1964, 7, 149–154. [Google Scholar] [CrossRef]
- Hestenes, M.R.; Stiefel, E. Methods of Conjugate Gradients for Solving. J. Res. Natl. Bur. Stand. 1952, 49, 409. [Google Scholar] [CrossRef]
- Liu, Y.; Storey, C. Efficient generalized conjugate gradient algorithms, part 1: Theory. J. Optim. Theory Appl. 1991, 69, 129–137. [Google Scholar] [CrossRef]
- Polak, E.; Ribiere, G. Note sur la convergence de méthodes de directions conjuguées. ESAIM Math. Model. Numer. Anal. Model. Math. Anal. Numer. 1969, 3, 35–43. [Google Scholar] [CrossRef]
- Birgin, E.G.; Krejić, N.; Martínez, J.M. Globally convergent inexact quasi-Newton methods for solving nonlinear systems. Numer. Algorithms 2003, 32, 249–260. [Google Scholar] [CrossRef]
- Li, D.; Fukushima, M. A globally and superlinearly convergent gauss–Newton-based BFGS method for symmetric nonlinear equations. SIAM J. Numer. Anal. 1999, 37, 152–172. [Google Scholar] [CrossRef]
- Li, D.H.; Fukushima, M. A derivative-free line search and global convergence of Broyden-like method for nonlinear equations. Optim. Methods Softw. 2000, 13, 181–201. [Google Scholar] [CrossRef]
- Solodov, M.V.; Svaiter, B.F. A globally convergent inexact Newton method for systems of monotone equations. In Reformulation: Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods; Springer: Cham, Switzerland, 1998; pp. 355–369. [Google Scholar]
- Zhou, G.; Toh, K.C. Superlinear convergence of a Newton-type algorithm for monotone equations. J. Optim. Theory Appl. 2005, 125, 205–221. [Google Scholar] [CrossRef]
- Abubakar, A.B.; Kumam, P. A descent Dai-Liao conjugate gradient method for nonlinear equations. Numer. Algorithms 2019, 81, 197–210. [Google Scholar] [CrossRef]
- Dai, Z.; Zhu, H. A modified Hestenes-Stiefel-type derivative-free method for large-scale nonlinear monotone equations. Mathematics 2020, 8, 168. [Google Scholar] [CrossRef]
- Su, Z.; Li, M. A Derivative-Free Liu–Storey Method for Solving Large-Scale Nonlinear Systems of Equations. Math. Probl. Eng. 2020, 2020, 6854501. [Google Scholar] [CrossRef]
- Yuan, G.; Wei, Z.; Lu, S. Limited memory BFGS method with backtracking for symmetric nonlinear equations. Math. Comput. Model. 2011, 54, 367–377. [Google Scholar] [CrossRef]
- Yuan, G.; Li, T.; Hu, W. A conjugate gradient algorithm for large-scale nonlinear equations and image restoration problems. Appl. Numer. Math. 2020, 147, 129–141. [Google Scholar] [CrossRef]
- Zhang, L.; Zhou, W. Spectral gradient projection method for solving nonlinear monotone equations. J. Comput. Appl. Math. 2006, 196, 478–484. [Google Scholar] [CrossRef]
- Ibrahim, A.H.; Deepho, J.; Abubakar, A.B.; Kamandi, A. A globally convergent derivative-free projection algorithm for signal processing. J. Interdiscip. Math. 2022, 25, 2301–2320. [Google Scholar] [CrossRef]
- Cruz, W.L. A spectral algorithm for large-scale systems of nonlinear monotone equations. Numer. Algorithms 2017, 76, 1109–1130. [Google Scholar] [CrossRef]
- Dennis, J.E.; Moré, J.J. A characterization of superlinear convergence and its application to quasi-Newton methods. Math. Comput. 1974, 28, 549–560. [Google Scholar] [CrossRef]
- Stanimirovic, P.S.; Shaini, B.I.; Sabiu, J.; Shah, A.; Petrovic, M.J.; Ivanov, B.; Cao, X.; Stupina, A.; Li, S. Improved Gradient Descent Iterations for Solving Systems of Nonlinear Equations. Algorithms 2023, 16, 64. [Google Scholar] [CrossRef]
- Waziri, M.Y.; Muhammad, H.U.; Halilu, A.S.; Ahmed, K. Modified matrix-free methods for solving system of nonlinear equations. Optimization 2021, 70, 2321–2340. [Google Scholar] [CrossRef]
- Kumam, P.; Abubakar, A.B.; Ibrahim, A.H.; Kura, H.U.; Panyanak, B.; Pakkaranang, N. Another hybrid approach for solving monotone operator equations and application to signal processing. Math. Methods Appl. Sci. 2022, 45, 7897–7922. [Google Scholar] [CrossRef]
- Dai, Z.; Wen, F. Global convergence of a modified Hestenes-Stiefel nonlinear conjugate gradient method with Armijo line search. Numer. Algorithms 2012, 59, 79–93. [Google Scholar] [CrossRef]
- Ibrahim, A.H.; Kumam, P.; Abubakar, A.B.; Abubakar, J. A derivative-free projection method for nonlinear equations with non-Lipschitz operator: Application to LASSO problem. Math. Methods Appl. Sci. 2023. [Google Scholar] [CrossRef]
- Alnowibet, K.A.; Mahdi, S.; El-Alem, M.; Abdelawwad, M.; Mohamed, A.W. Guided Hybrid Modified Simulated Annealing Algorithm for Solving Constrained Global Optimization Problems. Mathematics 2022, 10, 1312. [Google Scholar] [CrossRef]
- EL-Alem, M.; Aboutahoun, A.; Mahdi, S. Hybrid gradient simulated annealing algorithm for finding the global optimal of a nonlinear unconstrained optimization problem. Soft Comput. 2020, 25, 2325–2350. [Google Scholar] [CrossRef]
- Kramer, O.; Ciaurri, D.E.; Koziel, S. Derivative-free optimization. In Computational Optimization, Methods and Algorithms; Springer: Cham, Switzerland, 2011; pp. 61–83. [Google Scholar]
- Larson, J.; Menickelly, M.; Wild, S.M. Derivative-free optimization methods. Acta Numer. 2019, 28, 287–404. [Google Scholar] [CrossRef]
- Shi, H.J.M.; Xie, Y.; Xuan, M.Q.; Nocedal, J. Adaptive Finite-Difference Interval Estimation for Noisy Derivative-Free Optimization. arXiv 2021, arXiv:2110.06380. [Google Scholar] [CrossRef]
- Shi, H.J.M.; Xuan, M.Q.; Oztoprak, F.; Nocedal, J. On the numerical performance of derivative-free optimization methods based on finite-difference approximations. arXiv 2021, arXiv:2102.09762. [Google Scholar]
- Berahas, A.S.; Cao, L.; Choromanski, K.; Scheinberg, K. A theoretical and empirical comparison of gradient approximations in derivative-free optimization. Found. Comput. Math. 2022, 22, 507–560. [Google Scholar] [CrossRef]
- Curtis, A.; Reid, J. The choice of step lengths when using differences to approximate Jacobian matrices. IMA J. Appl. Math. 1974, 13, 121–126. [Google Scholar] [CrossRef]
- Frontini, M.; Milovanović, G.V. Numerical differentiation of analytic functions using quadratures on the semicircle. Comput. Math. Appl. 1991, 22, 99–106. [Google Scholar]
- Gill, P.E.; Murray, W.; Saunders, M.A.; Wright, M.H. Computing forward-difference intervals for numerical optimization. SIAM J. Sci. Stat. Comput. 1983, 4, 310–321. [Google Scholar] [CrossRef]
- Xie, Y. Methods for Nonlinear and Noisy Optimization. Ph.D. Thesis, Northwestern University, Evanston, IL, USA, 2021. [Google Scholar]
- Alnowibet, K.A.; Mahdi, S.; Alshamrani, A.M.; Sallam, K.M.; Mohamed, A.W. A Family of Hybrid Stochastic Conjugate Gradient Algorithms for Local and Global Minimization Problems. Mathematics 2022, 10, 3595. [Google Scholar] [CrossRef]
- Alshamrani, A.M.; Alrasheedi, A.F.; Alnowibet, K.A.; Mahdi, S.; Mohamed, A.W. A Hybrid Stochastic Deterministic Algorithm for Solving Unconstrained Optimization Problems. Mathematics 2022, 10, 3032. [Google Scholar] [CrossRef]
- Conn, A.R.; Scheinberg, K.; Vicente, L.N. Introduction to Derivative-Free Optimization; SIAM: Philadelphia, PA, USA, 2009. [Google Scholar]
- Hassan Ibrahim, A.; Kumam, P.; Hassan, B.A.; Bala Abubakar, A.; Abubakar, J. A derivative-free three-term Hestenes–Stiefel type method for constrained nonlinear equations and image restoration. Int. J. Comput. Math. 2022, 99, 1041–1065. [Google Scholar] [CrossRef]
- Ali, M.M.; Khompatraporn, C.; Zabinsky, Z.B. A numerical evaluation of several stochastic algorithms on selected continuous global optimization test problems. J. Glob. Optim. 2005, 31, 635–672. [Google Scholar] [CrossRef]
- Barbosa, H.J.; Bernardino, H.S.; Barreto, A.M. Using performance profiles to analyze the results of the 2006 CEC constrained optimization competition. In Proceedings of the IEEE Congress on Evolutionary Computation, Barcelona, Spain, 18–23 July 2010; pp. 1–8. [Google Scholar]
- Dolan, E.D.; Moré, J.J. Benchmarking optimization software with performance profiles. Math. Program. 2002, 91, 201–213. [Google Scholar] [CrossRef]
- Moré, J.J.; Wild, S.M. Benchmarking derivative-free optimization algorithms. SIAM J. Optim. 2009, 20, 172–191. [Google Scholar] [CrossRef]
- Vaz, A.I.F.; Vicente, L.N. A particle swarm pattern search method for bound constrained global optimization. J. Glob. Optim. 2007, 39, 197–219. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).