Abstract
In this article, we propose a modified self-adaptive conjugate gradient algorithm for handling nonlinear monotone equations with the constraints being convex. Under some nice conditions, the global convergence of the method was established. Numerical examples reported show that the method is promising and efficient for solving monotone nonlinear equations. In addition, we applied the proposed algorithm to solve sparse signal reconstruction problems.
Keywords:
non-linear equations; conjugate gradient method; projection method; convex constraints; signal reconstruction problem MSC:
65K05; 90C52; 90C56; 52A20
1. Introduction
Suppose is a nonempty, closed and convex subset of , F a continuous function from to . A constrained nonlinear monotone equation involves finding a point , such that
Many algorithms have been proposed in the literature to solve nonlinear constrained equations, some of which are the trust region and the Levenberg-Marquardt method [1]. However, the need for these methods to compute and store matrix in every iteration, make them unsuitable for solving large-scale nonlinear equations.
Conjugate gradient (CG) methods is an iterative method developed for handling the unconstrained optimization problem [2,3,4,5,6,7,8,9]. CG methods do not require matrix storage, which make them one of the most efficient methods for handling large-scale unconstrained optimization problems. Moreover, generating a descent direction does not always hold based on the secant conditions. In order to obtain a descent direction, Narushima et al. [6] and Zhang et al. [8] proposed three term CG methods, which always generate a descent direction and established the convergence of the methods under some suitable conditions. Also in Reference [5], Narushima proposed a smoothing CG algorithm, which combine the smoothing approach with the Polak–Ribière–Polyak CG methods in Reference [9], to handle unconstrained non-smooth equations. The convergence of the method was established under some mild conditions.
Methods for solving unconstrained problems sometimes become less useful (see for example References [10,11,12,13]), as in many practical applications, such as equilibrium problems, the solution of the unconstrained problem may lie outside the constrained set . This reason made researchers shift their attention to the constrained case (1). In the last few years, many kinds of algorithms for solving nonlinear monotone equations with convex constrained set have been developed and one of the popular is the projection method. For example, in Reference [14] Wang et al. proposed a projection method for solving systems of monotone nonlinear equations with convex constraints. The method was based on the inexact Newton backtracking technique and the direction was obtained by minimization of a linear system together with the constrained condition at each iteration. Also, in Reference [15] Wang et al. presented a modification of the method in Reference [16], and the global convergence as well as the super-linear rate of convergence were established under same conditions in Reference [14]. However, the direction of the methods in Reference [15,16] were determined by minimization of linear equations at each step. In trying to avoid solving the linear equation to obtain the direction at each step, Xiao and Zhu [17] proposed a projected CG methods, which combines the CG-DESCENT method in [4] and the projection technique by Solodov and Svaiter [18]. In Reference [19], a modification of the method in Reference [17] was proposed by Liu and Li. The advantage of this modification was that it improves the numerical performance of the method in Reference [17] and still retains its nice properties. Furthermore, Wang et al. [20] proposed a self-adaptive three-term CG methods for solving constrained nonlinear monotone equations. The method can be viewed as combination of the CG methods, the projection method and the self-adaptive method. Motivated by the above methods, we propose a modification of the method in Reference [20] for solving nonlinear monotone equations with convex constraints. The modification improves the numerical performance of the method in Reference [20] and still inherits its nice properties. The difference between the two methods is that in Reference [20] is replaced by (More details can be found in the next section). Under appropriate conditions, the global convergence of the proposed method is established. Numerical results presented show that the proposed method is efficient and promising compared to some similar existing algorithms.
The remaining part of this paper is organized as follows. In Section 2, we state some preliminaries and then present the algorithm. The global convergence of the proposed method is proved in Section 3. In Section 4, we report some numerical experiments to show its performance in solving nonlinear monotone equations with convex constraints, and lastly apply it to solve some signal recovery problems.
2. Preliminaries and Algorithm
This section gives some basic concepts and properties of the projection mapping as well as some assumptions. denotes the Euclidean norm throughout the paper.
Definition 1.
Let be a nonempty closed convex set. Then for any , its orthogonal projection onto Ω, denoted by , is defined by
The following Lemma provides us with some well-known properties of the projection mapping.
Lemma 1.
[20] Let be a nonempty, closed and convex set. Then the following statements are true:
- 1.
- ,
- 2.
- ,
- 3.
- ,
All through this article, we assume the following
- ()
- The solution set of (1), denoted by , is nonempty.
- ()
- The mapping F is monotone, that is,
- ()
- The mapping is Lipschitz continuous, that is there exists a positive constant L such that , .
| Algorithm 1: Modified Self-adaptive CG method (MSCG). | |
| Step 0. Given an arbitrary initial point , parameters , , , , , , and set . | |
| Step 1. If , stop, otherwise go to Step 2. | |
| Step 2. Compute
| |
| Step 3. Compute the step length and is the smallest non-negative integer m such that
| |
| Step 4. Set and compute
| |
| Step 5. Let and go to Step 1. | |
It can be observed that the modification made is by replacing , in [20] with , respectively in the proposed algorithm.
Remark 1.
Using Cauchy-Schwartz inequality, we get
Remark 2.
From the definition of , and (8), we have
The above inequality shows that the denominator of and cannot be zero except at the solution. However, there is no any guarantee that the denominator of and defined in Reference [20] cannot be zero. In addition, the conditions imposed in step 2 of Algorithm 2.1 of [20] were not considered in our case.
3. Convergence Analysis
To prove the global convergence of Algorithm 1, the following Lemmas are needed. The following Lemma shows that Algorithm 1 is well-defined.
Lemma 2.
Suppose that assumptions ()–() hold, then there exists a step-length satisfying the line search (6) .
Proof.
Lemma 3.
Suppose that () hold and the sequences and be generated by Algorithm 1. Then we have
Proof.
Lemma 4.
Suppose that assumptions ()–() hold, then the sequences and generated by Algorithm 1 are bounded. Moreover, we have
and
Proof.
We will start by showing that the sequences and are bounded. Suppose , then by monotonicity of F, we get
Also by definition of and the line search (6), we have
So, we have
Thus the sequence is non increasing and convergent and hence is bounded. Furthermore, from Equation (16), we have
and we can deduce recursively that
Then from Assumption (), we obtain
If we let , then the sequence is bounded, that is,
By the definition of , Equation (15), monotonicity of F and the Cauchy-Schwatz inequality, we get
The boundedness of the sequence together with Equations (18) and (19), implies that the sequence is bounded.
Since is bounded, then for any , the sequence is also bounded, that is, there exists a positive constant such that
This together with Assumption () yields
Therefore, using Equation (16), we have
which implies
Equation (20) implies
However, using statement 2 of Lemma 1, the definition of and the Cauchy-Schwatz inequality, we have
which yields
□
Remark 3.
By Equation (12) and definition of , we have
Theorem 1.
Suppose that assumptions ()–() hold and let the sequence be generated by Algorithm 1, then
Proof.
Assume that Equation (23) is not true, then there exists a constant such that
We will first show that the sequence is bounded. From the definition of , we have
Also from definition of and assumption (), we have
Furthermore by definition of , (25) and (26), we obtain
Therefore, by (2), (9), (18), (27) and Cauchy-Schwatz inequality, we have
Letting , then
4. Numerical Examples
This section reports some numerical results to show the efficiency of Algorithm 1. For convenience sake, we denote Algorithm 1 by the modified self-adaptive method MSCG. We also divide this section into two. First we compare the MSCG method with the projected conjugate gradient PCG and the self-adaptive three-term conjugate gradient SATCGM methods in References [19,20] respectively, by solving some monotone nonlinear equations with convex constraints using different initial points and several dimensions. Secondly, the MSCG method is applied to solve signal recovery problems. All codes were written in MATLAB R2017a and run on a PC with intel COREi5 processor with 4 GB of RAM and CPU 2.3 GHZ.
4.1. Numerical Examples on Some Convex Constrained Nonlinear Monotone Equations
Same line search implementation was used for both MSCG, PCG and SATCGM. The specific parameters used for each method are as follows:
- MSCG method:
- , , , , .
- PCG method:
- All parameters are chosen as in [19].
- SATCGM method:
- All parameters are chosen as in [20]. All runs were stopped whenever
We test problems 1 to 9 with dimensions of n = 1000, 5000, 10,000, 50,000, 100,000 and different initial points: , , , , , , , . The numerical results in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9 report the number of iterations (Iter), number of function evaluations (Fval), CPU time in seconds (Time) and the norm at the approximate solution (Norm). The symbol ‘−’ is used to indicate that the number of iterations exceeds 1000 and/or the number of function evaluations exceeds 2000.
Table 1.
Numerical Results for modified self-adaptive method MSCG, projected conjugate gradient PCG and the self-adaptive three-term conjugate gradient SATCGM for Problem 1 with given initial points and dimensions.
Table 2.
Numerical Results for MSCG, PCG and SATCGM for Problem 2 with given initial points and dimensions.
Table 3.
Numerical Results for MSCG, PCG and SATCGM for Problem 3 with given initial points and dimensions.
Table 4.
Numerical Results for MSCG, PCG and SATCGM for Problem 4 with given initial points and dimensions.
Table 5.
Numerical Results for MSCG, PCG and SATCGM for Problem 5 with given initial points and dimensions.
Table 6.
Numerical Results for MSCG, PCG and SATCGM for Problem 6 with given initial points and dimensions.
Table 7.
Numerical Results for MSCG, PCG and SATCGM for Problem 7 with given initial points and dimensions.
Table 8.
Numerical Results for MSCG, PCG and SATCGM for Problem 8 with given initial points and dimensions.
Table 9.
Numerical Results for MSCG, PCG and SATCGM for Problem 9 with given initial points and dimensions.
The problem functions and feasible sets tested are listed as follows:
Problem 1 Modified exponential function
Problem 2 Logarithmic Function
Problem 3 [21]
Problem 4 [22]
Problem 5 Strictly convex function [14]
Problem 6 Linear monotone problem
Problem 7 Tridiagonal Exponential Problem [23]
Problem 8
Problem 9
The numerical results indicate that the MSCG method is more effective than the PCG and SATCGM methods for the given problems as it solves and win 6 out of 9 of the problems tested both in terms of number of iterations, number of function evaluations and CPU time (see Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8 and Table 9). In particular, the PCG method fails to solve problems 4 completely while MSCG was able to solve all the problems except for the initial points and (see Table 4). In addition, the SATCGM method fails to solve problem 4 except in the case of dimension 1000 and initial point for the remaining dimensions considered. Therefore, we can conclude that the MSCG method is a very efficient tool for solving nonlinear monotone equations with convex constraints, especially for large-scale dimensions.
4.2. Experiments on Solving Some Signal Recovery Problems in Compressive Sensing
There are many problems in signal processing and statistical inference involving finding sparse solutions to ill-conditioned linear systems of equations. Among popular approaches is minimizing an objective function which contains quadratic () error term and a sparse regularization term, that is,
where , is an observation, () is a linear operator, is a nonnegative parameter, denotes the Euclidean norm of x and is the norm of x. It is easy to see that problem (30) is a convex unconstrained minimization problem. Due to the fact that if the original signal is sparse or approximately sparse in some orthogonal basis, problem (30) frequently appears in compressive sensing, and hence an exact restoration can be produced by solving (30).
Iterative methods for solving (30) have been presented in the literature, (see References [24,25,26,27,28,29]). The most popular method among these methods is the gradient based method and the earliest gradient projection method for sparse reconstruction (GPRS) was proposed by Figueiredo et al. [27]. The first step of the GPRS method is to express (30) as a quadratic problem using the following process.
Let and splitting it into its positive and negative parts. Then x can be formulated as
where , for all , and . By definition of -norm, we have , where . Now (30) can be written as
which is a bound-constrained quadratic program. However, from Reference [27], Equation (31) can be written in standard form as
where , , , .
Clearly, D is a positive semi-definite matrix, which implies that Equation (32) is a convex quadratic problem.
Xiao et al. [17] translated (32) into a linear variable inequality problem which is equivalent to a linear complementarity problem. Furthermore, they pointed out that z is a solution of the linear complementarity problem if and only if it is a solution of the nonlinear equation:
It was proved in Reference [30,31] that is continuous and monotone. Therefore problem (30) can be translated into problem (1) and thus MSCG method can be applied to solve (30).
In this experiment, we consider a simple compressive sensing possible situation, where our goal is to reconstruct a sparse signal of length n from k observations. The quality of restoration is assessed by mean of squared error (MSE) to the original signal ,
where is the recovered or restored signal. The signal size is chosen as , and the original signal contains randomly nonzero elements. A is the Gaussian matrix generated by the command in MATLAB. In addition, the measurement y is distributed with noise, that is, , where is the Gaussian noise distributed normally with mean 0 and variance ().
To show the performance of the MSCG method in compressive sensing, we compare it with the PCG method. The parameters in both MSCG and PCG methods are chosen as , , , and and the merit function used is . To achieve fairness in comparison, each code was run from same initial point, same continuation technique on the parameter , and observed only the behaviour of the convergence of each method to have a similar accurate solution. The experiment is initialized by and terminates when
where is the function evaluation at . In Figure 1, MSCG and PCG methods recovered the disturbed signal almost exactly. In order to show the performance of both methods visually, four figures were plotted to demonstrate their convergence behaviour based on MSE, objective function values, number of iterations and CPU time (see Figure 2, Figure 3, Figure 4 and Figure 5). Furthermore, the experiment was repeated for 25 different noise samples (see Table 10). From the Table, it can be observed that the MSCG is more efficient in terms of iterations and CPU time than the PCG method in most cases.
Figure 1.
From top to bottom: the original image, the measurement, and the recovered signals b they PCG and MSCG methods.
Figure 2.
Iterations.
Figure 3.
CPU time (seconds).
Figure 4.
Iterations.
Figure 5.
CPU time (seconds).
Table 10.
Twenty five experiment results of norm regularization problem for MSCG and PCG methods.
5. Conclusions
In this paper, a modified three-term conjugate gradient method for solving monotone nonlinear equations with convex constraints was presented. The proposed algorithm is suitable for solving non-smooth equations because it requires no Jacobian information of the nonlinear equations. Under some assumptions, global convergence properties of the proposed method were proved. The numerical experiments presented clearly show how effective the MSCG algorithm is compared to the PCG and SATCGM methods of References [19,20] for the given constrained problems. In addition, the MSCG algorithm was shown to be effective in signal recovery problems.
Author Contributions
Conceptualization, A.B.A.; methodology, A.B.A.; software, A.B.A.; validation, P.K., A.M.A. and P.T.; formal analysis, P.K. and P.T.; investigation, P.K.; resources, P.K. and P.T.; data curation, A.M.A.; writing–original draft preparation, A.B.A.; writing–review and editing, A.M.A.; visualization, P.T.; supervision, P.K.; project administration, P.K. and P.T.; funding acquisition, P.K. and P.T.
Funding
Petchra Pra Jom Klao Doctoral Scholarship for Ph.D. program of King Mongkut’s University of Technology Thonburi (KMUTT) and Theoretical and Computational Science (TaCS) Center. Moreover, this project was partially supported by the Thailand Research Fund (TRF) and the King Mongkut’s University of Technology Thonburi (KMUTT) under the TRF Research Scholar Award (Grant No. RSA6080047). Also, this research work was funded by King Mongkut’s University of Technology North Bangkok, contract no. KMUTNB-61-GOV-D-68.
Acknowledgments
This project was supported by Center of Excellence in Theoretical and Computational Science (TaCS-CoE) Center under Computational and Applied Science for Smart Innovation research Cluster (CLASSIC), Faculty of Science, KMUTT. The first author thanks for the support of the Petchra Pra Jom Klao Doctoral Scholarship Academic for Ph.D. Program at King Mongkut’s University of Technology Thonburi (KMUTT). Moreover, this research work was financially supported by King Mongkut’s University of Technology Thonburi through the KMUTT 55th Anniversary Commemorative Fund.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Kanzow, C.; Yamashita, N.; Fukushima, M. Levenberg–Marquardt methods with strong local convergence properties for solving nonlinear equations with convex constraints. J. Comput. Appl. Math. 2004, 172, 375–397. [Google Scholar] [CrossRef]
- Al-Baali, M.; Narushima, Y.; Yabe, H. A family of three term conjugate gradient methods with sufficient descent property for unconstrained optimization. Comput. Optim. Appl. 2015, 60, 89–110. [Google Scholar] [CrossRef]
- Hager, W.W.; Zhang, H. A survey of nonlinear conjugate gradient methods. Pac. J. Optim. 2006, 2, 35–58. [Google Scholar]
- Hager, W.; Zhang, H. A New Conjugate Gradient Method with Guaranteed Descent and an Efficient Line Search. SIAM J. Optim. 2005, 16, 170–192. [Google Scholar] [CrossRef]
- Narushima, Y. A smoothing conjugate gradient method for solving systems of nonsmooth equations. Appl. Math. Comput. 2013, 219, 8646–8655. [Google Scholar] [CrossRef]
- Narushima, Y.; Yabe, H.; Ford, J. A Three-Term Conjugate Gradient Method with Sufficient Descent Property for Unconstrained Optimization. SIAM J. Optim. 2011, 21, 212–230. [Google Scholar] [CrossRef]
- Sugiki, K.; Narushima, Y.; Yabe, H. Globally Convergent Three-Term Conjugate Gradient Methods that Use Secant Conditions and Generate Descent Search Directions for Unconstrained Optimization. J. Optim. Theory Appl. 2012, 153, 733–757. [Google Scholar] [CrossRef]
- Zhang, L.; Zhou, W.; Li, D. Some descent three-term conjugate gradient methods and their global convergence. Optim. Methods Softw. 2007, 22, 697–711. [Google Scholar] [CrossRef]
- Zhang, L.; Zhou, W.; Li, D.H. A descent modified Polak–Ribière–Polyak conjugate gradient method and its global convergence. IMA J. Numer. Anal. 2006, 26, 629–640. [Google Scholar] [CrossRef]
- Abubakar, A.B.; Kumam, P. An improved three-term derivative-free method for solving nonlinear equations. Comput. Appl. Math. 2018, 37, 6760–6773. [Google Scholar] [CrossRef]
- Abubakar, A.B.; Kumam, P. A descent Dai-Liao conjugate gradient method for nonlinear equations. Numer. Algorithms 2019, 81, 197–210. [Google Scholar] [CrossRef]
- Muhammed, A.A.; Kumam, P.; Abubakar, A.B.; Wakili, A.; Pakkaranang, N. A New Hybrid Spectral Gradient Projection Method for Monotone System of Nonlinear Equations with Convex Constraints. Thai J. Math. 2018, 16, 125–147. [Google Scholar]
- Mohammad, H.; Abubakar, A.B. A positive spectral gradient-like method for nonlinear monotone equations. Bull. Comput. Appl. Math. 2017, 5, 99–115. [Google Scholar]
- Wang, C.; Wang, Y.; Xu, C. A projection method for a system of nonlinear monotone equations with convex constraints. Math. Methods Oper. Res. 2007, 66, 33–46. [Google Scholar] [CrossRef]
- Wang, C.; Wang, Y. A superlinearly convergent projection method for constrained systems of nonlinear equations. J. Glob. Optim. 2009, 44, 283–296. [Google Scholar] [CrossRef]
- Solodov, M.; Svaiter, B. A New Projection Method for Variational Inequality Problems. SIAM J. Control Optim. 1999, 37, 765–776. [Google Scholar] [CrossRef]
- Xiao, Y.; Zhu, H. A conjugate gradient method to solve convex constrained monotone equations with applications in compressive sensing. J. Math. Anal. Appl. 2013, 405, 310–319. [Google Scholar] [CrossRef]
- Solodov, M.V.; Svaiter, B.F. A Globally Convergent Inexact Newton Method for Systems of Monotone Equations. In Reformulation: Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods; Fukushima, M., Qi, L., Eds.; Applied Optimization; Springer: Boston, MA, USA; Berlin/Heidelberg, Germany, 1998; Volume 22, pp. 355–369. [Google Scholar]
- Liu, J.; Li, S. A projection method for convex constrained monotone nonlinear equations with applications. Comput. Math. Appl. 2015, 70, 2442–2453. [Google Scholar] [CrossRef]
- Wang, X.Y.; Li, S.J.; Kou, X.P. A self-adaptive three-term conjugate gradient method for monotone nonlinear equations with convex constraints. Calcolo 2016, 53, 133–145. [Google Scholar] [CrossRef]
- Zhou, W.; Li, D. Limited memory BFGS method for nonlinear monotone equations. J. Comput. Math. 2007, 25, 89–96. [Google Scholar]
- La Cruz, W. A spectral algorithm for large-scale systems of nonlinear monotone equations. Numer. Algorithms 2017, 76, 1109–1130. [Google Scholar] [CrossRef]
- Bing, Y.; Lin, G. An efficient implementation of Merrill’s method for sparse or partially separable systems of nonlinear equations. SIAM J. Optim. 1991, 1, 206–221. [Google Scholar] [CrossRef]
- Figueiredo, M.A.; Nowak, R.D. An EM algorithm for wavelet-based image restoration. IEEE Trans. Image Process. 2003, 12, 906–916. [Google Scholar] [CrossRef]
- Hale, E.T.; Yin, W.; Zhang, Y. A fixed-point continuation method for ℓ1-regularized minimization with applications to compressed sensing. CAAM TR07-07 Rice Univ. 2007, 43, 44. [Google Scholar]
- Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
- Figueiredo, M.A.; Nowak, R.D.; Wright, S.J. Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE J. Sel. Top. Signal Process. 2007, 1, 586–597. [Google Scholar] [CrossRef]
- Van Den Berg, E.; Friedlander, M.P. Probing the Pareto frontier for basis pursuit solutions. SIAM J. Sci. Comput. 2008, 31, 890–912. [Google Scholar] [CrossRef]
- Birgin, E.G.; Martínez, J.M.; Raydan, M. Nonmonotone spectral projected gradient methods on convex sets. SIAM J. Optim. 2000, 10, 1196–1211. [Google Scholar] [CrossRef]
- Xiao, Y.; Wang, Q.; Hu, Q. Non-smooth equations based method for ℓ1-norm problems with applications to compressed sensing. Nonlinear Anal. Theory Methods Appl. 2011, 74, 3570–3577. [Google Scholar] [CrossRef]
- Pang, J.S. Inexact Newton methods for the nonlinear complementarity problem. Math. Program. 1986, 36, 54–71. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).