Abstract
In this work, we integrate some new approximate functions using the logarithmic penalty method to solve nonlinear optimization problems. Firstly, we determine the direction by Newton’s method. Then, we establish an efficient algorithm to compute the displacement step according to the direction. Finally, we illustrate the superior performance of our new approximate function with respect to the line search one through a numerical experiment on numerous collections of test problems.
Keywords:
interior point methods; logarithmic penalty method; line search; approximate functions; nonlinear optimization MSC:
90C25; 90C30; 90C51
1. Introduction
The nonlinear optimization is a fundamental subject in the modern optimization literature. It focuses on the problem of optimizing an objective function in the presence of inequality and/or equality constraints. Furthermore, the optimization problem is obviously linear if all the functions are linear, otherwise it is called a nonlinear optimization problem.
This research field is motivated by the fact that it arises in various problems encountered in practice, such as business administration, economics, agriculture, mathematics, engineering, and physical sciences.
In our knowledge, Frank and Wolfe are the deans in nonlinear optimization problems. They established a powerful algorithm in [1] to solve them. Later, they used another method in [2] based on the application of the Simplex method on the nonlinear problem after converting it to a linear one.
This pioneer work inspired many authors to propose and develop several methods and techniques to solve this class of problems. We refer to [3,4] for interior point methods to find the solution of nonlinear optimization problems with a high dimension.
In order to make this theory applicable in practice, other methods are designed on the linear optimization history, among robust algorithms with polynomial complexity. In this perception, Khachian succeeded in 1979 to introduce a new ellipsoid method from approaches applied originally to nonlinear optimization.
Interior point methods outperform the Simplex ones, and they have recently been the subject of several monographs including Bonnans and Gilbert [5], Evtushenko and Zhadan [6], Nesterov and Emirovski [7], and Wright [8] and Ye [9].
Interior point methods can be classified into three different groups as follows: projective methods and their alternatives as in Powell [10] and Rosen [11,12], central trajectory methods (see Ouriemchi [13] and Forsgren et al. [14]), and barrier/penalty methods, where majorant functions were originally proposed by Crouzeix and Merikhi [15] to solve a semidefinite optimization problem. Inspired by this work, Menniche and Benterki [16] and Bachir Cherif and Merikhi [17] applied this idea to linear and nonlinear optimizations, respectively.
A majorant function for the penalty method in convex quadratic optimization was proposed by Chaghoub and Benterki [18]. On the other hand, A. Leulmi et al. [19,20] used new minorant functions for semidefinite optimization, and this idea was extended to linear programming by A. Leulmi and S. Leulmi in [21].
As far as we know, our new approximate function has not been studied in the nonlinear optimization literature. These approximate functions are more convenient and efficient than the line search method for rapidly computing the displacement step.
Therefore, in our work, we aim to optimize a nonlinear problem based on prior efforts. Thus, we propose a straightforward and effective barrier penalty method using new minorant functions.
More precisely, we first introduce the position of the problem and its perturbed problem with the results of convergence in Section 2 and Section 3 of our paper. Then, in Section 4, we establish the solution of the perturbed problem by finding new minorant functions. Section 5 is devoted to presenting a concise description of the algorithm and to illustrating the outperformance of our new approach by carrying out a simulation study. Finally, we summarize our work in the conclusion.
Throughout this paper, the following notations are adopted. Let and denote the scalar product and the Euclidean norm, respectively, given by the following:
and
2. The Problem
We aim to present an algorithm for solving the following optimization problem:
where and A is a full-rank matrix with
For this purpose, we need the following hypothesis:
Hypothesis 1.
f is nonlinear, twice continuously differentiable, and convex on , where is the set of realizable solutions of (P).
Hypothesis 2.
(P) satisfies the condition of interior point (IPC), i.e., there exists such that
Hypothesis 3.
The set of optimal solutions of (P) is nonempty and bounded.
Notice that these conditions are standard in this context. We refer to [17,20].
If is an optimal solution, there exist two Lagrange multipliers and such that
3. Formulation of the Perturbed Problem of (P)
Let us first consider the function defined on by the following:
where is a convex, lower semicontinuous and proper function given by the following:
Thus, is a proper, convex, and lower semicontinuous function.
Furthermore, the function g defined by
is convex. Notice that for the perturbed problem coincides with the initial problem (P); then,
3.1. Existence and Uniqueness of Optimal Solution
To show that the perturbed problem has a unique optimal solution, it is sufficient to demonstrate that the recession cone of is reduced to zero.
Proof.
For a fixed the function is proper, convex, and lower semicontinuous. The asymptotic function of is defined by the following:
thus, the asymptotic functions of f and satisfy the relation:
Moreover, hypothesis H3 is equivalent to
Then,
and from [17], for each non-negative real number the strictly convex problem admits a unique optimal solution noted by The solution of the problem (P) is the limit of the solutions sequence of the perturbed problem when tends to 0. □
3.2. Convergence of the Solution
Now, we are in a position to state the convergence result of to (P), which is proved in Lemma 1 on [18].
Let for all we define
Lemma 1
([18]). We consider If the perturbed problem admits an optimal solution such that , then the problem (P) admits an optimal solution
We use the classical prototype of penalty methods. We begin our process with
where
and the iteration scheme is divided into the following steps:
1. Select
2. Establish an approximate solution for It is obvious that
Remark 1.
If the values of the objective functions of the problem (P) and the perturbed problem are equal and finite, then (P) will have an optimal solution if and only if has an optimal solution.
The iterative process stops when we obtain an acceptable approximation of
4. Computational Resolution of the Perturbed Problem
Our approach to the numerical solution of the perturbed problem consists of two stages. In the first one, we calculate the descent direction using the Newton approach, and in the second one, we propose an efficient new-minorant-functions approach to compute the displacement step easily and quickly relative to the line search method.
4.1. The Descent Direction
As is strictly convex, the necessary and sufficient optimality conditions state that is an optimal solution of if and only if it satisfies the nonlinear system:
Using the Newton approach, a penalty method is provided to solve the above system, where the vector in each is given by
The solution of the following quadratic convex optimization problem is necessary to obtain the Newton descent direction d:
where and
with the diagonal matrix
The Lagrangian is given by the following:
where is the Lagrange multiplier. It is sufficient for solving the linear system equations with
then,
It is simple to prove that system (3) is non-singular. We obtain
As and we obtain
The system can also be written as follows:
Thus, the Newton descent direction is obtained.
Throughout this paper, we take x instead of
4.2. Computation of the Displacement Step
This section deals with the numerical solution of the displacement step. We give a brief highlight of the line search methods used in nonlinear optimization problems. Then, we collect some important results of approximate function approaches applied to both semidefinite and linear programming problems. Finally, we propose our new approximate function method for the nonlinear optimization problem (P).
4.2.1. Line Search Methods
The line search methods consists of determining a displacement step , which ensures the sufficient decrease in the objective at each iteration where along the descent direction in other words, it involves solving the following one-dimensional problem:
The disadvantage of this method is that the solution is not necessarily optimal, which make the feasibility of not guaranteed.
The line search techniques of Wolfe, Goldstein-Armijo, and Fibonacci are the most widely used ones. However, generally, their computational volume is costly. This is what made us search for another alternative.
4.2.2. Approximate Functions Techniques
These methods are based on sophisticated techniques introduced by J.P. Crouzeix et al. [15] and A. Leulmi et al. [20] to obtain the solution of a semidefinite optimization problem.
The aim of these techniques is to give a minimized approximation of one real-variable function defined by
The function is convex, and we obtain the following:
We find that deduced from (4), which is expected since d is the direction of Newton’s descent direction.
We aim to avoid the disadvantages of line search methods and accelerate the convergence of the algorithm. For this reason, we have to identify an that yields a significant decrease in the function This is the same as solving a polynomial equation of degree where f is a linear function.
Now, we include a few helpful inequalities below, which are used throughout the paper.
H. Wolkowicz et al. [22] see also Crouzeix and Seeger [23] presented the following inequalities:
where and represent the mean and the standard deviation, respectively, of a statistical real numbers series . The later quantities are defined as follows:
Theorem 1
([15]). Let for We have the following:
where , and
We will proceed to present the paper’s principal result.
4.2.3. New Approximate Functions Approach
Let
be defined on such that
To find the displacement step, it is necessary to solve Considering the difficulty of solving a non-algebraic equation, approximate functions are recommended alternatives.
Two novel approximation functions of are introduced in the following lemma.
Lemma 2.
For all with we have
and for all with we obtain
where
and
with
Furthermore, we have
Proof.
We start by proving that
Theorem 1 gives
then,
and
Hence,
Therefore,
Let us consider the following:
We have
Because of the fact that and it is easy to see that
Therefore,
then, □
Hence, the domain of is included in the domain of which is where
Let us remark that
Thus, is well approximated by in a neighborhood of Since is strictly convex, it attains its minimum at one unique point which is the unique root of the equation . This point belongs to the domain of Therefore, is bounded from below by
And it is also bounded from below by
Then, gives an apparent decrease in the function
4.3. Minimize an Auxiliary Function
We now consider the minimization of the function
and we also have the following approximate function:
where is defined in (7). Then, we have the following:
and
We remark that for
We present the conditions and The function is strictly convex. It attains its minimum at one unique point such that which is one of the roots of the equations
and
For Equation (8), the roots are explicitly calculated, and we distinguish the following cases:
- If we obtain
- If we obtain
- If we have
- If is the only root of the second-degree equation that belongs to the domain of definition of We obtain Both roots areand
Then, the root of Equation (9) is explicitly calculated, and we have
Consequently, we compute the two values , explicitly. Then, we take , where is a fixed precision and
Remark 2.
The computation of is performed through a dichotomous procedure in the cases where and , as follows:
Put .
While do
If then,
else so
This computation guarantees a better approximation of the minimum of while remaining in the domain of
4.4. The Objective Function f Is
4.4.1. Linear
For all there exists such that
The minimum of is reached at the unique root of the equation Then,
Take in the auxiliary function The two functions and coincide.
yields a significant decrease in the function along the descent direction It is interesting to note that the condition implies the following:
4.4.2. Convex
is no longer constant, and the equation is not reduced to one equation of a second degree for
We consider another function less than Given we have, for all , the following:
then,
and
We choose in the auxiliary function , and we compute the root of the equation with
Therefore, we have two cases:
- Where We have the following:and, thus, along the direction we obtain a significant decrease in the function The approximation accuracy of by being better for small values of it is recommended to use a new value of situated between and the former for the next iteration. Moreover, the cost of the supplementary computation is small since it is the cost of one evaluation of f and the resolution of a second-order equation.
- Where The computation of is performed through a dichotomous procedure (see Remark 3).
5. Description of the Algorithm and Numerical Simulations
5.1. Description of the Algorithm
This section is devoted to introducing our algorithm for obtaining an optimal solution of (P).
Begin
Initialization
is a given precision. and are given.
is a strictly realizable solution from
Iteration
- Start with
- Calculate d and
- If calculate , and
- Determine following (8), (10), or (9) depending on the linear or nonlinear case.
- Take the new iterate and go back to step 2.
- If a well approximation of has been obtained.
- (a)
- If and return to step 2.
- (b)
- If STOP: a well approximate solution of has been obtained.
End algorithm.
The aim of this method is to reduce the number of iterations and the time consumption. In the next section, we provide some examples.
5.2. Numerical Simulations
To assess the superior performance and accuracy of our algorithm, based on our minorant functions, numerical tests are conducted to make comparisons between our new approach and the classical line search method.
For this purpose, in this section, we present comparative numerical tests on different examples taken from the literature [5,24].
We report the results obtained by implementing the algorithm in MATLAB on an Intel Core i7-7700HQ (2.80 GHz) machine with 16.00 Go RAM.
5.2.1. Examples with a Fixed Size
Nonlinear Convex Objective
Example 1.
Let us take the following problem:
The optimal value is and the optimal solution is
Example 2.
Let us take the following problem:
The optimal value is and the optimal solution is
Example 3.
Let us consider the following problem:
The optimal value is and the optimal solution is
This table presents the results of the previous examples:
| Example | st1 | st2 | LS | ||||||
| iter | Time (s) | iter | Time (s) | iter | Time (s) | ||||
| 1 | 12 | 0.0006 | 19 | 0.0015 | 6 | 0.0091 | |||
| 2 | 5 | 0.0004 | 9 | 0.0009 | 44 | 0.099 | |||
| 3 | 3 | 0.0001 | 5 | 0.0006 | 65 | 0.89 |
5.2.2. Example with a Variable Size
The Objective Function f Is
1-Linear: Let us consider the linear programming problem:
where A is an matrix given by the following:
where
The results are presented in the table below.
| Size | st1 | st2 | LS | |||||
| iter | Time (s) | iter | Time (s) | iter | Time (s) | |||
| 1 | 0.0021 | 2 | 0.0039 | 9 | 0.0512 | |||
| 1 | 0.0031 | 3 | 0.0045 | 13 | 0.0821 | |||
| 2 | 0.0049 | 3 | 0.0032 | 17 | 0.3219 | |||
| 2 | 0.0053 | 4 | 0.0088 | 19 | 0.5383 | |||
| 2 | 0.0088 | 4 | 0.0098 | 22 | 0.9220 | |||
| 3 | 0.0096 | 5 | 0.0125 | 26 | 9.2647 |
2-Nonlinear:
Example 4
(Quadratic case [13]). Let the quadratic problem be as follows:
with Q is the matrix defined for by the following:
This example is tested for many values of
The obtained results are given by the following table:
| ex() | st1 | st2 | LS | ||||
| iter | Time (s) | iter | Time (s) | iter | Time (s) | ||
| 5 | 0.9968 | 4 | 0.9699 | 26 | 19.5241 | ||
| 7 | 18.1448 | 5 | 9.6012 | 35 | 86.1259 | ||
| 12 | 36.3259 | 5 | 19.0099 | 23 | 98.2354 | ||
| 21 | 56.9912 | 17 | 41.1012 | 33 | 109.2553 | ||
| 28 | 140.1325 | 23 | 95.6903 | 40 | 1599.1596 |
Example 5
(The problem of Erikson [25]). Let the following be the quadratic problem:
where and are fixed.
This example is tested for different values of , and
The following table resumes the obtained results in the case
| ex() | st1 | st2 | LS | ||||
| iter | Time (s) | iter | Time (s) | iter | Time (s) | ||
| 1 | 0.0001 | 2 | 0.0012 | 4 | 0.0236 | ||
| 2 | 0.0021 | 3 | 0.0033 | 5 | 0.7996 | ||
| 2 | 0.0043 | 3 | 0.0201 | 5 | 1.5289 | ||
| 2 | 3.0901 | 4 | 5.9619 | 12 | 22.1254 |
In the above tables, we take = 1.0 × .
We also denote the following:
- (iter) is the number of iterations.
- (time) is the computational time in seconds (s).
- (sti)i=1,2 represents the strategy of approximate functions introduced in this paper.
- (LS) represents the classical line search method.
Commentary: The numerical tests carried out show, without doubt, that our approach leads to a very significant reduction in the cost of calculation and an improvement in the result. When comparing the approximate functions to the line search approach, the number of iterations and computing time are significantly reduced.
6. Conclusions
The contribution of this paper is particular focused on the study of nonlinear optimization problems by using the logarithmic penalty method based on some new approximate functions. We first formulate the problems and with the results of the convergence. Then, we find their solutions by using new approximate functions.
Finally, to lend further support to our theoretical results, a simulation study is conducted to illustrate the good accuracy of the studied approach. More precisely, our new approximate functions approach outperforms the line search one as it significantly reduces the cost and computing time.
Funding
This work has been supported by: The General Directorate of Scientific Research and Technological Development (DGRSDT-MESRS) under the PRFU project number C00L03UN190120220009. Algeria.
Data Availability Statement
Data are contained within the article.
Acknowledgments
The author is very pleased to thank the editor and the reviewers for their helpful suggestions and comments.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Frank, M.; Wolfe, B. An algorithm for quadratic programming. Nav. Res. Logist. Q. 1956, 3, 95–110. [Google Scholar] [CrossRef]
- Wolfe, P. A Duality Theorem for Nonlinear Programming. Q. Appl. Math. 1961, 19, 239–244. [Google Scholar] [CrossRef]
- Bracken, J.; M1cCormiek, G.P. Selected Applications of Nonlinear Programming; John Wiley & Sons, Inc.: New York, NY, USA, 1968. [Google Scholar]
- Fiacco, A.V.; McCormick, G.P. Nonlinear Programming: Sequential Unconstrained Minimization Techniques; John Wiley & Sons, Inc.: New York, NY, USA, 1968. [Google Scholar]
- Bonnans, J.-F.; Gilbert, J.-C.; Lemaréchal, C.; Sagastizàbal, C. Numerical Optimization: Theoretical and Practical Aspects; Mathematics and Applications; Springer: Berlin/Heidelberg, Germany, 2003; Volume 27. [Google Scholar]
- Evtushenko, Y.G.; Zhadan, V.G. Stable barrier-projection and barrier-Newton methods in nonlinear programming. In Optimization Methods and Software; Taylor & Francis: Abingdon, UK, 1994; Volume 3, pp. 237–256. [Google Scholar]
- Nestrov, Y.E.; Nemiroveskii, A. Interior-Point Polynomial Algorithms in Convex Programming; SIAM: Philadelphia, PA, USA, 1994. [Google Scholar]
- Wright, S.J. Primal–Dual Interior Point Methods; SIAM: Philadelphia, PA, USA, 1997. [Google Scholar]
- Ye, Y. Interior Point Algorithms: Theory and Analysis. In Discrete Mathematics Optimization; Wiley-Interscience Series; John Wiley & Sons: New York, NY, USA, 1997. [Google Scholar]
- Powell, M.J.D. Karmarkar’s Algorithm: A View from Nonlinear Programming; Department of Applied Mathematics and Theoretical Physics, University of Cambridge: Cambridge, UK, 1989; Volume 53. [Google Scholar]
- Rosen, J.B. The Gradient Projection Method for Nonlinear Programming. Soc. Ind. Appl. Math. J. Appl. Math. 1960, 8, 181–217. [Google Scholar] [CrossRef]
- Rosen, J.B. The Gradient Projection Method for Nonlinear Programming. Soc. Ind. Appl. Math. J. Appl. Math. 1961, 9, 514–553. [Google Scholar] [CrossRef]
- Ouriemchi, M. Résolution de Problèmes non Linéaires par les Méthodes de Points Intérieurs. Théorie et Algorithmes. Doctoral Thesis, Université du Havre, Havre, France, 2006. [Google Scholar]
- Forsgren, A.; Gill, P.E.; Wright, M.H. Interior Methods for Nonlinear Optimization; SIAM: Philadelphia, PA, USA, 2002; Volume 44, pp. 525–597. [Google Scholar]
- Crouzeix, J.P.; Merikhi, B. A logarithm barrier method for semidefinite programming. RAIRO-Oper. Res. 2008, 42, 123–139. [Google Scholar] [CrossRef]
- Menniche, L.; Benterki, D. A Logarithmic Barrier Approach for Linear Programming. J. Computat. Appl. Math. 2017, 312, 267–275. [Google Scholar] [CrossRef]
- Cherif, L.B.; Merikhi, B. A Penalty Method for Nonlinear Programming. RAIRO-Oper. Res. 2019, 53, 29–38. [Google Scholar] [CrossRef]
- Chaghoub, S.; Benterki, D. A Logarithmic Barrier Method Based on a New Majorant Function for Convex Quadratic Programming. IAENG Int. J. Appl. Math. 2021, 51, 563–568. [Google Scholar]
- Leulmi, A. Etude d’une Méthode Barrière Logarithmique via Minorants Functions pour la Programmation Semi-Définie. Doctoral Thesis, Université de Biskra, Biskra, Algeria, 2018. [Google Scholar]
- Leulmi, A.; Merikhi, B.; Benterki, D. Study of a Logarithmic Barrier Approach for Linear Semidefinite Programming. J. Sib. Fed. Univ. Math. Phys. 2018, 11, 300–312. [Google Scholar]
- Leulmi, A.; Leulmi, S. Logarithmic Barrier Method via Minorant Function for Linear Programming. J. Sib. Fed. Univ. Math. Phys. 2019, 12, 191–201. [Google Scholar] [CrossRef]
- Wolkowicz, H.; Styan, G.P.H. Bounds for Eigenvalues Using Traces. Lin. Alg. Appl. 1980, 29, 471–506. [Google Scholar] [CrossRef]
- Crouzeix, J.-P.; Seeger, A. New bounds for the extreme values of a finite sample of real numbers. J. Math. Anal. Appl. 1996, 197, 411–426. [Google Scholar] [CrossRef][Green Version]
- Bazraa, M.S.; Sherali, H.D.; Shetty, C.M. Nonlinear Programming, Willey-Interscience; John Wiley & Sons, Inc.: Hoboken, NJ, USA; Toronto, ON, Canada, 2006. [Google Scholar]
- Shannon, E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423+623–656. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).