Abstract
The convergence rate of the augmented Lagrangian method (ALM) for solving the nonlinear semidefinite optimization problem is studied. Under the Jacobian uniqueness conditions, when a multiplier vector and the penalty parameter are chosen such that is larger than a threshold and the ratio is small enough, it is demonstrated that the convergence rate of the augmented Lagrange method is linear with respect to and the ratio constant is proportional to , where is the multiplier corresponding to a local minimizer. Furthermore, by analyzing the second-order derivative of the perturbation function of the nonlinear semidefinite optimization problem, we characterize the rate constant of local linear convergence of the sequence of Lagrange multiplier vectors produced by the augmented Lagrange method. This characterization shows that the sequence of Lagrange multiplier vectors has a Q-linear convergence rate when the sequence of penalty parameters has an upper bound and the convergence rate is superlinear when is increasing to infinity.
Keywords:
nonlinear semidefinite optimization problem; Jacobian uniqueness conditions; augmented Lagrange method; convergence rate; perturbation function MSC:
90C30
1. Introduction
Consider the following nonlinear semidefinite optimization problem
where function , mapping and mapping are assumed to be twice continuously differentiable in a neighborhood of a given feasible point , where is the space of real symmetric matrices.
The augmented Lagrange method was initiated by Powell [1] and Hestenes [2] for solving equality constrained nonlinear programming problems and was extended by Rockafellar [3] to inequality constrained optimization problems. For convex programming, Rockafellar [3,4] adopted the augmented Lagrange function for establishing a saddle point theorem and demonstrated the global convergence of the augmented Lagrange method when the penalty parameter is chosen as an arbitrary positive number. Rockafellar [5] gave a deep study of the augmented Lagrange method for convex optimization.
The study of local convergence properties for the augmented Lagrange method is fairly comprehensive. For the optimization problems with equality constraints, Powell [1] proved the local linear convergence of the augmented Lagrange method to a local minimum point when the second-order sufficient condition and the linear independence constraint qualification are satisfied. This result was extended by Bertsekas ([6], Chapter 3) to the optimization problem with inequality constraints under the strict complementarity condition, in which the linear rate constant is proportional to . If the strict complementarity condition is not satisfied, Ito and Kunisch [7], Conn et al. [8] and Contesse-Becker [9] proved that the augmented Lagrange method has linear convergence rate.
The Lagrange function for Problem (1) can be written as
The augmented Lagrange function for (1) is defined by
where denotes the projection operator to . The augmented Lagrange method for Problem (1) can be expressed in the following form:
- Step 0
- Given parameter , initial point , initial multiplier and .
- Step 1
- Ifthen stop and is a Karush–Kuhn–Tucker (KKT) pair.
- Step 2
- Solve the following problemand calculate
- Step 3
- Update , set to k, and go to Step 1.
For nonlinear semidefinite optimization problem, in the appendix of Sun et al. [10], they used a direct way to derive the linear rate of convergence when the strict complementarity condition holds. However, this result on the rate of convergence of the augmented Lagrange method obtained in [10] has the possibility for improvement. For example, can we obtain a similar result to ([6], Chapter 3) for equality constrained optimization problems when is very small? How can we characterize the rate constant of the local linear convergence of the augmented Lagrangian method? In this paper, we will give positive answers to these two questions.
It should be noted that there are a lot of important applications for augmented Lagrangian methods in different types of optimization problems; for examples, see [11,12,13].
The paper is organized as follows. In the next section, we develop properties of the augmented Lagrange function under the Jacobian uniqueness conditions for the semidefinite optimization problem (1), which will be required to prove results on the convergence rate of the augmented Lagrange method. In Section 3, we demonstrate the linear rate of convergence of the augmented Lagrangian method for the semidefinite optimization problem when the Jacobian uniqueness conditions are satisfied. In Section 4, we establish the asymptotical convergence rate of Lagrange multipliers, which shows that the sequence of Lagrange multiplier vectors produced by the augmented Lagrange method is convergent to the optimal Lagrange multiplier superlinearly when the sequence is increasing to ∞. Finally, we draw a conclusion in Section 5.
We list two technical results at the end of this section, which will be used in developing properties of the augmented Lagrange function for proving the main theorem about the convergence rate of the ALM. The first technical result is a variant of [14] and the second result is an implicit function theorem from page 12 of Bertsekas [6].
Lemma 1.
Let X and Y be two finite dimensional Hilbert spaces and be continuous and positive homogeneous of degree 2, namely
Suppose that there exists a real number such that for any w satisfying , where is a linear mapping. Then, there exist positive real numbers and such that
Lemma 2.
Assume that be an open subset of , Σ be a nonempty compact subset of , and be a mapping and on for some . Assume that exists, and it is continuous on . Assume that is a vector such that for , and the Jacobian is nonsingular for all . Then, there exist scalars , and a mapping such that on , for all , and for all . The mapping ψ is unique in the sense that if , and , then . Furthermore, if , then we have
2. Properties of the Augmented Lagrangian
Assume that is a given feasible point of Problem (1) and f, h and g are twice differentiable in a neighborhood of . The following conditions, which are called Jacobian uniqueness conditions, are needed in our analysis.
Definition 1.
Jacobian uniqueness conditions at are the following conditions:
- (i)
- The point satisfies the Karush–Kuhn–Tucker condtions:
- (ii)
- The constraint nondegeneracy condition is satisfied at :where denotes the linearity space of a closed convex cone.
- (iii)
- The strict complementarity condition at holds, namely .
- (iv)
- At , the second-order sufficiency optimality conditions holds, namely for any satisfying ,where is the Moore–Penrose pseudoinverse of and is the critical cone at defined by
In this section, we will give some properties of the Jacobian uniqueness conditions of Problem (1) and properties of the augmented Lagrange function under this set of conditions. These properties are crucial for studying the convergence rate of augmented Lagrange method.
Let be a KKT pair. Assume that (iii) holds; then, is nonsingular. Let the eigenvalues of be and
Then, an orthogonal matrix exists such that
where
If Jacobian uniqueness conditions (i)–(iii) hold, then the cone is reduced to the following subspace
If Jacobian uniqueness condition (iv) holds, then there exists such that
In fact, if this is not true, then a sequence with exists such that
There exists a subsequence and with such that . The closedness of implies . Taking the limit of (5) along the subsequence , we obtain
which contradicts Jacobian uniqueness condition (iv).
Define
Then, the Jacobian of , denoted by , is expressed as
Lemma 3.
Let be a given point and f, h and g be twice continuously differentiable in a neighborhood of . Let the Jacobian uniqueness conditions at are satisfied. Then, is a nonsingular linear operator.
Proof.
Consider the equation
where . This equation is equivalent to
This implies the following relations
From and , we have . Multiplying to the first equality of (8) we obtain
which implies from Jacobian uniqueness condition (iv). This comes from the fact that implies
from Jacobian uniqueness condition (iv). Then, from the first equality of (8) we obtain
which is equivalent to
This, from , implies
From Jacobian uniqueness condition (ii) we obtain
and this implies and . Combining , we obtain that is a nonsingular linear operator. □
Proposition 1.
Let be a Karush–Kuhn–Tucker point of Problem (1) and the Jacobian uniqueness conditions are satisfied at . Then there exist positive numbers and such that is positively definite when and .
Proof.
It is easy to obtain
If is nonsingular, then is differentiable at and
Define
Then, from (4), we obtain for any vector ,
This implies for any vector that
It follows from (ii) that the linear mapping
is onto. Then, we have from Lemma 1 that there exists such that is positively definite if . Therefore, there exists a positive real number such that is positively definite if and . □
Suppose that is nonsingular such that is differentiable at . In this case, we define a linear operator:
Proposition 2.
Let be a Karush–Kuhn–Tucker point of Problem (1) and the Jacobian uniqueness conditions are satisfied at . Then, there exists a positive real number large enough, such that is nonsingular and
for some positive constant if , where .
Proof.
We divide the proof into two steps.
- Step 1:
- we prove that for sufficiently small is nonsingular.
Since we have from Lemma 3 that is nonsingular. Now, we consider the case where . Consider the equation
where . This equation is equivalent to
From the second equality of (11), we have
This implies the following relations
Then, multiplying to the first equality of (11), we obtain
which implies from Proposition 1 when . Therefore, we obtain , , and so that is nonsingular when .
- Step 2:
- We prove thatfor some positive constant if small enough.
Noting, for we have and . Therefore, we get
For any we have that
For any matrix , we have
where
Thus, we have, for , that
Therefore, there exists a sufficiently large positive number , for , if ; then, is nonsingular and
for some positive constant . The proof is complete. □
Proposition 3.
The corresponding Löwner operator F is twice (continuously) differentiable at X if and only if f is twice (continuously) differentiable at , .
Proposition 4.
Let be a Karush–Kuhn–Tucker point of Problem (1) and the Jacobian uniqueness conditions are satisfied at . Then, there exist , , and , for , is nonsingular and
if and .
Proof.
We have from Proposition 2 that the operator is nonsingular. Since the norm of is less than 1 and
we have
Since is twice continuously differentiable at we obtain
for . For we obtain
where
Note for that
we have
Therefore, from
we get
where
3. The Convergence Rate of the Augmented Lagrange Method
In this section, we focus on the local convergence of the augmented Lagrange method for nonlinear semidefinite optimization problems under the Jacobian uniqueness conditions. Now we estimate the solution error of the augmented Lagrange subproblem
and the error for the updating multiplier when is around . The local convergence and the linear rate of multipliers can be obtained by using these estimates.
For a real number define
Theorem 1.
Let Jacobian uniqueness conditions be satisfied at . Then, there exist and and such that for any the problem
has a unique solution , which is smooth on . Moreover, for ,
where
Proof.
If x is a local minimum point of Problem
then, from the definition of , we obtain
Obviously, we have
and
Obviously, from the definition in (10), we have . Then, from Proposition 4, we have that is nonsingular when .
Define and and
From the implicit function theorem, we have that there exists with , and mapping
which is smooth on and satisfies
From Propositions 1 and 4, we may choose and small enough such that constraint nondegeneracy condition holds at , , is positively definite and for all .
Differentiating the three equations in (21) with respect to , we obtain
where . Define and . Then, we have from (22), for that
Noting that for and , we obtain from (23) and that
Noting that is twice continuously differentiable at , we have
It is easy to check the equality . Then, when is chosen small enough, there exists a positive constant such that
when and .
Combining this estimate with (24), we obtain
Substituting by in (25) yields
Define
From the definitions of and K, we have that
Noting that and and we have from Proposition 1 that
According to the above theorem, it is easy for us to prove the local convergence properties of the augmented Lagrange method for the nonlinear semidefinite optimization problem.
Proposition 5.
Let satisfy Jacobian uniqueness conditions. Let and be given in Theorem 1. Suppose that , and satisfy
Then, the sequence generated by the ALM is convergent to with
if . The sequence converges superlinearly to when .
Proof.
For the sequence generated by the ALM, we obtain from Theorem 1 that
which implies
and
Suppose that satisfies and , then for , from Theorem 1 we have that
which implies
and
Therefore, by induction, we obtain that for any and . Then for , we obtain
which implies
4. Asymptotical Superlinear Convergence of Multipliers
In Theorem 1, the convergence rate of the augmented Lagrange method is characterized by (19), which involves a constant . The means by which to give an estimate of are an important topic. In this section, we estimate using the eigenvalues of the second-order derivative of the perturbation function of Problem (1).
Let be a Kurash–Kuhn–Tucker point of Problem (1), consider the following system of equations in ,
then, is a solution of (32) for any where . According to the implicit function theorem, there exist a constant and mappings such that
and for , where ,
Moreover, there exists such that
for . Define the function as follows
In view of the Jacobian uniqueness conditions, and can be taken small enough so that is actually a local minimum point in of the following perturbed problem
Thus, the function p is actually the following perturbation function:
Lemma 4.
Suppose that Jacobian uniqueness conditions hold and ε and δ are taken sufficiently small such that is a local minimum point of Problem (35). Then,
Proof.
We use to denote the Lagrange function of Problem (35), namely
Then, is expressed as follows
Noting and , from the above expression of we obtain
The proof is complete. □
Lemma 5.
Suppose that Jacobian uniqueness conditions hold and ε and δ are taken sufficiently small so that is a local minimum point of the perturbed problem (35). Then,
Proof.
Thus, Equation (41) is equivalent to
Therefore, we get that
which implies
It follows from Page 20 of [15] that the inverse of can be expressed as
It is easy to check
which implies
where
namely, the equality (38) holds. □
Corollary 1.
Let Jacobian uniqueness conditions be satisfied at . Then,
where .
Proof.
By using the above properties, we are able to analyze the rate of convergence of multipliers generated by the augmented Lagrange method. For this purpose, we first give an equivalent expression of
which is a key property for analyzing the superlinear rate of the sequence of multipliers.
Theorem 2.
Let the Jacobian uniqueness conditions be satisfied at . Let , δ and ε be given by Theorem 1. Then, for all ,
where is defined by
and .
Proof.
Define
Noting that is equivalent to = 0, we have
Differentiating the last three equations in (47) with respect to , we obtain
Denoting and , we have from (48) that
We can easily obtain the following expression of :
From the equality
that
with
Then, we get
Substituting , , , , and , . We obtain the desired result. □
Theorem 3.
Assume that satisfies the Jacobian uniqueness y conditions, , δ and ε are the constants given by Theorem 1. Suppose that
Then, there exists a scalar such that if and satisfy
then the sequence generated by
is well-defined, and and . Furthermore if
and for all k, then
while if and for all k, then
Proof.
In view of of (46), we have that
5. Conclusions
In this paper, we have studied the convergence rate of the augmented Lagrangian method for the nonlinear semidefinite optimization problem. We have proven the local linear rate of convergence of the sequence of multipliers and that the ratio constant is proportional to when exceeds a threshold and the ratio is sufficiently small. Importantly, based on the second-order derivative of the perturbation function of the nonlinear semidefinite optimization problem, we have obtained an accurate estimation for the rate constant of the linear convergence of multiplier vectors generated by the augmented Lagrange method, which shows that the sequence of multipliers is superlinear convergent if is increasing to ∞.
There are many unsolved problems left in the augmented Lagrange method for nonlinear semidefinite optimization problems. First, in Theorem 1, the result on the convergence rate of the augmented Lagrange method is obtained when the subproblems are exactly solved. A natural problem is how to analyze the convergence rate of the ALM when the subproblems are solved inexactly. Second, all results in this paper are about local convergence of the augmented Lagrange method, global convergent augmented Lagrangian methods are worth studying. Third, for estimating the rate constant of linear convergence, we need the strict complementarity condition; this is a critical condition. What about the convergence properties of the augmented Lagrange method when this condition does not hold?
Author Contributions
Methodology, Y.Z., J.W. and H.L.; validation, J.W.; formal analysis, Y.Z. and J.Z.; writing—original draft preparation, Y.Z. and H.L.; writing—review and editing, J.W., J.Z. and H.L.; funding acquisition, Y.Z. and J.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Key R&D Program of China under project number 2022YFA1004000 and National Natural Science Foundation of China (Nos. 12201097 and 12071055).
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
Acknowledgments
The authors have reviewed and edited the output and take full responsibility for the content of this publication.
Conflicts of Interest
The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.
References
- Powell, M.J.D. A method for nonlinear constraints in minimization problems. In Optimization; Academic Press: London, UK, 1969; pp. 283–298. [Google Scholar]
- Hestenes, M.R. Multiplier and gradient methods. J. Optim. Theory Appl. 1969, 4, 303–320. [Google Scholar] [CrossRef]
- Rockafellar, R.T. A dual approach to solving nonlinear programming problems by unconstrained optimization. Math. Program. 1973, 5, 354–373. [Google Scholar] [CrossRef]
- Rockafellar, R.T. The multiplier method of Hestenes and Powell applied to convex programming. J. Optim. Theory Appl. 1973, 12, 555–562. [Google Scholar] [CrossRef]
- Rockafellar, R.T. Augmented Lagrangians and applications of the proximal point algorithm in convex programming. Math. Oper. Res. 1976, 1, 97–116. [Google Scholar] [CrossRef]
- Bertsekas, D.P. Constrained Optimization and Lagrange Multiplier Methods; Academic Press: Cambridge, MA, USA, 1982. [Google Scholar]
- Ito, K.; Kunisch, K. The augmented Lagrangian method for equality and inequality constraints in Hilbert spaces. Math. Program. 1990, 46, 341–360. [Google Scholar] [CrossRef]
- Conn, A.R.; Gould, N.I.M.; Toint, P.L. A globally convergent augmented Lagrangian algorithm for optimization with general constraints and simple bounds. SIAM J. Numer. Anal. 1991, 28, 545–572. [Google Scholar] [CrossRef]
- Contesse-Becker, L. Extended convergence results for the method of multipliers for non-strictly binding inequality constraints. J. Optim. Theory Appl. 1993, 79, 273–310. [Google Scholar] [CrossRef]
- Sun, D.F.; Sun, J.; Zhang, L.W. The rate of convergence of the augmented Lagrangian method for nonlinear semidefinite programming. Math. Program. 2008, 114, 349–391. [Google Scholar] [CrossRef]
- Hang, N.T.V.; Sarabi, M.E. local convergence analysis of augmented lagrangian methods for piecewise linear-quadratic composite optimization problems. SIAM J. Optim. 2021, 31, 2665–2694. [Google Scholar] [CrossRef]
- Kočvara, M.; Stingl, M. PENNON: A code for convex nonlinear and semidefinite programming. Optim. Methods Softw. 2003, 18, 317–333. [Google Scholar] [CrossRef]
- El Yazidi, Y.; Ellabib, A. Augmented Lagrangian approach for a bilateral free boundary problem. J. Appl. Math. Comput. 2021, 67, 69–88. [Google Scholar] [CrossRef]
- Debreu, G. Definite and semidefinite quadratic forms. Econometrica 1952, 20, 295–300. [Google Scholar] [CrossRef]
- Zhang, F.Z. The Schur Complement and Its Application; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).