Next Article in Journal
Nonlinear Analysis for a Type-1 Diabetes Model with Focus on T-Cells and Pancreatic β-Cells Behavior
Previous Article in Journal
Mechanism of Coup and Contrecoup Injuries Induced by a Knock-Out Punch
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Existence of Generalized Augmented Lagrange Multipliers for Constrained Optimization Problems

1
Department of Statistics, School of Mathematics and Statistics, Shandong University of Technology, Zibo 255049, China
2
School of Mathematics and Statistics, Xinyang Normal University, Xinyang 464000, China
*
Author to whom correspondence should be addressed.
Math. Comput. Appl. 2020, 25(2), 24; https://doi.org/10.3390/mca25020024
Submission received: 17 March 2020 / Revised: 4 April 2020 / Accepted: 22 April 2020 / Published: 24 April 2020

Abstract

:
The augmented Lagrange multiplier as an important concept in duality theory for optimization problems is extended in this paper to generalized augmented Lagrange multipliers by allowing a nonlinear support for the augmented perturbation function. The existence of generalized augmented Lagrange multipliers is established by perturbation analysis. Meanwhile, the relations among generalized augmented Lagrange multipliers, saddle points, and zero duality gap property are developed.

1. Introduction

This paper is concerned with the following nonlinear programming problem:
( P ) min x Ω f ( x ) s . t . g i ( x ) 0 , i = 1 , , m , h j ( x ) = 0 , j = 1 , , l ,
where Ω is a nonempty and closed subset in R n , g i ( x ) : R n R for i = 1 , , m , and h j ( x ) : R n R for j = 1 , , l are continuous functions. For simplification of notation, let us denote g ( x ) : = ( g 1 ( x ) , g 2 ( x ) , , g m ( x ) ) , and h ( x ) : = ( h 1 ( x ) , h 2 ( x ) , , h l ( x ) ) . Note that the feasible region of ( P ) can be written as Ω F , where
F : = { x | g i ( x ) 0 , i = 1 , , m ; h j ( x ) = 0 , j = 1 , , l } .
The classical Lagrangian function for the problem ( P ) is defined as
L ( x , λ , μ ) : = f ( x ) + λ , g ( x ) + μ , h ( x ) , ( λ , μ ) R + m × R l .
A non-zero duality gap maybe arise for nonconvex optimization problems when using the above Lagrangian functions. Hence some modifications are necessary to overcome this difficulty, such as the augmented Lagrangian by introducing an augmented term, or the nonlinear Lagrangian by replacing the multiplier item and augmented term together by a nonlinear function. For example, the Hestenes–Powell–Rockafellar augmented Lagrangian [1,2,3], the cubic augmented Lagrangian [4], Mangasarian’s augmented Lagrangian [5,6], the exponential penalty function [7,8], the log-sigmoid Lagrangian [9], modified barrier functions [8,10], the p-th power augmented Lagrangian [11], and nonlinear augmented Lagrangian functions [12,13,14,15]. The other related discussion on augmented Lagrangians regarding special constrained optimization includes second-order cone programming [16,17], semidefinite programming [18,19,20], cone programming [21,22,23], semi-infinite programming [24,25], min-max programming [26], distributed optimization [27], mixed integer programming [28], stochastic mixed-integer programs [29], generalized Nash equilibrium problems [30], quasi-variational inequalities [31], composite convex programming [32], and sparse discrete problems [33].
The duality theory is closely related to the perturbation of primal problem. Precisely, for a given ( y , z ) R m × R l , the perturbation problem of ( P ) is
( P ( y , z ) ) min x Ω f ( x ) s . t . g i ( x ) + y i 0 , i = 1 , , m , h j ( x ) + z j = 0 , j = 1 , , l .
Denote by val ( P ) and v ( y , z ) ( : = val ( P ( y , z ) ) ) the optimal values of ( P ) and ( P ( y , z ) ) , respectively. Clearly, v ( 0 , 0 ) = val ( P ) . Denote by X * the optimal solution set of problem ( P ) , and assume throughout the paper that the optimal value val ( P ) is finite.
The augmented perturbation function is
v r ( y , z ) : = v ( y , z ) + r σ ( y , z ) , ( y , z ) R m × R l .
Here σ is called an augmenting function (see Section 2 below for details). Its properties are weakened from convex to level-bounded, or valley-at-zero. For example, in Rockafellar and Wets [34], a nonnegative convex augmenting function and the corresponding augmented Lagrangian dual problem of primal problem were introduced. A sufficient condition for the zero duality gap and a necessary and sufficient condition for the existence of an exact penalty representation were obtained. It was extended in [35] by replacing the convexity condition of the augmenting function with a level-boundedness condition. Using the theory of abstract convexity, a family of augmenting functions with almost peak at zero property and a class of corresponding augmented Lagrangian dual problems were introduced in [36]. Valley-at-zero property (similar to almost peak-at-zero property) was used in [37].
A vector ( λ , μ ) is said to be an augmented Lagrange multiplier for problem ( P ) (cf. [22,25]), if
v r ( y , z ) v r ( 0 , 0 ) + λ , y + μ , z , ( y , z ) R m × R l .
That means that ( λ , μ ) is a subgradient of v r ( · , · ) at ( y , z ) = ( 0 , 0 ) . The set of all subgradients ( λ , μ ) is called the subdifferential of v r ( y , z ) at ( y , z ) = ( 0 , 0 ) and denoted by v r ( 0 , 0 ) . Augmented Lagrange multipliers are an important concept in duality theory. Their existence is important for the global convergence analysis of primal-dual type algorithms based on the use of augmented Lagrangians [7,19,29,32,33]. In addition, augmented Lagrange multipliers are closely related to saddle points, the zero duality gap property, and exact penalty representation. Some results on the existence of augmented Lagrange multipliers are discussed for semi-infinite programming [25], cone programming [22,23], and eigenvalue composite optimization problems [38]. Moreover, CQ-free duality was proposed in the classical monograph [39] by Bonnans and Shapiro. The stronger results on CQ-free strong duality for semidefinite and general convex programming can be found in [40,41], and in more recent publications for semi-infinite, semidefinite, and copositive programming by Kostyukova and others [42,43]. Recently, Dolgopolik [44] studied the existence of augmented Lagrange multipliers for geometric constraint optimization by using the localization principle.
Recall that for convex programming, Lagrangian multiplier is a subgradient of perturbation function v at u = 0 in the sense of convex analysis; i.e.,
v ( u ) v ( 0 ) + λ , u , u .
For nonconvex programming, the Lagrangian multiplier can be used to estimate the subdifferential of the perturbation function at the origin. Precisely, for a minimization problem
min f ( x ) + θ ( g ( x ) ) , x X ,
where f : R n R , g : R n R m , X is a closed set in R n , and θ : R m R ¯ : = ( , + ] is proper, lsc, and convex. This model includes the constrained optimization problems (by letting θ be a indicator function) and composite optimization problems. Denote by S * the solution set. For x ¯ S * , let
M ( x ¯ ) : = { λ | 0 f ( x ¯ ) + λ T g ( x ¯ ) + N X ( x ¯ ) , λ θ ( g ( x ¯ ) ) }
and
M ( x ¯ ) : = { λ | 0 λ T g ( x ¯ ) + N X ( x ¯ ) , λ N dom θ ( g ( x ¯ ) ) } .
If X is regular and M ( x ¯ ) = { 0 } for every x ¯ S * , then
v ( 0 ) x ¯ S * M ( x ¯ ) , lip v ( 0 ) max λ M ( x ¯ ) x ¯ S * λ .
It should be pointed out that the subdifferential that appeared in (3) is the limiting/Mordukhovich subdifferential, not a subdifferential in the sense of convex analysis. Here M ( x ¯ ) = { 0 } can be regarded as constraint qualification. In particular, if dom θ : = R l × { 0 } l and X : = R n , then this condition is Mangasarian–Fromovitz constraint qualification; if dom θ is a convex cone with nonempty interior and X : = R n , then this condition is Robinson’s constraint qualification. The result (3) indicates that the Lagrangian multiplier provides an upper bound on the subdifferential of perturbation function and gives an estimate on the Lipschitz constant of perturbed function. It is very important for the convergence analysis of numerical algorithms.
Compared with the classical Lagrangian function, the augmented Lagrangian function has been successfully applied to study nonconvex programming. Hence an interesting question is how to use the augmented Lagrangian multiplier to study the subdifferential of v τ , and further give an estimate on Lipschitz constant on v τ . On subdifferentiability in nonconvex setting, Clarke’s pioneering work on generalized gradient opened the door to the study of general nonsmooth functions. Many concepts were introduced in the past few decades. Frequently used concepts include limiting/Mordukhovich subdifferential, Ioffe’s approximate and G-subdifferential, Michel and Penot’s subdifferential, Treiman’s linear subdifferential, Sussmann’s semidifferential, etc. Compared with the abstract subdifferential (pioneered by Warga), which is defined by a set of axioms, many subdifferentials have reasonable geometric explanations. For example, a convex subdifferential means a linear support, Frech e ´ t subdifferential means a smooth support, and a proximal subdifferential means a local quadratic support. The detailed discussion on other subdifferentials and their properties (particularly on calculus rules and the robust property) can be found in [34].
Clearly, the definition of an augmented Lagrangian multiplier given in (2) indicates that the augmented perturbation function is supported by a linear function at the origin. It corresponds to the subdifferential in the convex analysis. However, for a nonconvex setting, it is natural to consider whether a nonlinear support is available. Once it is done, we can establish and apply the duality theory in a more flexible environment. Define ω : R + R + such that ω ( η ) + as η + .
Definition 1.
A vector ( λ , μ ) is said to be a generalized augmented Lagrange multiplier of ( P ) , if there exists r 0 such that
v r ( y , z ) v r ( 0 , 0 ) + ϕ 1 ( λ , y ) + ϕ 2 ( μ , z ) , ( y , z ) R m × R l ,
where ϕ i for i = 1 , 2 possesses the following properties:
(A1) 
ϕ i is continuous and ϕ i ( · , 0 ) = 0 ;
(A2) 
ϕ i ( x , y + z ) ϕ i ( x , y ) + ϕ i ( x , z ) ;
(A3) 
x F , there exist a nonzero vector ( u 0 , v 0 ) and γ < 0 such that
ϕ 1 ( η u 0 , y ) + ϕ 2 ( η v 0 , z ) ω ( η ) γ ,
whenever ( y , z ) satisfies y + g ( x ) 0 , z + h ( x ) = 0 , and η > 0 is sufficiently large.
Since ϕ i includes the inner product as special cases, (4) is an essential extension of (2) from linear support to nonlinear support.
As mentioned above, the augmented Lagrange multiplier is a subgradient (in the sense of convex analysis) of an augmented perturbation function at the origin. That means the augmented perturbation function has a linear support. The augmented Lagrange multiplier is extended in this paper to a new concept called the generalized augmented Lagrangian multiplier, in which a nonlinear support is allowed. The main aim of this paper is to study the existence of generalized augmented Lagrange multipliers. It helps us to better understand properties of an augmented perturbation function at the origin. Based on this nonlinear support, we need to re-investigate the corresponding duality theory, particularly be discussing the relations among generalized augmented Lagrange multipliers, saddle points, and the zero duality gap property. The existence of generalized augmented Lagrange multipliers is established by perturbation analysis of the primal problem.
We organize our paper as follows. Section 2 introduces the preliminaries. In Section 3, we present the duality theory based on generalized augmented Lagrangians. Section 4 discusses the existence of generalized augmented Lagrange multipliers by perturbation analysis.

2. Preliminaries

In this section we clarify the notation, recall some background materials we need from duality theory, and develop some preliminary results.
Recall that
v r ( y , z ) : = v ( y , z ) + r σ ( y , z ) , ( y , z ) R m × R l .
where σ : R m + l R + : = [ 0 , + ) satisfies the following valley-at-zero property:
(i)
σ is continuous at 0 with σ ( 0 , 0 ) = 0 ;
(ii)
inf { σ ( y , z ) | ( y , z ) η , y R m , z R l } > 0 for all η > 0 .
The definition of the growth condition defined below was introduced in [23], as an extension of the one given in [3], where the augmenting function is restricted to be a quadratic function.
Definition 2.
A function v ( y , z ) is said to satisfy the growth condition with σ, if for any τ > 0 , there exist a , c R such that
v ( y , z ) c a σ ( y , z ) , ( y , z ) R m + l \ τ B R m + l ,
where B R m + l denotes the closed unit ball in R m + l .
The dualizing parametrization function of the primal problem is defined as
F ( x , y , z ) : = f ( x ) , if x Ω and y + g ( x ) 0 , z + h ( x ) = 0 , + , otherwise .
For ( x , λ , μ ) R n × R + m × R l , the corresponding generalized augmented Lagrangian is
L ( x , λ , μ , r ) : = inf F ( x , y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r σ ( y , z ) | ( y , z ) R m + l .
The generalized Lagrangian function is defined as
L 0 ( x , λ , μ ) : = f ( x ) ϕ 1 λ , g ( x ) ϕ 2 μ , h ( x ) ,
which reduces to the classical Lagrangian of (P) when ϕ 1 ( λ , y ) = λ , y and ϕ 2 ( μ , z ) = μ , z .
If in particular x Ω , the generalized augmented Lagrangian can be rewritten as
L ( x , λ , μ , r ) = inf y + g ( x ) 0 z + h ( x ) = 0 f ( x ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r σ ( y , z ) inf ξ 1 0 , ξ 2 = 0 { f ( x ) ϕ 1 λ , g ( x ) ϕ 2 μ , h ( x ) ϕ 1 ( λ , ξ 1 ) ϕ 2 ( μ , ξ 2 ) + r σ ξ 1 g ( x ) , ξ 2 h ( x ) }
= inf ξ 1 0 , ξ 2 = 0 L 0 ( x , λ , μ ) ϕ 1 ( λ , ξ 1 ) + r σ ξ 1 g ( x ) , ξ 2 h ( x ) ,
where the inequality comes from ( A 2 ) .
Definition 3.
A solution ( x * , λ * , μ * ) Ω × R + m × R l is said to be a global saddle point of the generalized augmented Lagrangian L for r 0 , if
L ( x * , λ , μ , r ) L ( x * , λ * , μ * , r ) L ( x , λ * , μ * , r ) , x Ω , ( λ , μ ) R + m × R l .
If the above inequalities hold for all x B R n ( x * , δ ) Ω , where B R n ( x * , δ ) denotes the ball with center x * and radius δ > 0 , then ( x * , λ * , μ * ) is said to be a local saddle point of L.
The generalized augmented Lagrangian dual problem of ( P ) is defined as
( D ) sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) ,
where θ ( λ , μ , r ) is the generalized augmented Lagrangian dual function given as
θ ( λ , μ , r ) : = inf { L ( x , λ , μ , r ) | x Ω } .
Taking into account of (7) and (10), we have
inf ( y , z ) R m × R l v r ( y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) = inf ( y , z ) R m × R l inf x Ω f ( x ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r σ ( y , z ) | g ( x ) + y 0 , h ( x ) + z = 0 = inf x Ω inf y + g ( x ) 0 , z + h ( x ) = 0 f ( x ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r σ ( y , z ) = inf x Ω L ( x , λ , μ , r ) = θ ( λ , μ , r ) .
In addition, it also follows from (5) that
v ( y , z ) = inf x Ω g ( x ) + y 0 , h ( x ) + z = 0 f ( x ) = inf x Ω F ( x , y , z ) .
It is well known that a zero duality gap between the problem ( P ) and its generalized augmented Lagrangian dual problem ( D ) holds if
val ( P ) = sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) .
For r 0 , consider the following r-dual problem of ( P ) , denoted by ( D r ) ,
( D r ) sup ( λ , μ ) R + m × R l θ ( λ , μ , r ) = sup ( λ , μ ) R + m × R l inf x Ω L ( x , λ , μ , r ) .
Similarly, if for some fixed r 0 such that
val ( P ) = sup ( λ , μ ) R + m × R l θ ( λ , μ , r ) ,
then the zero duality gap property holds for the pair of problems ( P ) and ( D r ) .
Define the optimal values of problems ( D ) and ( D r ) by val ( D ) and val ( D r ) , respectively. It is clear that val ( D ) = sup r R + val ( D r ) .

3. Duality Theory Based on Generalized Augmented Lagrangian Functions

In this section, we study the relationships among generalized augmented Lagrange multipliers, global saddle points, and the zero duality gap property between the primal problem and its generalized augmented Lagrangian dual problem. The related conclusions are given in Theorem 3 and Theorem 4.
Firstly, the weak duality theorem is given below, which shows that the dual problem provides a lower bound for ( P ) .
Proposition 1.
Let x be a feasible point of ( P ) and ( λ , μ , r ) R + m × R l × R + . Then
θ ( λ , μ , r ) v a l ( P ) f ( x ) .
Proof. 
Since x is feasible, i.e., x Ω and g ( x ) 0 , h ( x ) = 0 , then g ( x ) 0 , h ( x ) = 0 . So
L ( x , λ , μ , r ) = inf y g ( x ) , z = h ( x ) f ( x ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r σ ( y , z ) f ( x ) ,
where the inequality follows by letting y = 0 , z = 0 and ϕ ( · , 0 ) = 0 . Hence
θ ( λ , μ , r ) = inf x Ω L ( x , λ , μ , r ) f ( x ) .
The arbitrariness of x ensures
θ ( λ , μ , r ) inf x Ω g ( x ) 0 , h ( x ) = 0 f ( x ) = v a l ( P ) .
Theorem 1.
Let σ : R m + l R + and ( λ * , μ * , r * ) R + m × R l × R + . Then ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) with r * if and only if ( λ * , μ * , r * ) is an optimal solution of ( D ) and the zero duality gap property holds for problems ( P ) and ( D ) .
Proof. 
(Necessity). If ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) with r * , then
v r * ( 0 , 0 ) = inf ( y , z ) R m × R l v r * ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z ) ,
where the above equation is due to Definition 1. According to (11), we have
v a l ( P ) = v r * ( 0 , 0 ) = θ ( λ * , μ * , r * ) .
This implies
v a l ( P ) = θ ( λ * , μ * , r * ) sup ( λ , μ ) R + m × R l θ ( λ , μ , r * ) v a l ( D ) = sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) v a l ( P ) ,
where the third inequality is due to (13). Hence, ( λ * , μ * , r * ) is an optimal solution of ( D ) and
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) = v a l ( P ) .
(Sufficiency). Suppose ( λ * , μ * , r * ) is an optimal solution of ( D ) and the zero duality gap property between ( P ) and ( D ) holds. Then
v a l ( P ) = θ ( λ * , μ * , r * ) v a l ( D r * ) v a l ( D ) v a l ( P ) .
Hence
θ ( λ * , μ * , r * ) = v a l ( D r * ) = v a l ( P ) = v ( 0 , 0 ) = v r * ( 0 , 0 ) ,
which together with (11) implies
v r * ( 0 , 0 ) = θ ( λ * , μ * , r * ) = inf ( y , z ) R m × R l v r * ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z ) .
Therefore, ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) with r * . □
From the proof of Theorem 1, we can see that ( λ * , μ * ) is an optimal solution of ( D r * ) and the zero duality gap property holds between ( P ) and ( D r * ) . It should be emphasized that the existence of generalized augmented Lagrange multipliers does not require that the primal problem ( P ) must be solvable. Indeed, in general, the optimal solution of a primal problem cannot be known in advance. The relation between the zero duality gap property and global saddle points is given below.
Theorem 2.
Let σ : R m + l R + and r * 0 . Then ( x * , λ * , μ * ) is a global saddle point of L ( x , λ , μ , r * ) if and only if val ( P ) = val ( D r * ) , and x * Ω , ( λ * , μ * ) R + m × R l are optimal solutions of ( P ) and ( D r * ) , respectively.
Proof. 
We first claim that
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) = f ( x ) , x Ω and g ( x ) 0 , h ( x ) = 0 , + , otherwise .
Consider the following two cases:
Case 1.x is infeasible. Then either x Ω or x Ω while x F . If x Ω , from (5) and (6) we get
L ( x , λ , μ , r * ) = + , ( λ , μ ) R + m × R l .
If x Ω , but x F , it follows from the property ( A 3 ) that there exist nonzero ( λ 0 , μ 0 ) R + m × R l and γ < 0 such that
ϕ 1 ( η λ 0 , y ) + ϕ 2 ( η μ 0 , z ) ω ( η ) γ ,
whenever ( y , z ) satisfies y + g ( x ) 0 , z + h ( x ) = 0 , and η > 0 sufficiently large. Hence
L ( x , η λ 0 , η μ , r * ) = inf y + g ( x ) 0 , z + h ( x ) = 0 f ( x ) ϕ 1 ( η λ 0 , y ) ϕ 2 ( η μ , z ) + r * σ ( y , z ) inf y + g ( x ) 0 , z + h ( x ) = 0 f ( x ) ϕ 1 ( η λ 0 , y ) ϕ 2 ( η μ 0 , z ) f ( x ) ω ( η ) γ ,
where the first inequality comes from the nonnegativity of σ , and the second inequality is due to (16). This together with γ < 0 further implies that
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) L ( x , η λ 0 , η μ 0 , r * ) f ( x ) ω ( η ) γ + , a s η + ;
i.e.,
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) = + .
Therefore, either x Ω or x Ω , x F , so it follows from (15) and (17) that
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) = + .
Case 2.x is feasible i.e., x Ω and g ( x ) 0 , h ( x ) = 0 . In this case, it follows from (12) that for any ( λ , μ ) R + m × R l ,
L ( x , λ , μ , r * ) = inf y g ( x ) , z = h ( x ) { f ( x ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) + r * σ ( y , z ) } f ( x ) .
According to the nonnegativity of σ , we also have
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) L ( x , 0 , 0 , r * ) f ( x ) ,
which together with (19) means that
sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) = f ( x ) .
Putting (18) and (20) together yields the desired formula (14). Hence
v a l ( P ) = inf x Ω sup ( λ , μ ) R + m × R l L ( x , λ , μ , r * ) .
On the other hand, note that the dual problem can be rewritten as
v a l ( P ) = sup ( λ , μ ) R + m × R l inf x Ω L ( x , λ , μ , r * ) .
The desired result follows by applying the minimax relations theorem (Theorem 11.50 [34]). □
Indeed, Theorem 2 shows that val ( P ) = val ( D ) , and x * , ( λ * , μ * , r * ) are optimal solutions of ( P ) and ( D ) respectively, provided that val ( D ) = val ( D r ) , i.e., val ( D ) = sup r R + val ( D r ) by Proposition 1, and the maximum can be attained at some r. The converse statement obviously holds true. As just mentioned above, compared with the existence of augmented Lagrange multipliers, global saddle points require that the primal problem is solvable.
Theorem 3.
Suppose that σ : R m + l R + has a valley at zero, v satisfies the growth condition with σ, and
lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) < + .
The following statements hold:
(i)
sup ( 0 , 0 , r ) R + m × R l × R + θ ( 0 , 0 , r ) = lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) ;
(ii)
v is lower semi-continuous at the origin if and only if the zero duality gap property holds for problems ( P ) and ( D ) .
Proof .
(i). First, according to the condition (21) we show that
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) = lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) .
Assume that { ( y ( s ) , z ( s ) ) } is the sequence such that the liminf in (21) is attained; i.e.,
lim s ( y ( s ) , z ( s ) ) = 0 , lim s v ( ( y ( s ) , z ( s ) ) ) = lim inf s 0 v ( y , z ) .
Consider the following two cases:
Case 1. lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) = . For ( λ , μ , r ) R + m × R l × R + , it follows from (11) that
θ ( λ , μ , r ) = inf ( y , z ) R m + l v r ( y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) v ( y ( s ) , z ( s ) ) + r σ ( y ( s ) , z ( s ) ) ϕ 1 ( λ , y ( s ) ) ϕ 2 ( μ , z ( s ) ) ,
where the inequality comes from (1). Passing to limit (24), together with (23), we get
θ ( λ , μ , r ) lim s v ( y ( s ) , z ( s ) ) + r σ ( y ( s ) , z ( s ) ) ϕ 1 ( λ , y ( s ) ) ϕ 2 ( μ , z ( s ) ) = lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) = ,
where the first equality comes from the continuity of σ and ϕ by ( A 1 ) . Hence
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) = lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) .
Case 2. lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) > . Noting that θ ( λ , μ , r ) lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) , then
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) .
Conversely, take k satisfying lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) > k . Then there exists τ > 0 such that
v ( y , z ) + r σ ( y , z ) v ( y , z ) k , ( y , z ) τ B R m + l , r 0 ,
where the first inequality follows from the nonnegativity of σ . Since σ has a valley at zero, there exists ε > 0 such that
σ ( y , z ) ε , ( y , z ) R m + l \ τ B R m + l .
Using the growth condition of v with σ , for the above τ > 0 there exist a , c R such that
v ( y , z ) c a σ ( y , z ) , τ > 0 , ( y , z ) R m + l \ τ B R m + l .
This together with (27) yields
v ( y , z ) + r σ ( y , z ) c a σ ( y , z ) + r σ ( y , z ) c + ( r a ) ε k , ( y , z ) R m + l \ τ B R m + l , r r 0 ,
where r 0 : = max { a + ( k c ) / ε , 0 } + 1 . From (26) and (28) we get
θ ( 0 , 0 , r ) = inf ( y , z ) R m + l v ( y , z ) + r σ ( y , z ) k , r r 0 ,
and
sup r R + θ ( 0 , 0 , r ) k .
Since k < lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) is arbitrary, then
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) .
Taking into account (25), we get in the last inequality
sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) = lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) .
The statement (i) follows by letting ( λ , μ ) = ( 0 , 0 ) in (24) and (25) and by a similar argumentation as above.
(ii). If v is lower semi-continuous at origin, then
lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) v ( 0 , 0 ) = v a l ( P ) ,
which together with (22) yields
v a l ( P ) lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) = sup ( λ , μ , r ) R + m × R l × R + θ ( λ , μ , r ) v a l ( P ) .
Therefore, the zero duality gap property holds for ( P ) and ( D ) .
Conversely, according to (22), it is easy to see that the lower semi-continuity of v at the origin can be obtained if the zero duality gap property holds for problems (P) and (D). □
Corollary 1.
Suppose that σ : R m + l R + has a valley at zero, v satisfies the growth condition with σ, and
lim inf ( y , z ) ( 0 , 0 ) v ( y , z ) < + .
If v is lower semi-continuous at origin and r * arg sup r > 0 θ ( 0 , 0 , r ) , then the following statements hold:
(i)
( 0 , 0 ) is a generalized augmented Lagrange multiplier;
(ii)
If the primal problem (P) has the optimal solution x * , then ( x * , 0 , 0 ) and ( x * , 0 , 0 , r * ) are saddle points of D r * and D, respectively.
Proof. 
The results follow immediately from Theorem 3. □
Theorem 3 shows that the zero duality gap property is closely related with the lower semi-continuity of the perturbation function. In the definition of generalized augmented Lagrange multipliers, the inequality involved in (4) is required to be satisfied for all ( y , z ) R m + l , but Theorem 4 shows that this restriction can be weakened by just checking all ( y , z ) in some neighborhood of the origin once some additional assumptions are imposed on augmented functions. In the following, we further require the ϕ satisfying the following property:
( A 4 ) For any ( λ , μ ) R + m × R l , there exist ρ > 0 , τ > 0 such that
ρ σ ( y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) 0 , ( y , z ) R m + l \ τ B R m + l .
Theorem 4.
Suppose that σ has a valley at zero and v satisfies the growth condition with σ. Then (P) has a generalized augmented Lagrange multiplier ( λ * , μ * ) R + m × R l if and only if there exists r * R + such that
v r * ( y , z ) v r * ( 0 , 0 ) + ϕ 1 ( λ * , y ) + ϕ 2 ( μ * , z ) , ( y , z ) τ B R m + l .
Proof. 
(Necessity). The necessity is clear by the definition of generalized augmented Lagrange multiplier.
(Sufficiency). Since v satisfies the growth condition with σ , then for any τ > 0 there exist a , c R such that
v ( y , z ) c a σ ( y , z ) , ( y , z ) R m + l \ τ B R m + l .
Since σ has a valley at zero, there exists d > 0 such that
σ ( y , z ) d , ( y , z ) R m + l \ τ B R m + l .
Combining the property ( A 4 ) with (31) means that for any ( y , z ) R m + l \ τ B R m + l we have
v r ( y , z ) v r ( 0 , 0 ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z )     c a σ ( y , z ) v r ( 0 , 0 ) + r σ ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z )     = c val ( P ) + ( r a ρ ) σ ( y , z ) + ρ σ ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z )     c val ( P ) + ( r a ρ ) σ ( y , z )     c val ( P ) + ( r a ρ ) d .
Pick r > max r * , a + ρ , a + ρ c val ( P ) / d . Then
v r ( y , z ) v r ( 0 , 0 ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z ) 0 , ( y , z ) R m + l \ τ B R m + l .
It follows from (29) and (32) that
v r ( y , z ) v r ( 0 , 0 ) + ϕ 1 ( λ * , y ) + ϕ 2 ( μ * , z ) , ( y , z ) R m + l .
Hence ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) . □
Here we list two classes of nonlinear functions satisfying the above assumptions ( A 1 ) ( A 4 ) .
(1) Let θ : R R be sublinear, continuous, and increasing with θ ( 0 ) = 0 .
Let
ϕ ( x , y ) : = x θ ( x T y ) .
(1-1) ϕ ( x , 0 ) = x θ ( 0 ) = 0 , x .
(1-2) For any x , y , z ,
ϕ ( x , y + z ) = x θ ( x T ( y + z ) ) = x θ ( x T y + x T z ) x θ ( x T y ) + x θ ( x T z ) = ϕ ( x , y ) + ϕ ( x , z ) .
(1-3) For any x F , then ( g ( x ) , h ( x ) ) R m × { 0 } l , i.e., ( 0 , 0 ) R m × { 0 } l ( g ( x ) , h ( x ) ) . If 0 R m g ( x ) , then according to convex set sperate theorem, there exist a nonzero vector u 0 and ξ < 0 such that u 0 T y < ξ whenever y + g ( x ) 0 . Hence taking v 0 : = 0 and γ : = u 0 θ ( ξ ) , we have
ϕ 1 ( η u 0 , y ) + ϕ 2 ( η v 0 , z ) = ϕ 1 ( η u 0 , y ) = η u 0 θ ( η u 0 T y ) η u 0 θ ( η ξ ) η u 0 θ ( ξ ) = ω ( η ) γ ,
where ω ( η ) : = η . Similarly, if 0 { 0 } l h ( x ) , there exists a nonzero vector v 0 and ξ < 0 such that v 0 T z ξ for z = h ( x ) . Hence taking u 0 : = 0 and γ : = v 0 θ ( ξ ) , we have
ϕ 1 ( η u 0 , y ) + ϕ 2 ( η v 0 , z ) = ϕ 2 ( η v 0 , z ) = η v 0 θ ( η v 0 T z ) η v 0 θ ( ξ ) = ω ( η ) γ .
(1-4) Let β : R + R satisfy β ( t ) > 0 as t > 0 . Assume that there exist α > 0 , τ > 0 such that for all t > α and u with u τ , we have θ ( t u ) β ( t ) σ ( u ) .
For any ( λ , μ ) , letting ϱ : = λ + μ + α we have
ϕ 1 ( λ , y ) + ϕ 2 ( μ , z ) = λ θ ( λ T y ) + μ θ ( μ T z ) λ θ ( λ y ) + μ θ ( μ z ) ϱ θ ( ϱ y ) + ϱ θ ( ϱ z ) 2 ϱ θ ( ϱ ( y , z ) ) .
As ( y , z ) τ , we have 2 ϱ θ ( ϱ ( y , z ) ) 2 ϱ β ( ϱ ) σ ( y , z ) . Hence for all ( y , z ) τ and ρ : = 2 ϱ β ( ϱ ) ,
ρ σ ( y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) 0 .
In particular, we can take θ as a piecewise linear function or a support function over a bounded closed interval, w ( t ) : = t , β ( t ) : = t 2 , and σ ( u ) : = u .
(2) Let θ : R + R satisfy θ ( t ) > 0 if t 0 and θ ( t ) t q as t > 0 are sufficiently large, where q is positive integer.
Define
ϕ ( x , y ) : = θ ( x ) ( x T A y ) ,
where A is a symmetric and invertible matrix.
(2-1) ϕ ( x , 0 ) = 0 , x .
(2-2) For any x , y , z ,
ϕ ( x , y + z ) = θ ( x ) ( x T A ( y + z ) ) = θ ( x ) ( x T A y + x T A z ) = ϕ ( x , y ) + ϕ ( x , z ) .
(2-3) Similar to the argument given in (1-3), if 0 R m g ( x ) , then according to convex set sperate theorem, there exists a nonzero vector u ˜ 0 and ξ < 0 such that u ˜ 0 T y < ξ whenever y + g ( x ) 0 . Hence taking u 0 : = A 1 u ˜ 0 , v 0 : = 0 , and γ : = θ ( u 0 ) ξ , we have
ϕ 1 ( η u 0 , y ) + ϕ 2 ( η v 0 , z ) = ϕ 1 ( η u 0 , y ) = θ ( η u 0 ) η u 0 T A y η q + 1 u 0 q ξ η q θ ( u 0 ) ξ = ω ( η ) γ ,
whenever η θ ( u 0 ) / u 0 q and ω ( η ) : = η q .
If 0 { 0 } l h ( x ) , there exists a nonzero vector v ˜ 0 and ξ < 0 such that v ˜ 0 T z ξ for z = h ( x ) . Hence taking u 0 : = 0 , v 0 : = A 1 v ˜ 0 , and γ : = θ ( v 0 ) ξ , we have
ϕ 1 ( η u 0 , y ) + ϕ 2 ( η v 0 , z ) = ϕ 2 ( η v 0 , z ) = θ ( η v 0 ) η v 0 T A z η q + 1 v 0 q ξ η q θ ( v 0 ) ξ ω ( η ) γ ,
whenever η θ ( v 0 ) / v 0 q .
(2-4) Assume that there exists τ > 0 such that σ ( u ) u as u with u τ . For any ( λ , μ ) , we have
ϕ 1 ( λ , y ) + ϕ 2 ( μ , z ) = θ ( λ ) λ T A y + θ ( μ ) μ T A z θ ( λ ) ( A λ y ) + θ ( μ ) ( A μ z ) θ ( λ ) A λ + θ ( μ ) A μ ( y , z ) θ ( λ ) A λ + θ ( μ ) A μ σ ( y , z ) .
Let ρ : = θ ( λ ) A λ + θ ( μ ) A μ . Then for any ( y , z ) with ( y , z ) τ ,
ρ σ ( y , z ) ϕ 1 ( λ , y ) ϕ 2 ( μ , z ) 0 .
In particular, we can take q : = 2 , ω ( t ) : = t 2 , θ ( t ) : = t 3 , and σ ( u ) : = u 2 .

4. Existence of Generalized Augmented Lagrange Multipliers

In this section, we develop some sufficient conditions for the existence of generalized augmented Lagrange multipliers. Given ε 0 , define
W 1 ( ε ) : = x Ω | dist g ( x ) , h ( x ) ; R m × { 0 } l ε ,
and
W 2 ( ε ) : = x Ω | f ( x ) v ( 0 , 0 ) ε .
Lemma 1.
Suppose that σ : R m + l R + has a valley at zero and
inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) > .
Then for any ε > 0 , we have
lim r + inf x Ω / W 1 ( ε ) L ( x , λ * , μ * , r ) = + ,
and
x Ω | L ( x , λ * , μ * , r ) v ( 0 , 0 ) W 1 ( ε ) W 2 ( ε ) ,
whenever r > 0 is sufficiently large.
Proof. 
The proofs of (34) and (35) are given in parts (a) and (b), respectively.
(a) For any fixed x Ω / W 1 ( ε ) , it follows from the definition of W 1 ( ε ) that
dist g ( x ) , h ( x ) ; R m × { 0 } l > ε ,
which implies that for any ( ξ 1 , ξ 2 ) with ξ 1 0 , ξ 2 = 0 we have
inf x Ω / W 1 ( ε ) ( ξ 1 , ξ 2 ) g ( x ) , h ( x ) inf x Ω / W 1 ( ε ) dist g ( x ) , h ( x ) ; R m × { 0 } l ε .
According to the valley-at-zero property of σ , for any x Ω / W 1 ( ε ) , there exists ζ > 0 such that
σ ξ 1 g ( x ) , ξ 2 h ( x ) ζ .
It follows from (8) that
L ( x , λ * , μ * , r ) inf ξ 1 0 , ξ 2 = 0 L 0 ( x , λ * , μ * ) ϕ 1 ( λ * , ξ 1 ) + r σ ξ 1 g ( x ) , ξ 2 h ( x ) inf ξ 1 0 { L 0 ( x , λ * , μ * ) ϕ 1 ( λ * , ξ 1 ) + r ζ } = L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) + r ζ ,
where the second inequality comes from (36). This implies that
inf x Ω / W 1 ( ε ) L ( x , λ * , μ * , r ) inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) + r ζ .
Passing to limit in the above inequality, we get
lim r + inf x Ω / W 1 ( ε ) L ( x , λ * , μ * , r ) inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) + lim r + r ζ = + ,
where the equality comes from the fact that inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) is finite by (33). Hence, (34) is true.
(b) First prove that
{ x Ω | L ( x , λ * , μ * , r ) v ( 0 , 0 ) } W 1 ( ε ) .
We argue it by contradiction. If there exist ε 0 > 0 , r k , and x k Ω such that
L ( x k , λ * , μ * , r k ) v ( 0 , 0 ) , x k W 1 ( ε 0 ) ,
then
L ( x k , λ * , μ * , r k ) inf x Ω / W 1 ( ε 0 ) L ( x , λ * , μ * , r k ) .
Passing to limit in the above inequality, we get
lim inf k L ( x k , λ * , μ * , r k ) lim k inf x Ω / W 1 ( ε 0 ) L ( x , λ * , μ * , r k ) = + ,
where the equality comes from part ( a ) . Clearly, this contradicts the finiteness of v ( 0 , 0 ) .
Next, we claim that
{ x Ω | L ( x , λ * , μ * , r ) v ( 0 , 0 ) } W 2 ( ε ) .
Suppose, on the contrary, that there exist ε 0 > 0 , r k and x k Ω such that
L ( x k , λ * , μ * , r k ) v ( 0 , 0 ) , x k W 2 ( ε 0 ) .
From (7) and (37), we conclude that there exist y k + g ( x k ) 0 , z k + h ( x k ) = 0 such that
v ( 0 , 0 ) + ε 0 2 L ( x k , λ * , μ * , r k ) + ε 0 2 f ( x k ) ϕ 1 ( λ * , y k ) ϕ 2 ( μ * , z k ) + r k σ ( y k , z k ) .
By the property ( A 4 ) , for above ( λ * , μ * ) , there exist ρ > 0 , τ > 0 such that
ρ σ ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z ) 0 , ( y , z ) R m + l \ τ B R m + l .
Further using valley-at-zero property of σ , for above τ > 0 there exists d 1 > 0 such that
σ ( y , z ) d 1 , ( y , z ) R m + l \ τ B R m + l .
Now, let us prove that ( y k , z k ) 0 . Let us consider the following cases:
Case 1. There exists an infinite subset N 1 N such that ( y k , z k ) τ for all k N 1 . Note that
v ( 0 , 0 ) + ε 0 f ( x k ) ϕ 1 ( λ * , y k ) ϕ 2 ( μ * , z k ) + ρ σ ( y k , z k ) + ( r k ρ ) σ ( y k , z k ) f ( x k ) + ( r k ρ ) σ ( y k , z k ) f ( x k ) + ( r k ρ ) d 1 ,
where the second inequality comes from the assumption (39), and the third step is due to (40). The right side in (41) can be arbitrary large as N 1 k , which contradicts the finiteness of v ( 0 , 0 ) .
Case 2. ( y k , z k ) τ as k sufficiently large. Since ϕ 1 and ϕ 2 are continuous by the property A 1 , for above ( λ * , μ * ) and τ > 0 , we can find d 2 R such that
ϕ 1 ( λ * , y ) + ϕ 2 ( μ * , z ) d 2 , ( y , z ) R m + l \ τ B R m + l .
Then
v ( 0 , 0 ) + ε 0 f ( x k ) ϕ 1 ( λ * , y k ) ϕ 2 ( μ * , z k ) + r k σ ( y k , z k ) f ( x k ) d 2 + r k σ ( y k , z k ) d 2 + r k σ ( y k , z k ) .
Due to the boundedness of ( y k , z k ) , we have
σ ( y k , z k ) v ( 0 , 0 ) + ε 0 + d 2 r k 0 , a s k ,
which in turn implies that ( y k , z k ) 0 by the valley-at-zero property of σ .
Applying ( y k , z k ) 0 into (38) yields v ( 0 , 0 ) + ε 0 2 f ( x k ) . It contradicts x k W 2 ( ε 0 ) by the definition of W 2 ( ε ) . Therefore, (35) holds. □
Remark 1.
Note that ϕ 2 ( μ * , ξ 2 ) is not used in the assumption (33). The reason is ϕ 2 ( μ * , ξ 2 ) = 0 as ξ 2 = 0 , since the perturbation for equality constraint is restricted to the subspace { 0 } l .
Theorem 5.
Suppose that σ : R m + l R + has a valley at zero and
inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) > .
For any x * X * , ( x * , λ * , μ * ) is a local saddle point of L ( x , λ * , μ * , r ) for some r * > 0 and there exist a bounded subset Λ R n and ε 0 > 0 such that
x Ω | dist g ( x ) , h ( x ) ; R m × { 0 } l ε 0 , f ( x ) v ( 0 , 0 ) ε 0 Λ .
Then ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) .
Proof. 
According to the relationship among the generalized augmented Lagrange multiplier, the zero duality gap property, and global saddle points established in Theorems 1 and 2, we only need to justify that ( x * , λ * , μ * ) is a global saddle point of L ( x , λ , μ , r ) .
According to the definition of local saddle points, there exists δ > 0 such that
L ( x * , λ , μ , r * ) L ( x * , λ * , μ * , r * ) L ( x , λ * , μ * , r * ) , x B R n ( x * , δ ) Ω , ( λ , μ ) R + m × R l .
It follows by invoking (14) and the first inequality in (43) that
L ( x * , λ * , μ * , r * ) = f ( x * ) .
By the monotonicity of L ( x * , λ * , μ * , r ) in r, we also have
L ( x * , λ * , μ * , r * ) L ( x * , λ * , μ * , r ) f ( x * ) , r r * ,
where the second inequality comes from (12). Combining (44) and (45) implies
L ( x * , λ * , μ * , r ) = f ( x * ) , r r * ,
which together with (12) again yields
L ( x * , λ * , μ * , r ) = f ( x * ) L ( x * , λ , μ , r ) , r r * , ( λ , μ ) R + m × R l .
Now, we establish the first inequality in (9). To complete the proof, it remains to show that
L ( x * , λ * , μ * , r ) L ( x , λ * , μ * , r ) , x Ω \ B R n ( x * , δ ) ,
whenever r is sufficiently large. Suppose on the contrary that we can find r k + and x k Ω \ B R n ( x * , δ ) such that
L ( x k , λ * , μ * , r k ) < L ( x * , λ * , μ * , r k ) .
Hence, applying (46) into (49) and together with the fact x * X * yields
L ( x k , λ * , μ * , r k ) < f ( x * ) = v ( 0 , 0 ) ,
which means that x k belongs to the set { x Ω | L ( x , λ * , μ * , r k ) v ( 0 , 0 ) } . Taking into account of (35) in Lemma 1, we obtain that for any ε ( 0 , ε 0 ) , x k W 1 ( ε ) W 2 ( ε ) , which further implies that x k Λ by (42). We can assume without loss of generality that x k converges to x ¯ . According to the continuity of f ( x ) , g ( x ) and h ( x ) , together with the closedness of W 1 ( ε ) and W 2 ( ε ) , we obtain that x ¯ W 1 ( ε ) W 2 ( ε ) . Therefore, x ¯ W 1 ( 0 ) W 2 ( 0 ) by the arbitrariness of ε > 0 , which further implies that x ¯ X * . By assumption, ( x ¯ , λ * , μ * ) is also a local saddle point of L ( x , λ , μ , r ) for some r ¯ > 0 ; i.e., there exists δ ¯ > 0 such that
L ( x ¯ , λ , μ , r ¯ ) L ( x ¯ , λ * , μ * , r ¯ ) L ( x , λ * , μ * , r ¯ ) , x B R n ( x ¯ , δ ¯ ) Ω , ( λ , μ ) R + m × R l .
Similar to the above argument, it follows from (44) that
L ( x ¯ , λ * , μ * , r ¯ ) = f ( x ¯ ) = v ( 0 , 0 ) = v a l ( P ) .
Since x k B R n ( x ¯ , δ ¯ ) and r k r ¯ for k large enough, from (51) and (52), it follows
L ( x k , λ * , μ * , r k ) L ( x k , λ * , μ * , r ¯ ) L ( x ¯ , λ * , μ * , r ¯ ) = f ( x ¯ ) = f ( x * ) = v ( 0 , 0 ) ,
which contradicts (50). This justifies (48).
By the fact (43), (47) and (48), we conclude that ( x * , λ * , μ * ) is a global saddle point of L ( x , λ , μ , r ) for r large enough. Therefore, ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) . □
Theorem 6.
Suppose that σ : R m + l R + has a valley at zero and
inf x Ω L 0 ( x , λ * , μ * ) sup ξ 1 0 ϕ 1 ( λ * , ξ 1 ) > .
Let x * be the unique global optimal solution of ( P ) . If ( x * , λ * , μ * ) is a local saddle point of L ( x , λ , μ , r ) for some r 0 , and there exists ε 0 > 0 such that
{ x Ω | dist ( g ( x ) , h ( x ) ; R m × { 0 } l ) ε 0 } Λ ,
where Λ is a bounded subset in R n , then ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) .
Proof. 
For ε 0 > 0 , it follows from (34) in Lemma 1 that there exists r 1 > 0 such that
inf x Ω / W 1 ( ε 0 ) L ( x , λ * , μ * , r 1 ) f ( x * ) .
That is to say, for any x Ω \ W 1 ( ε 0 ) , we have
L ( x , λ * , μ * , r 1 ) f ( x * ) , x Ω \ W 1 ( ε 0 ) .
To complete the proof, we next need to show that there exists r 2 > 0 such that
L ( x , λ * , μ * , r 2 ) f ( x * ) , x W 1 ( ε 0 ) .
Suppose on the contrary that there exist r k and { x k } W 1 ( ε 0 ) such that
f ( x * ) > L ( x k , λ * , μ * , r k ) .
According to (53) and Λ being bounded, W 1 ( ε 0 ) is bounded, which further implies that { x k } has at least a cluster point x ¯ . We assume without loss of generality that x k converges x ¯ .
We now claim that x ¯ is a feasible point of ( P ) . If x ¯ is not feasible, then dist g ( x ) , h ( x ) ; R m × { 0 } l > 2 m 0 for some m 0 > 0 . Therefore, dist g ( x k ) , h ( x k ) ; R m × { 0 } l > m 0 as k sufficiently large. It in turn implies
L ( x k , λ * , μ * , r k ) inf x Ω / W 1 ( m 0 ) L ( x , λ * , μ * , r k ) .
Taking the limits on both sides yields
lim inf k L ( x k , λ * , μ * , r k ) lim k inf x Ω / W 1 ( m 0 ) L ( x , λ * , μ * , r k ) = + ,
where the equality comes from Lemma 1. Combining (56) with (57) together yields a contradiction to the finiteness of f ( x * ) . This justifies the feasibility of x ¯ for ( P ) .
By hypothesis, ( x * , λ * , μ * ) is a local saddle point of L ( x , λ , μ , r ) for some r 0 ; then there exists a neighborhood B R n ( x * , δ ) such that
L ( x * , λ , μ , r ) L ( x * , λ * , μ * , r ) L ( x , λ * , μ * , r ) , x B R n ( x * , δ ) Ω , ( λ , μ ) R + m × R l .
Putting (14), (58), and the monotonicity of L ( x * , λ * , μ * , r ) with respect to r together means that for any r r ,
f ( x * ) = sup ( λ , μ ) R + m × R l L ( x * , λ , μ , r ) = L ( x * , λ * , μ * , r ) L ( x * , λ * , μ * , r ) sup ( λ , μ ) R + m × R l L ( x * , λ , μ , r ) = f ( x * ) .
That is,
f ( x * ) = L ( x * , λ * , μ * , r ) , r r .
Taking into account of (56), we obtain that
L ( x * , λ * , μ * , r ) = f ( x * ) > L ( x k , λ * , μ * , r k ) L ( x k , λ * , μ * , r ) ,
where the last step is due to the monotonicity of L ( x , λ * , μ * , r ) with respect to r. Thus, it follows from (58) that x k B R n ( x * , δ ) whenever k is sufficiently large; i.e., x ¯ x * . Since x * is the unique global optimal solution of (P) and invoking that x ¯ is feasible, we have
f ( x ¯ ) f ( x * ) > 0 .
Define ϱ : = f ( x ¯ ) f ( x * ) / 2 > 0 . Using (7) and (56), there exist y k g ( x k ) , z k = h ( x k ) such that
f ( x * ) + ϱ > L ( x k , λ * , μ * , r k ) + ϱ f ( x k ) ϕ 1 ( λ * , y k ) ϕ 2 ( μ * , z k ) + r k σ ( y k , z k ) .
Similarly to the argument given in Lemma 1, we conclude from the property ( A 4 ) that ( y k , z k ) 0 . Hence f ( x * ) + ϱ f ( x k ) . Passing to limit, we get f ( x * ) + ϱ f ( x ¯ ) , which together with the definition of ϱ implies that f ( x * ) f ( x ¯ ) . It is clearly the case that x ¯ is also an optimal solution. Hence x ¯ = x * , since the optimal solution is unique. This justifies (55).
Let r * : = max { r 1 , r 2 } . Taking into account of (54) and (55), we obtain that
L ( x , λ * , μ * , r * ) f ( x * ) = L ( x * , λ * , μ * , r * ) , x Ω ,
where the equation comes from (59). Hence, according to (11) and (60),
inf ( y , z ) R m + l v r * ( y , z ) ϕ 1 ( λ * , y ) ϕ 2 ( μ * , z ) = inf x Ω L ( x , λ * , μ * , r * ) = f ( x * ) = v ( 0 , 0 ) ,
where the last step comes from that x * is the optimal solution of (P). This further implies
v r * ( y , z ) v r * ( 0 , 0 ) + ϕ 1 ( λ * , y ) + ϕ 2 ( μ * , z ) , ( y , z ) R m + l .
Hence ( λ * , μ * ) is a generalized augmented Lagrange multiplier of ( P ) . □
The existence of generalized augmented Lagrange multipliers is established in two different scenarios: one is applicable to the case of unique solution while another is applicable to the case of multiple optimal solutions.

5. Conclusions

In this paper, we studied the generalized augmented Lagrangian multiplier, which is an extension of the augmented Lagrangian multiplier from linear support to nonlinear support for an augmented perturbation function. Some sufficient conditions for the existence of generalized augmented Lagrangian multipliers were developed. In particular, the relationships among global saddle points, generalized augmented Lagrangian multipliers, and the zero duality gap property between the primal problem and its generalized augmented Lagrangian dual problem were established. Several interesting topics are left for further investigation. For example, one is developing some necessary and sufficient conditions for the existence of generalized augmented Lagrangian multipliers by using the localization principle; another is studying the generalized differentiation of support functions from the subdifferential view.

Author Contributions

All authors contributed equally and significantly in writing this article. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by National Natural Science Foundation of China (11771255, 11801325) and Young Innovation Teams of Shandong Province (2019KJI013).

Acknowledgments

The authors are gratefully indebted to the anonymous referees for their valuable suggestions that helped us greatly improve the original presentation of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Birgin, E.G.; Martinez, J.M. Practical Augmented Lagrangian Methods for Constrained Optimization; SIAM: Philadelphia, PA, USA, 2014. [Google Scholar]
  2. Curtis, F.E.; Jiang, H.; Robinson, D.P. An adaptive augmented Lagrangian method for large-scale constrained optimization. Math. Program. 2015, 152, 201–245. [Google Scholar] [CrossRef]
  3. Rockafellar, R.T. Augmented Lagrange multiplier functions and duality in nonconvex programming. SIAM J. Control Optim. 1974, 12, 268–285. [Google Scholar] [CrossRef]
  4. Kiwiel, K.C. On the twice differentiable cubic augmented Lagrangian. J. Optim. Theory Appl. 1996, 88, 233–236. [Google Scholar] [CrossRef]
  5. Mangasarian, O.L. Unconstrained Lagrangians in nonlinear programming. SIAM J. Control Optim. 1975, 12, 772–791. [Google Scholar] [CrossRef] [Green Version]
  6. Wu, H.X.; Luo, H.Z. Saddle points of general augmented Lagrangians for constrained nonconvex optimization. J. Glob. Optim. 2012, 53, 683–697. [Google Scholar] [CrossRef]
  7. Tseng, P.; Bertsekas, D.P. On the convergence of the exponential multiplier method for convex programming. Math. Program. 1993, 60, 1–19. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, C.Y.; Li, D. Unified theory of augmented Lagrangian methods for constrained global optimization. J. Glob. Optim. 2009, 44, 433–458. [Google Scholar] [CrossRef]
  9. Polyak, R.A. Log-sigmoid multipliers method in constrained optimization. Ann. Oper. Res. 2001, 101, 427–460. [Google Scholar] [CrossRef]
  10. Polyak, R.A. Modified barrier functions: Theory and methods. Math. Program. 1992, 54, 177–222. [Google Scholar] [CrossRef]
  11. Wu, H.X.; Luo, H.Z. A note on the existence of saddle points of p-th power Lagrangian for constrained nonconvex optimization. Optimization 2012, 61, 1231–1245. [Google Scholar] [CrossRef]
  12. Wang, C.Y.; Yang, X.Q.; Yang, X.M. Unified nonlinear Lagrangian approach to duality and optimal paths. J. Optim. Theory Appl. 2007, 135, 85–100. [Google Scholar] [CrossRef]
  13. Burachik, R.S.; Iusem, A.N.; Melo, J.G. Duality and exact penalization for general augmented Lagrangians. J. Optim. Theory Appl. 2010, 147, 125–140. [Google Scholar] [CrossRef]
  14. Wang, C.Y.; Yang, X.Q.; Yang, X.M. Nonlinear augmented Lagrangian and duality theory. Math. Oper. Res. 2012, 38, 740–760. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, C.; Liu, Q.; Qu, B. Global saddle points of nonlinear augmented Lagrangian functions. J. Glob. Optim. 2017, 68, 125–146. [Google Scholar] [CrossRef]
  16. Zhang, L.W.; Gu, J.; Xiao, X.T. A class of nonlinear Lagrangians for nonconvex second-order cone programming. Comput. Optim. Appl. 2011, 49, 61–99. [Google Scholar] [CrossRef]
  17. Zhou, J.C.; Chen, J.S. On the existence of saddle points for nonlinear second-order cone programming problems. J. Glob. Optim. 2015, 62, 459–480. [Google Scholar] [CrossRef]
  18. Fukuda, E.H.; Lourenco, B.F. Exact augmented Lagrangian functions for nonlinear semidefinite programming. Comput. Optim. Appl. 2018, 71, 457–482. [Google Scholar] [CrossRef] [Green Version]
  19. Sun, D.F.; Sun, J.; Zhang, L.W. The rate of convergence of the augmented Lagrangian method for nonlinear semidefinite programming. Math. Program. 2008, 114, 349–391. [Google Scholar] [CrossRef]
  20. Zhao, X.Y.; Sun, D.F.; Toh, K.-C. A Newton-CG augmented Lagrangian method for semidefinite programming. SIAM J. Optim. 2010, 20, 1737–1765. [Google Scholar] [CrossRef]
  21. Dolgopolik, M.V. Augmented Lagrangian functions for cone constrained optimization: the existence of global saddle points and exact penalty property. J. Glob. Optim. 2018, 71, 237–296. [Google Scholar] [CrossRef] [Green Version]
  22. Shapiro, A.; Sun, J. Some properties of the augmented Lagrangian in cone constrained optimization. Math. Oper. Res. 2004, 29, 479–491. [Google Scholar] [CrossRef]
  23. Zhou, Y.Y.; Zhou, J.C.; Yang, X.Q. Existence of augmented Lagrange multipliers for cone constrained optimization problems. J. Glob. Optim. 2014, 58, 243–260. [Google Scholar] [CrossRef]
  24. Burachik, R.S.; Yang, X.Q.; Zhou, Y.Y. Existence of augmented Lagrange multipliers for semi-infinite programming problems. J. Optim. Theory Appl. 2017, 173, 471–503. [Google Scholar] [CrossRef]
  25. Ru¨ckmann, J.-J.; Shapiro, A. Augmented Lagrangians in semi-infinite programming. Math. Program. 2009, 116, 499–512. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, C.Y.; Zhou, J.C.; Xu, X.H. Saddle points theory of two classes of augmented Lagrangians and its applications to generalized semi-infinite programming. Appl. Math. Optim. 2009, 59, 413–434. [Google Scholar] [CrossRef]
  27. Chatzipanagiotis, N.; Dentcheva, D.; Zavlanos, M.M. An augmented Lagrangian method for distributed optimization. Math. Program. 2015, 152, 405–434. [Google Scholar] [CrossRef]
  28. Feizollahi, M.J.; Ahmed, S.; Sun, A. Exact augmented Lagrangian duality for mixed integer linear programming. Math. Program. 2017, 161, 365–387. [Google Scholar] [CrossRef]
  29. Boland, N.; Christiansen, J.; Dandurand, B.; Eberhard, A.; Oliveira, F. A parallelizable augmented Lagrangian method applied to large-scale non-convex-constrained optimization problems. Math. Program. 2019, 175, 503–536. [Google Scholar] [CrossRef] [Green Version]
  30. Kanzow, C.; Steck, D. Augmented Lagrangian methods for the solution of generalized Nash equilibrium Problems. SIAM J. Optim. 2016, 26, 2034–2058. [Google Scholar] [CrossRef] [Green Version]
  31. Kanzow, C.; Steck, D. Quasi-variational inequalities in Banach spaces: theory and augmented Lagrangian methods. SIAM J. Optim. 2019, 29, 3174–3200. [Google Scholar] [CrossRef]
  32. Liu, Y.F.; Liu, X.; Ma, S.Q. On the nonergodic convergence rate of an inexact augmented Lagrangian framework for composite convex programming. Math. Oper. Res. 2019, 44, 632–650. [Google Scholar] [CrossRef] [Green Version]
  33. Teng, Y.; Yang, L.; Song, X.L.; Yu, B. An augmented Lagrangian proximal alternating method for sparse discrete optimization problems. Numer. Algorithms 2020, 83, 833–866. [Google Scholar] [CrossRef]
  34. Rockafellar, R.T.; Wets, J.-B. Variational Analysis; Springer: New York, NY, USA, 1998. [Google Scholar]
  35. Huang, X.X.; Yang, X.Q. A unified augmented Lagrangian approach to duality and exact penalization. Math. Oper. Res. 2003, 28, 533–552. [Google Scholar] [CrossRef] [Green Version]
  36. Rubinov, A.M.; Huang, X.X.; Yang, X.Q. The zero duality gap property and lower semicontinuity of the perturbation function. Math. Oper. Res. 2002, 27, 775–791. [Google Scholar] [CrossRef] [Green Version]
  37. Burachik, R.S.; Rubinov, A. Abstract convexity and augmented Lagrangians. SIAM J. Optim. 2007, 18, 413–436. [Google Scholar] [CrossRef]
  38. Kan, C.; Song, W. Second-order conditions for existence of augmented Lagrange multipliers for eigenvalue composite optimization problems. J. Glob. Optim. 2015, 63, 77–97. [Google Scholar] [CrossRef]
  39. Bonnans, J.F.; Shapiro, A. Perturbation Analysis of Optimization Problems; Springer: New York, NY, USA, 2000. [Google Scholar]
  40. Ramana, M.; Tuncel, L.; Wolkowicz, H. Strong duality for semidefinite programming. SIAM J. Optim. 1997, 7, 641–662. [Google Scholar] [CrossRef] [Green Version]
  41. Borwein, J.M.; Wolkowicz, H. Characterization of optimality for the abstract convex program with finite-dimensional range. J. Aust. Math. Soc. 1981, 30, 390–411. [Google Scholar] [CrossRef] [Green Version]
  42. Kostyukova, O.I.; Tchemisova, T.V. Optimality conditions for convex semi-infinite programming problems with finitely representable compact index sets. J. Optim. Theory Appl. 2017, 175, 76–103. [Google Scholar] [CrossRef] [Green Version]
  43. Kostyukova, O.I.; Tchemisova, T.V. Optimality criteria without constraint qualification for linear semidefinite problems. J. Math. Sci. 2012, 182, 126–143. [Google Scholar] [CrossRef] [Green Version]
  44. Dolgopolik, M.V. Existence of augmented Lagrange multipliers: Reduction to exact penalty functions and localization principle. Math. Program. 2017, 166, 297–326. [Google Scholar] [CrossRef] [Green Version]

Share and Cite

MDPI and ACS Style

Wang, Y.; Zhou, J.; Tang, J. Existence of Generalized Augmented Lagrange Multipliers for Constrained Optimization Problems. Math. Comput. Appl. 2020, 25, 24. https://doi.org/10.3390/mca25020024

AMA Style

Wang Y, Zhou J, Tang J. Existence of Generalized Augmented Lagrange Multipliers for Constrained Optimization Problems. Mathematical and Computational Applications. 2020; 25(2):24. https://doi.org/10.3390/mca25020024

Chicago/Turabian Style

Wang, Yue, Jinchuan Zhou, and Jingyong Tang. 2020. "Existence of Generalized Augmented Lagrange Multipliers for Constrained Optimization Problems" Mathematical and Computational Applications 25, no. 2: 24. https://doi.org/10.3390/mca25020024

Article Metrics

Back to TopTop