Next Article in Journal
Bivariate Burr X Generator of Distributions: Properties and Estimation Methods with Applications to Complete and Type-II Censored Samples
Next Article in Special Issue
Regarding New Wave Patterns of the Newly Extended Nonlinear (2+1)-Dimensional Boussinesq Equation with Fourth Order
Previous Article in Journal
Fuzzy Reduced Hypergroups
Previous Article in Special Issue
On C-To-R-Based Iteration Methods for a Class of Complex Symmetric Weakly Nonlinear Equations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modified Accelerated Bundle-Level Methods and Their Application in Two-Stage Stochastic Programming

College of Mathematics and Information Science, Guangxi University, Nanning 540004, China
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(2), 265; https://doi.org/10.3390/math8020265
Submission received: 25 December 2019 / Revised: 31 January 2020 / Accepted: 11 February 2020 / Published: 17 February 2020
(This article belongs to the Special Issue Computational Methods in Applied Analysis and Mathematical Modeling)

Abstract

:
The accelerated prox-level (APL) and uniform smoothing level (USL) methods recently proposed by Lan (Math Program, 149: 1–45, 2015) can achieve uniformly optimal complexity when solving black-box convex programming (CP) and structure non-smooth CP problems. In this paper, we propose two modified accelerated bundle-level type methods, namely, the modified APL (MAPL) and modified USL (MUSL) methods. Compared with the original APL and USL methods, the MAPL and MUSL methods reduce the number of subproblems by one in each iteration, thereby improving the efficiency of the algorithms. Conclusions of optimal iteration complexity of the proposed algorithms are established. Furthermore, the modified methods are applied to the two-stage stochastic programming, and numerical experiments are implemented to illustrate the advantages of our methods in terms of efficiency and accuracy.

1. Introduction

In the fields of production planning, finance risk, telecommunication, and electricity, decision makers need to take into consideration uncertainty about the information and the model itself. Lack of data, calculation errors as well as unpredictability, etc. lead to the uncertainty in information. The uncertainty of model is derived from the structure of problem, the features of constraints, as well as the risks and profiles of decisions. Stochastic programming (SP) is an effective tool for dealing with optimization problems under uncertainty. The expectation model in stochastic programming is widely used, which maximizes the expectation of the benefit or minimizes the loss under the expected constraints. In this paper, we are concerned with the two-stage stochastic programming with recourse. Next, through a practical problem, network planning with random demand, we introduce the mathematical model of two-stage stochastic programming.
Due to demand for higher bandwidth and dedicated lines, network capacity is becoming a scarce resource. Consider a situation where a network provider plans bandwidth allocation between network links, under total network capacity b that is available for allocation. In addition, there are n different links needed to be expanded in capacity. The extra capacity allocated to link j is x j , where j = 1 , 2 , , n and vector x R n consists of elements x j . In network planning, demands refer to the number of connections requested between point-to-point pairs provided by the network at a certain time. Here, the demand D R m related to the m point-to-point pairs is modeled as a random variable.
Suppose that the extra capacities x j , j = 1 , 2 , , n are given and the demand D is observed. Then the capacity planning model introduced in [1] is to minimize the total number of unoffered requests. Let i = 1 , , m be the indices of point-to-point pairs, and the paths that can be offered to connections related to point-to-point pair i be represented in P ( i ) . c R n is a vector whose elements c j denote the capacity of current link j. f i p and s i denote the number of connections and unoffered requests related to pair i, respectively. For an observation d = ( d 1 , , d m ) of the random variable D, one can obtain the optimal decision by solving the linear programming problem as follows
min i = 1 m s i s . t . i = 1 m p P ( i ) A i p f i p x + c , p P ( i ) f i p + s i = d i , f i p 0 , s i 0 , i = 1 , 2 , , m , and p P ( i ) .
Here, 0-1 vector A i p is introduced as an incidence whose j-th component, a i p j , is 1 if link j lies in path p and 0 otherwise. Let Q ( x , d ) denote the optimal value of the above linear programming (1) which depends on the observation d of D and capacity x. It is obvious that Q ( x , D ) is also a random variable due to the randomness of D. Thus, the capacities x j , j = 1 , 2 , , n can be obtained from the following programming problem
min E [ Q ( x , D ) ] s . t . j = 1 n x j b , x 0 .
Here E [ · ] refers to the expectation of the probability distribution of random variable D. In a word, the purpose of programming problem (2) is to minimize the expectation of unoffered requests, while satisfying the constraint within the total network capacity b. The above optimization model (1) and (2) is an instance of two-stage stochastic programming with recourse, in which (1) is the second-stage problem and (2) is the first-stage problem.
From the above practical instance we can see that the stochastic programming plays a key role in network planning with random demand. In addition to stochastic programming, the unit commitment problems in the field of energy systems usually involve mixed-integer quadratic programming. Recently, there have been some studies on solving such problems [2,3,4]. For solving the programming problems (1) to (2), Sen et al. [1] applied the stochastic decomposition method, which is one of the most efficient methods for stochastic programming. There are still many applications of stochastic programming in other fields, such as insurance and finance, valuation of electricity, telecommunications, hydrothermal power production planning and pollution control, etc. Stochastic models of these application problems can be found in [5]. In what follows, as an extension of the network planning problem, the standard mathematical model of two-stage stochastic programming with recourse is given, which is more convenient for generalization and theoretical analysis.
The two-stage stochastic programming with fixed recourse is in the following form
min c T x + E [ Q ( x , D ) ] s . t . A x = b , x 0 ,
where x is the decision variable, c is the the cost of production and Q ( x , D ) denotes the optimal objective value of the second-stage problem
min q T y s . t . W y = h T x , y 0 .
Here, D is the demand vector with the elements q T , h T , and T. Let W be the fixed recourse and D = ( D 1 , , D n ) be a random variable with the support of probability distribution Ξ R d . E [ Q ( x , D ) ] in the first-stage problem (3) is the expectation with respect to D.
The mathematical model of the two-stage stochastic programming is given and next we will summarize the methods mainly used to solve it. Since the two and multi-stage stochastic programming problems have very large dimension and special structures, they can be solved by means of decomposition. In two-stage models (3) and (4), suppose that the probability space Θ is finite, ω s , s = 1 , , S and that p s , s = 1 , , S represent all basic events and their probabilities. Then the first-stage problem can be rewritten in the form
min c T x + s = 1 S p s Q s ( x ) s . t . A x = b , x 0 .
Here, Q s ( x ) is the optimal value of the second-stage problem
min ( q s ) T y s . t . W s y = h s T s x , y 0 ,
where, T s = T ( ω s ) , W s = W ( ω s ) , h s = h ( ω s ) , y s = y ( ω s ) , q s = q ( ω s ) .
The main idea of the basic dual decomposition methods is to establish certain approximations to the first-stage problem by means of solving subproblems in structure (6). As an original decomposition-type method, the cutting-plane method establishes a linear model of s = 1 S p s Q s ( x ) , and its main scheme can be found in [6,7]. Simplicity and cheapness in calculation are advantages of the cutting-plane model. With the increasing of cuts, however, the demand for storage growths which is a typical difficulty of such method. On the other hand, although there is a good initial iteration, such method may generate a longer stepsize. In order to tackle these difficulties, Ruszczynski [8] proposed the regularized decomposition method which is considered as an improvement of the basic one. Another way to avoid generating longer steps is the trust region method which was extended to two-stage stochastic programming by Linderoth and Wright [9]. In addition, bundle methods, which can be viewed as stabilized variants of the original cutting-plane methods, were developed in [10,11,12]. Algorithmic modifications based on bundle methods were introduced in [13,14]. The bundle level (BL) method was first proposed by Lemaréchal et al. [15] as another kind of bundle method. On this basis, in [16,17,18] the “restricted-memory” version of BL method is developed which performs well in numerical experiments. In recent years, there has been tremendous development in “asynchronous” and “partial” versions of BL methods, see references [19,20,21,22,23]. Considering that the research on BL methods is only focused on general non-smooth convex optimization (CP) problems, Lan [24] proposed an accelerated BL-type method, namely the accelerated bundle level (ABL) method and its restricted memory version, the accelerated prox-level (APL) method. Benefiting from the multi-step strategy introduced by Nesterov [25] and later applied in [26,27,28,29,30], both ABL and APL methods are uniformly optimal for solving non-smooth, weakly smooth and smooth CP problems. In addition, by incorporating Nesterov’s smoothing technique [16,17] into APL method, Lan [24] presented the uniform smoothing level (USL) method for solving structure non-smooth CP problems with optimal iteration complexity. In particular, the USL method does not require any input of problem parameters. Moreover, Lan [24] illustrated that the APL and USL methods, when applied to solving semidefine programming and two-stage stochastic programming, have obvious advantages in computation time and accuracy over the related gradient-type algorithms and some existing methods.
Our main work in this paper includes several aspects. First of all, on the basis of Lan’s work, we make further improvement and present the modified accelerated prox-level (MAPL) method. By selecting the proper proximal function, the MAPL method only need to solve one subproblem to update the prox-center and the lower bound synchronously, which improves the computational efficiency. In addition, MAPL method can achieve uniformly optimal complexity for solving smooth, weakly smooth, and non-smooth CP problems uniformly. Furthermore, we extend the MAPL method to solve the structure non-smooth CP problems and present the modified uniform smoothing level (MUSL) method. Finally, we apply the proposed methods to solve the two-stage stochastic programming problems with recourse. The numerical results show that the MAPL and MUSL methods have certain advantages in iterations and computation time.
The present paper is built up as follows. In Section 2, some related works about BL-type methods are reviewed. The MAPL method and its complexity analysis are presented in Section 3. The MUSL method and its complexity analysis are presented in Section 4. The application of MAPL and MUSL methods to two-stage stochastic programming as well as numerical experiments are shown in Section 5. Conclusions are presented in the final section.

2. Related Work

This section reviews some related work on BL methods. Specifically, the main ideas of BL methods and related notation are reviewed in Section 2.1. The APL method and its gap reduction procedure G M A P L introduced in Section 2.2 are the basis of the main work in this paper.

2.1. The Bundle Level (BL) Method

Consider the general CP problem
f * : = min x X f ( x ) ,
where the constraint set X R n is convex and compact, and the objective function f : X R is closed and convex over X. The function f is known only through a first-order oracle returning the function value f ( x ) and a subgradient f ( x ) f ( x ) for a given point x X , where f ( x ) : = { ξ R n | f ( y ) f ( x ) + ξ , y x , y R n } is the subdifferential of f at x.
Given a sequence of iteration points, the first-order information { f ( x k ) } and { f ( x k ) } are provided by oracle. The cutting-plane approximation of f is generated by
f ˇ k ( x ) : = max { h ( x i , x ) : 1 i k } f ( x ) ,
where h ( x i , x ) : = f ( x i ) + f ( x i ) , x x i . In addition, x , y : = i x i y i , x , y R n refers to the inner product. The level set X k of f ˇ k with level parameter k is defined by X k : = { x X : f ˇ k ( x ) k } . The BL method [15] generates the next iteration point x k + 1 by
x k + 1 = argmin x X k x x k 2 .
The main steps of the BL method are listed below.
Step 1. Set f ¯ k : = min f x i , 1 i k and compute a lower bound f ̲ k : = min x X f ˇ k ( x ) ;
Step 2. Set the level parameter k = β f ̲ k + ( 1 β ) f ¯ k for some β ( 0 , 1 ) ;
Step 3. Set X k : = { x X : f ˇ k ( x ) k } and generate the new iterate by solving (8).

2.2. The Accelerated Prox-Level (APL) Method and Its Gap Reduction Procedure G M A P L

In this subsection we consider the CP problem (7), where f satisfies the following inequality
f ( y ) f ( x ) f ( x ) , y x M 1 + ρ y x 1 + ρ , x , y X ,
for some M > 0 , ρ [ 0 , 1 ] and f ( x ) f ( x ) . It is easy to show that non-smooth ( ρ = 0 ) , smooth ( ρ = 1 ) and weakly smooth ( ρ ( 0 , 1 ) ) CP problems are contained in this family of problems. Lan [24] generalized the BL method to accelerated bundle level (ABL) and accelerated prox-level (APL) methods, such that they achieve uniformly optimal complexity bounds for functions satisfying (9). Here, we mainly introduce the APL method and its principle.
Firstly, Lan [24] introduced three related iteration sequences, { x k l } , { x k u } and { x k } to establish the cutting-plane approximation f ˇ k ( x ) , generate the upper bounds f ¯ k , and update prox-center x k , respectively. Specifically, { x k l } and { x k u } are updated by x k l = ( 1 α k ) x k u + α k x k 1 and x k u = ( 1 α k ) x k 1 u + α k x k . Secondly, Lan introduced an internal procedure G A P L to reduce the gap between the upper and lower bounds of f * . The algorithmic framework of the APL method is as follows (Algorithm 1).
Algorithm 1: The APL method.
  Input: Choose p 0 X , stopping tolerance ε > 0 and parameters λ , θ ( 0 , 1 ) .
  Step 0: (Initialization) Set p 1 Argmin x X h p 0 , x , lb 1 = h p 0 , p 1 and ub 1 = f p 1 .
  Let s = 1 .
  Step 1: (Stopping test) If ub s lb s ε , terminate.
  Step 2: (Call G A P L procedure) Set p s + 1 , lb s + 1 = G A P L p s , lb s , λ , θ and ub s + 1 = f p s + 1 .
  Step 3: (Loop) Set s = s + 1 , and return to Step 1.
Next, we focus on describing the gap reduction procedure G A P L . Denote the level set X f ( ) : = { x X : f ( x ) } . For a given iterate z, denote
f ¯ : = min { h ( z , x ) : x X f ( ) } .
Then, it can be verified that
min { , f ¯ } f ( x ) , x X .
This implies that min { , f ¯ } a lower bound of f * . However, the problem (10) is difficult to solve in general. To overcome the difficulty and obtain the lower bound in a convenient way, Lan used a compact and convex set X to replace X f ( ) in problem (10). Thus, one can solve the relaxation of (10)
f ̲ : = min { h ( z , x ) : x X } .
Here, the set X is called the localizer of the level set X f ( ) satisfying X f ( ) X X . Then we obtain a lower bound on f * as follow
min { , f ̲ } f ( x ) , x X .
Indeed, as shown in [24], if X , then f ̲ min { h ( z , x ) : x X f ( ) } f ( x ) , for all x X . If X = , we have that f ̲ = + and X f ( ) = , and therefore f ( x ) for all x X . Thus, (11) holds.
Moreover, in order to make better use of the structure of the feasible set X, similar to the NERML algorithm in [16,17], Lan introduced the prox-function to replace the Euclidean distance function · 2 . Here, a function ω : X R is served as a prox-function of a convex compact set X R n with coefficient σ ω , if it is a function with differentiability, strong convexity as well as coefficient σ ω i.e.
ω ( x ) ω ( z ) , x z σ ω x z 2 , x , z X .
Furthermore, one can redefine the diameter of X with d ω by
D ω , X 2 : = max { ω ( x ) ω ( z ) ω ( z ) , x z , x , z X } .
This lead to the following relation
x z 2 ( 2 / σ ω ) D ω , X 2 = : Ω ω , X , x , z X .
Furthermore, let
d ω ( x ) : = ω ( x ) [ ω ( x 0 ) + ω ( x 0 ) , x x 0 ] , x X ,
be the prox-function on X and x 0 : = arg min x X d ω ( x ) be the prox-center of d ω ( x ) . It follows that
d ω ( x ) ( 1 / 2 ) σ ω x x 0 2 ,
and d ω ( x ) is a strong convex function with coefficient σ ω .
The internal gap reduction procedure G A P L of APL method is as follows.
The APL gap reduction procedure: ( p + , lb + ) = G A P L ( p , lb , λ , θ )
  • Step 0: (Initialization) Set x 0 u = p , f ¯ 0 = f ( x 0 u ) , f ̲ 0 = lb and = λ f ̲ 0 + ( 1 λ ) f ¯ 0 . In addition, choose x 0 X and the initial localizer X 0 = X . The prox-function d ω ( x ) is defined in (15). Also let k = 1 .
  • Step 1: (Update lower bound) Let x k l = α k x k 1 + ( 1 α k ) x k 1 u , and h ( x k l , x ) = f ( x k l ) + f ( x k l ) , x x k l .
    h ̲ k : = min x X k 1 h ( x k l , x ) , and f ̲ k : = max f ̲ k 1 , min , h ̲ k .
    If f ̲ k θ ( f ̲ 0 ) , then stop the procedure and output p + = x k 1 u and lb + = f ̲ k .
  • Step 2: (Update prox-center) Set
    x k : = argmin x X k 1 d ω ( x ) : h ( x k l , x ) .
  • Step 3: (Update upper bound) Set f ¯ k = min { f ¯ k 1 , f ( α k x k + ( 1 α k ) x k 1 u ) } and x k u such that f ( x k u ) = f ¯ k . If f ¯ k + θ ( f ¯ 0 ) , then stop the procedure and output p + = x k 1 u and lb + = f ̲ k .
  • Step 4: (Update localizer) Choose X k such that X ̲ k X k X ¯ k , where
    X ̲ k : = x X k 1 : h ( x k l , x ) and X ¯ k = x X : d ω ( x k ) , x x k 0 .
  • Step 5: (Loop) Set k = k + 1 and return to Step 1.

3. The Modified Accelerated Prox-Level (MAPL) Method

In this section, we proposed the modified accelerated prox-level (MAPL) method which requires only one subproblem to be solved per iteration while achieving the uniformly optimal iteration complexity for solving the black-box CP problems. We first present the modified gap reduction procedure G M A P L , which generates a new search point p + and a new lower bound l b + of f * such that f ( p + ) l b + q [ f ( p ) l b ] from the given input p and lb and some q ( 0 , 1 ) . Here, the value of q depends on the parameters λ , θ ( 0 , 1 ) .
The MAPL gap reduction procedure: ( p + , lb + ) = G M A P L ( p , lb , λ , θ )
  • Step 0: (Initialization) Set x 0 u = p , f ¯ 0 = f ( x 0 u ) , f ̲ 0 = lb and = λ f ̲ 0 + ( 1 λ ) f ¯ 0 . In addition, choose x 0 X and the initial localizer X 0 = X . The prox-function d ω ( x ) is defined in (15). Let k = 1 .
  • Step 1: (Update level set) Set
    x k l = α k x k 1 + ( 1 α k ) x k 1 u ,
    h ( x k l , x ) = f ( x k l ) + f ( x k l ) , x x k l ,
    X ̲ k = x X k 1 : h ( x k l , x ) .
  • Step 2: (Update prox-center and lower bound) Set
    x k = argmin x X ̲ k d ω ( x ) .
    If X ̲ k = , then stop the procedure and output p + = x k 1 u , l b + = .
  • Step 3: (Update upper bound) Set
    x ˜ k u = α k x k + ( 1 α ) x k 1 u ,
    x k u = x ˜ k u , i f f ( x ˜ k u ) < f ¯ k 1 , x k 1 u , o t h e r w i s e ,
    and f ¯ k = f ( x k u ) . If f ¯ k + θ ( f ¯ 0 ) , then stop the procedure (there is a significant improvement on the upper bound) and output p + = x k u , l b + = l b .
  • Step 4: (Update localizer) Choose an arbitrary X k such that X ̲ k X k X ¯ k , where
    X ¯ k = x X : d ω ( x k ) , x x k 0 .
  • Step 5: (Loop) Set k = k + 1 and return to Step 1.
The following are a few remarks on the MAPL gap reduction procedure. Firstly, the upper bound f ¯ 0 = f ( x 0 u ) and lower bound f ̲ 0 = lb of f * in Step 0 are obtained from the outer iteration of the MAPL method (described below), and f ¯ 0 and f ̲ 0 are fixed throughout the entire progress of G M A P L . Furthermore, the level parameter is also fixed as a convex combination of f ¯ 0 and f ̲ 0 . But in the original BL methods, the level parameter changes in each iteration. Secondly, Step 2 and Step 3 have two exits. When the procedure stops at Step 2, is the lower bound on f * ; when the procedure stops at Step 3, there is a significant progress on f ¯ k , which depends on the parameter θ . Compared with the gap reduction procedure G A P L of APL method, the G M A P L updates the lower bound in a easier way. Indeed, in Step 1 of G A P L see [24], in order to determine whether the lower bound lb needs to be updated, a linear programming subproblem is first solved. However, in the G M A P L procedure, the prox-center x k and the lower bound lb are merged into one subproblem to be updated by selecting appropriate prox-function d ω . While solving the subproblem (20), it can be automatically checked that whether X ̲ k = . If X ̲ k = , the lower bound lb is directly updated to , which avoids solving the linear programming as G A P L and reduce the number of calculation. Thirdly, in Step 4, X k can be selected arbitrarily under condition X ̲ k X k X ¯ k . For convenience, one can directly choose X k = X ̲ k or X k = X ¯ k . However, the number of constraints in X ̲ k increases with k, and X ¯ k has only one more constraint than the feasible set X. In practice, we can choose X k between these two sets, which can control the number of constraints in (19) and reduce the computation cost. Moreover, a proper selection of the stepsize sequences { α k } is critical for G M A P L to terminate after finite iterations and achieve optimal iteration complexity. Lan proposed a generalized selection rule for the sequence { α k } in [24]. Chen et al. [31] proposed a more concise selection rule, i.e., the sequence { α k } satisfies the following conditions:
α 1 = 1 , 0 < α k 1 , α k c k , 1 α k + 1 α k + 1 1 + γ 1 α k 1 + γ , k 1 ,
for any γ [ 0 , 1 ] and some c > 0 . Two examples for { α k } are as follows [31]:
(1) α k = 2 / ( k + 1 ) , k = 1 , 2 , , with c = 2 ;
(2) α 1 = 1 , α k + 1 2 = ( 1 α k + 1 ) α k 2 , with c = 2 .
The following lemma shows the important properties of procedure G M A P L . These properties are similar to those of [24,31], whose proof can be found in Appendix A.
Lemma 1.
The following properties hold for the procedure G M A P L .
a. X k k 0 is a collection of localizers for the level set X f ( ) ;
b. f ¯ 0 f ¯ 1 f ¯ k f * holds for any k 1 ;
c. X f ( ) , and hence X f ( ) X ̲ k X k X ¯ k for any k 1 ;
d. When X ̲ k holds, the problem (20) has a unique solution. In addition, if the procedure G M A P L stops at Step 2, we have f * .
e. When the procedure G M A P L stops, the relation f ( p + ) lb + q ( f ( p ) lb ) holds, where
q q ( λ , θ ) = max λ , 1 λ ( 1 θ ) .
Referring to the theoretical analysis mode of [24], the following proposition will show that the gap between the upper bound f ( x k u ) and the level parameter will decrease with k, and it is proven that when the algorithm stops, the total number of iterations does not exceed an upper bound.
Proposition 1.
If { α k } satisfy the condition (24), take γ = ρ , let ℓ be the level parameter and d ω ( x ) denotes the prox-function, then the number of iterations of the procedure G M A P L does not exceed
N Δ 0 : = c 1 + ρ M Ω ω , X 1 + ρ 2 λ θ ( 1 + ρ ) Δ 0 2 1 + 3 ρ ,
where Δ 0 = f ¯ 0 f ̲ 0 and Ω ω , X is defined in (14).
Proof. 
Assume that the gap reduction procedure G M A P L does not stop at the K-th iteration. From (16) we have for x 1 X
σ ω 2 x 1 x 0 2 d ω x 1 .
By the definition of X ¯ k in (23) we have d ω x k , z x k 0 , for any z X ¯ k . This together with X k X ¯ k shows that
d ω x k , z x k 0 , z X k .
It follows from (19) and (20) that x k + 1 X ̲ k + 1 X k , for all 1 k K . Moreover, due to the strong convexity of d ω ( x ) and the optimality condition of subproblem (20), we have
σ ω 2 x k + 1 x k 2 d ω x k + 1 d ω x k d w x k , x k + 1 x k d ω x k + 1 d ω x k , k 1 .
It turns out that
σ ω 2 x k x k 1 2 d ω x k + 1 d ω x k , k 1 .
Summing up the above inequalities over k, we obtain that
σ ω 2 k = 1 K x k + 1 x k 2 d ω x K .
According to (22), (9) and (18), we have
f ( x k u ) f ( x ˜ k u ) f ( x k l ) + f ( x k l ) , x ˜ k u x k l + M 1 + ρ x ˜ k u x k l 1 + ρ = h ( x k l , x ˜ k u ) + M 1 + ρ x ˜ k u x k l 1 + ρ .
From (18), (21), the convexity of f, (19) and (22), we have
h ( x k l , x ˜ k u ) = f ( x k l ) + f ( x k l ) , ( 1 α k ) x k 1 u + α k x k x k l = ( 1 α k ) [ f ( x k l ) + f ( x k l ) , x k 1 u x k l ] + α k [ f ( x k l ) + f ( x k l ) , x k x k l ] = ( 1 α k ) h ( x k l , x k 1 u ) + α k h ( x k l , x k ) ( 1 α k ) f ( x k 1 u ) + α k h ( x k l , x k ) ( 1 α k ) f ( x k 1 u ) + α k .
By the definition of x ˜ k u (21) and x k l (17), it is easy to show that
x ˜ k u x k l = α k x k + ( 1 α k ) x k 1 u α k x k 1 ( 1 α k ) x k 1 u = α k x k α k x k 1 = α k ( x k x k 1 ) .
It then follows that
f ( x k u ) ( 1 α k ) f ( x k 1 u ) + α k + M α k 1 + ρ 1 + ρ x k x k 1 1 + ρ .
Subtracting both sides of the above inequality by and dividing both sides by α k 1 + ρ , from (24), we obtain that
1 α k 1 + ρ [ f ( x k u ) ] 1 α k α k 1 + ρ [ f ( x k 1 u ) ] + M 1 + ρ x k x k 1 1 + ρ 1 α k 1 1 + ρ f x k 1 u + M 1 + ρ x k x k 1 1 + ρ , k 2 .
As a result,
1 α 1 1 + ρ f x 1 u M 1 + ρ x 1 x 0 1 + ρ ,
1 α 2 1 + ρ f x 2 u 1 α 1 1 + ρ f x 1 u + M 1 + ρ x 2 x 1 1 + ρ ,
1 α K 1 + ρ f x K u 1 α K 1 1 + ρ f x K 1 u + M 1 + ρ x K x K 1 1 + ρ .
Summing up the above inequality over m and from the fact that α k > 0 , k 1 , we have
f x K u α K 1 + ρ M 1 + ρ k = 1 K x k x k 1 1 + ρ .
Applying the H o ¨ lder inequality to the above inequality and using (27), we have
k = 1 K x k x k 1 1 + ρ K 1 ρ 2 k = 1 K x k x k 1 2 1 + ρ 2 K 1 ρ 2 2 d ω x K σ ω 1 + ρ 2 K 1 ρ 2 2 D ω , X σ ω 1 + ρ 2 .
Applying (24) and (14) to (28), as well as the above inequality, we have
f x K u c 1 + ρ M K 1 + ρ ( 1 + ρ ) K 1 ρ 2 2 D ω , X σ ω 1 + ρ 2 c 1 + ρ M 1 + ρ Ω ω , X 1 + ρ 2 K 1 + ρ 2 .
From the fact that G M A P L does not stop at Step 3, we have
f ¯ k = f ( x K u ) > θ ( f ¯ 0 ) = θ λ ( f ¯ 0 f ̲ 0 ) = θ λ ( ub lb ) = λ θ Δ 0 .
This together with (29) shows that
λ θ Δ 0 M c 1 + ρ 1 + ρ Ω ω , X 1 + ρ 2 K 1 + 3 ρ 2 .
We finally conclude that
K < c 1 + ρ M Ω ω , X 1 + ρ 2 λ θ ( 1 + ρ ) Δ 0 2 1 + 3 ρ ,
that is
N Δ 0 = c 1 + ρ M Ω ω , X 1 + ρ 2 λ θ ( 1 + ρ ) Δ 0 2 1 + 3 ρ .
 □
After giving the relevant properties and complexity analysis of the procedure G M A P L , we are now in a position to introduce the so-called modified accelerated prox-level (MAPL) method, which repeatedly calls the procedure G M A P L during iteration progress until it finds an approximate solution with the given accuracy. The algorithmic framework of MAPL method is as follows (Algorithm 2).
Algorithm 2: The MAPL method.
  Parameters: Choose stopping tolerance ε > 0 and parameters λ , θ ( 0 , 1 ) .
  Step 0: (Initialization) Choose initial point p 0 , set p 1 arg min x X h p 0 , x , lb 1 = h p 0 , p 1 and ub 1 = min { f ( p 0 ) , f ( p 1 ) ) } . Let s = 1 ;
  Step 1: (Stopping test) If ub s lb s ε , terminate;
  Step 2: (Call G M A P L procedure) Set p s + 1 , lb s + 1 = G M A P L p s , lb s , λ , θ and ub s + 1 = f p s + 1 ;
  Step 3: (Loop) Set s = s + 1 , and return to Step 1.
Since the procedure G M A P L is called during the progress of the MAPL method, we consider that an iteration of G M A P L is also an iteration of the MAPL method. Take this fact into consideration, the following theorem establishes the convergence and iteration complexity of the MAPL method. The principle of the proof comes from reference [24].
Theorem 1.
For given ε > 0 , if α k ( 0 , 1 ] , k = 1 , 2 , , in G M A P L are chosen to satisfy condition (24) with some c > 0 , then
(1) The number of procedure G M A P L called by the MAPL method can be bounded by
S ( ε ) : = max 0 , log 1 q M Ω ω , X 1 + ρ 2 ( 1 + ρ ) ε ;
(2) The total number of iterations performed by the MAPL method does not exceed
S ( ε ) + 1 1 q 2 1 + 3 ρ c 1 + ρ M Ω ω , X 1 + ρ 2 ( 1 + ρ ) λ θ ε 2 1 + 3 ρ .
Proof. 
(1) Without loss of generality, we suppose that Δ 1 > ε , where Δ s = ub s lb s , s 1 . According to Step 1 of the MAPL method, (9) and (14), we have
Δ 1 = f ( p 1 ) h ( p 0 , p 1 ) = f ( p 1 ) [ f ( p 0 ) + f ( p 0 ) , p 1 p 0 ] M p 1 p 0 1 + ρ 1 + ρ M Ω ω , X 1 + ρ 2 1 + ρ .
From Lemma 1 and the fact that u b s = f ( p s ) , we have
Δ s + 1 q Δ s , s 1 .
Furthermore, suppose that the MAPL method finds an ε -solution after calling S ( ε ) times G M A P L , i.e., the MAPL method stops at s : = S ( ε ) + 1 . Then we have
Δ S ( ε ) + 1 ε Δ S ( ε ) ,
and
ε < q S ( ε ) 1 Δ 1 q S ( ε ) 1 M Ω ω , X 1 + ρ 2 .
It is easy to obtain
( 1 q ) S ( ε ) 1 M Ω ω , X 1 + ρ 2 ( 1 + ρ ) ε ,
that is
S ( ε ) : = max 0 , log 1 q M Ω ω , X 1 + ρ 2 ( 1 + ρ ) ε .
(2) Suppose that the procedure G M A P L has been called S ( ε ) times in MAPL method. Then by (30) and (31) we have
Δ s > ε q s S ( ε ) , s = 1 , 2 , , S ( ε ) .
Due to Δ S ( ε ) > ε , we know that
s = 1 S ( ε ) Δ s 2 1 + 3 ρ < s = 1 S ( ε ) q 2 1 + 3 ρ ( s S ( ε ) ) ε 2 1 + 3 ρ = t = 0 S ( ε ) 1 q 2 1 + 3 ρ t ε 2 1 + 3 ρ 1 ( 1 q 2 1 + 3 ρ ) ε 2 1 + 3 ρ .
This together with Lemma 1 and Proposition 1 shows that the total number of iterations performed by the MAPL method does not exceed
s = 1 S ( ε ) N s s = 1 S ( ε ) N ( Δ s ) S ( ε ) + s = 1 S ( ε ) c 1 + ρ M Ω ω , X 1 + ρ 2 ( 1 + ρ ) λ θ Δ s 2 1 + 3 ρ S ( ε ) + 1 1 q 2 1 + 3 ρ c 1 + ρ M Ω ω , X 1 + ρ 2 ( 1 + ρ ) λ θ ε 2 1 + 3 ρ .
Here, we denote the number of the internal iterations performed by the s-th procedure G M A P L with N s . □
We present a few remarks on the iteration complexity of MAPL method.
Remark 1.
According to the classic complexity theory [32] for CP problem (7), the number of iterations to find an ε-solution i.e., an approximate solution x ¯ X such that f ( x ¯ ) f * ε , does not exceed O ( 1 / ε 2 ) , if f is a general non-smooth Lipschitz continuous convex function. For smooth convex optimization, the optimal iteration complexity bound is O ( 1 / ε 1 2 ) . Furthermore, in case that f is weakly smooth and its gradient is H o ¨ lder continuous, the optimal iteration complexity is bounded by O 1 / ε 2 1 + 3 ρ for some ρ ( 0 , 1 ) . It follows from Theorem 1 that the iteration complexity bound of the MAPL method is
O M D X 1 + ρ ε 2 1 + 3 ρ .
In other words, the MAPL method can achieve uniformly optimal iteration complexity bounds for solving non-smooth ( ρ = 0 ), smooth ( ρ = 1 ) and weakly smooth ( ρ ( 0 , 1 ) ) CP problems.

4. The Modified Uniform Smoothing Level (MUSL) Method

In this section we consider the objective function f in (7) with the form of
f ( x ) : = f ^ + F ( x ) ,
where f ^ : X R is Lipschitz continuous and simple. In addition, F ( x ) has a special structure that
F ( x ) : = max y Y { A x , y g ^ ( y ) } .
Here, the compact convex set Y R n is nonempty, A : R n R m is a linear operator and g ^ : Y R is convex and continuous on Y.
Generally, the function F is convex and non-smooth. In this case, F can be approximated by constructing a series of smooth convex functions [16,17]. Let v ( y ) denote the prox-function of compact convex set Y with coefficient σ v . c v : = arg min y Y v ( y ) refers to the prox-center of v ( y ) . Let
V ( y ) : = v ( y ) v ( c v ) v ( c v ) , y c v .
The functions F ( x ) and f ( x ) are approximated by F η ( x ) and f η ( x ) (with some smoothing parameter η > 0 ), respectively, i.e.,
F η ( x ) : = max y Y { A x , y g ^ ( y ) η V ( y ) } ,
f η ( x ) : = f ^ + F η ( x ) .
As described in [17], from the first-order optimality conditions, the convexity of g ^ ( y ) and the strong convexity of V ( y ) , we know that the gradient of F η ( x ) is Lipschitz continuous with Lipschitz constant
L η L ( F η ) : = A 2 η σ v .
Furthermore, we have
F η F ( x ) F η + η D v , Y ,
and
f η f ( x ) f η + η D v , Y ,
for any x X .
Inspired by the smoothing technique of [16,17], Lan [24] proposed the uniform smoothing level (USL) method for solving structure non-smooth CP problem (7) with f being defined by (32). The advantage of the USL method is that the parameter η and estimation of D v , Y can be automatically adjusted and obtained during the gap reduction procedure G U S L , which makes the USL method free of parameters. However, similar to the APL method, each iteration of the USL method also involves two subproblems. Based on the USL method, combining with the analysis of the MAPL method in the previous section, we propose the modified uniform smoothing level (MUSL) method. Same as the USL method, the MUSL method can achieve the optimal iteration complexity when solving problem (7) with (32), but only one subproblem is required to solve in each iteration.
We next describe the gap reduction procedure G M U S L of MUSL method. As the internal procedure of MUSL method, G M U S L is called to compress the gap between upper and lower bounds of f * .
The MUSL gap reduction procedure: ( p + , lb + , D ˜ + ) = G M U S L ( p , lb , D ˜ , λ , θ )
  • Step 0: (Initialization) Set e r r o r t y p e T i t l e x 0 u : = p , f ¯ 0 : = f ( e r r o r t y p e T i t l e x 0 u ) , f ̲ 0 : = lb and : = λ f ̲ 0 + ( 1 λ ) f ¯ 0 , x 0 u : = e r r o r t y p e T i t l e x 0 u . Let
    η : = θ ( f ¯ 0 ) / ( 2 D ˜ ) .
    Choose x 0 X and the initial localizer X 0 = X . The prox-function d ω ( x ) is defined in (15). Let k = 1 .
  • Step 1: (Update level set) Let
    x k l : = 1 α k x ˜ k 1 u + α k x k 1 ,
    h η ( x k l , x ) : = f ^ ( x ) + F η ( x k l ) + F η ( x k l ) , x x k l ,
    X ̲ k = x X k 1 : h ( x k l , x ) .
  • Step 2: (Update prox-center and lower bound) Let
    x k = argmin x X ̲ k d ω ( x ) .
    If X ̲ k = , then stop the procedure and output p + = x k 1 u , l b + = , D ˜ + = D ˜ .
  • Step 3: (Update upper bound) Let
    x ˜ k u = α k x k + ( 1 α k ) x ˜ k 1 u ,
    x k u = x ˜ k u , i f f ( x ˜ k u ) < f ¯ k 1 , x k 1 u , o t h e r w i s e ,
    and f ¯ k : = f ( x k u ) . Check the following two possible stopping rules:
    • (3a) if f ¯ k + θ ( f ¯ 0 ) , stop the procedure with p + = x k u , lb + = lb , D ˜ + = D ˜ .
    • (3b) Otherwise, if f η ( x ˜ k u ) + θ 2 ( f ¯ 0 ) , stop the procedure with p + = x k u , 1 b + = lb , D ˜ + = 2 D ˜ .
  • Step 4: (Update localizer) Choose an arbitrary X k such that X ̲ k X k X ¯ k , where
    X ¯ k = x X : d ω ( x k ) , x x k 0 .
  • Step 5: (Loop) Set k = k + 1 and return to Step 1.
The following are a few remarks to the procedure G M U S L including the differences from G M A P L and some properties. Firstly, compared with G M A P L , the procedure G M U S L needs to input one more parameter D ˜ which is used to calculate the smoothing parameter η , so that the approximate function f η of f can be defined. Secondly, these two procedures approximate the objective function in different ways. G M A P L approximates the objective function f by the linearization of f (18), while G M U S L only linearizes the function F η and approximates the function f with h η (40). According to (40), convexity of F η , (36) and (32), we have
h η ( x k l , x ) f ^ ( x ) + F η ( x ) f ^ ( x ) + F ( x ) = f ( x ) , x X ,
which means that the function h η ( x k l , x ) is the lower estimate of f ( x ) . Thirdly, the procedure G M U S L has one more exit than G M A P L with three possible exits: Step 2, Step 3a, and Step 3b. If the procedure stops at Step 2, the lower bound lb is updated to ; if it stops at Step 3a, we say that a significant improvement has been made on the upper bound of f * and update the upper bound f ¯ k . In addition, if the procedure stops at Step 3b, it is considered that there is no significant improvement on the upper bound of f * , thus the parameter D ˜ needs to be adjusted to estimate D v , Y and we update D ˜ to 2 D ˜ .
Referring to the work of [24,31], the following lemma gives some simple observations of procedure G M U S L . Its proof can be found in Appendix A.
Lemma 2.
The following results hold for internal procedure G M U S L :
a. If G M U S L terminates at Step 2 or Step 3a, we have f ( p + ) lb + q ( f ( p ) lb ) . where q q ( λ , θ ) = max { λ , 1 λ ( 1 θ ) } ;
b. If G M U S L terminates at Step 3b, we have D ˜ < D v , Y and D ˜ + < 2 D v , Y .
Similar to the previous section, it is time to build the convergence results for procedure G M U S L . Referring to the theoretical analysis mode of [24], the detailed derivations of the iteration complexity are given as follows.
Proposition 2.
Let { α k } satisfy (24) and take γ = 1 . Let ( x k l , x k , x k u ) X × X × X , k 1 be the search points, ℓ be the level parameter, d ω ( · ) be the prox-function and η be the smoothing parameter. Then the number of iterations of internal procedure G M U S L can be bounded by,
N Δ 0 , D ˜ : = 2 c A λ θ Δ 0 D ω , X D ˜ σ ω σ v ,
where Δ 0 = ub 1 b = f ¯ 0 f ̲ 0 , D ω , X is defined in (14).
Proof. 
Suppose the gap reduction procedure G M U S L does not stop at the K-th iteration. From (34) and inequality (45), we have h η ( z , x ) f η ( x ) , z , x X . Because of (34), (40) and the Lipschitz continuity of the subgradient of F η , we have
f η ( x ) h η ( z , x ) = F η ( x ) [ F η ( z ) + F η ( z ) , x z ] L η 2 x z 2 , z , x X .
Hence,
f η ( x ) h η + L η 2 x z 2 , x , z X .
From (40), the definition of x ˜ k u (43), the convexity of f ^ , (45) and (44), we have for all k 1 ,
h η ( x k l , x ˜ k u ) = f ^ ( x ˜ k u ) + F η ( x k l ) + F η ( x k l ) , x ˜ k u x k l = f ^ ( α k x k + ( 1 α k ) x ˜ k 1 u ) + F η ( x k l ) + F η ( x k l ) , α k x k + ( 1 α k ) x ˜ k 1 u x k l ( 1 α k ) h η ( x k l , x ˜ k 1 u ) + α k h η ( x k l , x k ) ( 1 α k ) f η ( x ˜ k 1 u ) + α k h η ( x k l , x k ) ( 1 α k ) f η ( x ˜ k 1 u ) + α k .
Moreover, by (39) and (43) we can easily obtain
x ˜ k u x k l = α k x k + ( 1 α k ) x ˜ k 1 u α k x k 1 ( 1 α k ) x ˜ k 1 u = α k ( x k x k 1 ) .
It then follows that
f η ( x ˜ k u ) ( 1 α k ) f η ( x ˜ k 1 u ) + α k + L η α k 2 2 x k x k 1 2 .
Subtracting both sides of the above inequality by , then dividing both sides by α k 2 , we have
f η ( x ˜ k u ) α k 2 1 α k α k 2 [ f η ( x ˜ k 1 u ) ] + L η 2 x k x k 1 2 .
Summing up the above inequalities over k, together with (24) and H o ¨ lder inequality, we obtain
f η ( x ˜ k u ) L η α k 2 2 k = 1 K x k x k 1 2 c 2 L η d ω ( x K ) σ ω K 2 .
Thus, from the relation d ω ( x K ) D ω , X , (35) and (38), we have
f η ( x ˜ K u ) c 2 L η D ω , X σ ω K 2 = c 2 A 2 D ω , X η σ ω σ v K 2 = 2 c 2 A 2 D ω , X D ˜ θ ( f ¯ 0 ) σ ω σ v K 2 .
From the fact that the procedure does not stop at Step 3b at the K-th iteration, we obtain f η ( x ˜ K u ) > θ ( f ¯ 0 ) / 2 . Noticing that Δ 0 = ( f ¯ 0 f ̲ 0 ) = ( f ¯ 0 ) / λ , thus
f η ( x ˜ K u ) > λ θ Δ 0 2 .
In conclusion, we obtain
K < 2 c A λ θ Δ 0 D ω , X D ˜ σ ω σ v .
 □
Based on the results of the convergence of the procedure G M U S L above, we next give the algorithm of the MUSL method. Similar to the MAPL method, the MUSL method is also implemented with outernal algorithm framework and internal gap reduction procedure G M U S L . The outernal algorithm of the MUSL method is mainly to determine whether the gap between the upper and lower bounds on f * has reached the given tolerance ε in current iteration k. If the given tolerance ε is reached, the algorithm terminates and output an approximate optimal solution of f, otherwise the outernal algorithm call the internal procedure G M U S L continuously to compress the gap between the upper and lower bounds. The algorithmic framework of the MUSL method is as follows (Algorithm 3).
Algorithm 3: The MUSL method.
   Step 0: (Input) Choose p 0 X , stopping tolerance ε > 0 , initial estimate D ˜ 1 ( 0 , D v ] and parameters λ , θ ( 0 , 1 ) .
  Step 1: (Initialization) Set
p 1 Argmin x X h 0 ( p 0 , x ) : = f ^ ( x ) + F p 0 + F p 0 , x p 0 ,

   lb 1 = h p 0 , p 1 and ub 1 = min { f p 1 , f ( p ˜ 1 ) } . Let s = 1 ;
  Step 2: (Stopping test) If ub s lb s ε , terminate;
  Step 3: (Call G M U S L procedure) Set ( p s + 1 , lb s + 1 , D ˜ s + 1 ) = G M U S L p s , lb s , D ˜ s , λ , θ and ub s + 1 = f ( p s + 1 ) ;
  Step 4: (Loop) Set s = s + 1 , and return to Step 2
We now turn to analyze the optimal complexity bound for the MUSL method. Please note that the following results are modifications of those in [24].
Lemma 3.
Suppose that F is defined in (33) and v is a prox-function of Y with strong convex modulus σ v . Then, we have the following relation
F ( x 0 ) F ( x 1 ) F ( x 1 ) , x 0 x 1 2 2 A 2 D v , Y σ v 1 2 x 0 x 1 , x 0 , x 1 R n .
Here F ( x 1 ) F ( x 1 ) denotes the subgradient of F at x 1 and D v , Y is defined in (13).
Theorem 2.
Suppose that { α k } in procedure G M U S L is chosen to satisfy condition (24), and that γ = 1 . Then for given ε > 0 , the following statements hold for the MUSL method.
(1) the number of non-significant phases can be bounded by
S 1 ( D ˜ 1 ) : = log 2 D v , Y D ˜ 1 ,
and the number of significant phases can be bounded by
S 2 ( ε ) : = max 0 , log 1 q 2 A Ω ω , X 1 2 Ω v , Y 1 2 ε .
(2) the total number of iterations performed by the MUSL method does not exceed
S 1 ( D ˜ 1 ) + S 2 ( ε ) + c ¯ Ω ω , X 1 2 Ω v , Y 1 2 ε ,
where c ¯ = c A λ θ ( 2 / ( 1 q ) + 2 + 1 ) , and Ω ω , X and Ω v , Y are defined in (14).
Proof. 
(1) Let Δ s = ub s lb s , s 1 . Without loss of generality, we suppose that Δ s > ε . From conclusion (2) of Lemma 2, we konw that D ˜ S 1 ( D ˜ ) 1 + 1 = 2 D ˜ S 1 ( D ˜ 1 ) < D v , Y , if non-significant phase occurs. From D ˜ S 1 + 1 = 2 S 1 ( D ˜ 1 ) D ˜ 1 and 2 S 1 ( D ˜ 1 ) D ˜ 1 < D v , Y , we have
S 1 ( D ˜ ) < log 2 D v , Y D ˜ 1 .
Furthermore, by (47) and (48) in Lemma 3 we have
Δ 1 = f ( p 1 ) h 0 ( p 0 , p 1 ) = F ( p 1 ) F ( p 0 ) + F ( p 0 ) , p 1 p 0 2 2 A 2 D v , Y σ v p 1 p 0 2 A 2 D v , Y σ v , Y 2 D ω , X σ ω = 2 A Ω ω , X 1 2 Ω v , Y 1 2 .
From (e) in Lemma 1 and the definition of ub s and lb s , we know that ε < Δ S 2 ( ε ) + 1 q Δ S 2 ( ε ) q S 2 ( ε ) Δ 1 and ( 1 q ) S 2 ( ε ) Δ 1 ε . In conclusion, we obtain that
S 2 ( ε ) log 1 q Δ 1 ε log 1 q 2 A Ω w , X 1 2 Ω v , Y 1 2 ε .
(2) Let K 1 = { a 1 , a 2 , , a m } and K 2 = { b 1 , b 2 , , b m } denote the index sets of non-significant and significant phases respectively. For any 1 i m in the non-significant phases, we have D ˜ a i + 1 = 2 D ˜ a i . Then we know that D ˜ a i = 2 i 1 D ˜ a 1 = 2 i 1 D ˜ 1 and Δ a i > ε . Due to Proposition 2, we know that K decreases about the first variable and increases about the second variable monotonously. In addition, we conclude that the total number of iterations performed by the significant phases can be bounded by
N 1 = i = 1 m K MUSL Δ a i , D ˜ a i i = 1 m K MUSL ( ε , D ˜ a i ) m + i = 1 m 2 c A λ θ ε D ω , X D ˜ a i σ ω σ v m + i = 1 m 2 c A λ θ ε D ω , X 2 i 1 D ˜ a 1 σ w σ v S 1 D ˜ 1 + 2 c A λ θ ε D ω , X D ˜ 1 σ ω , σ v i = 1 S 1 D ˜ 1 2 i 1 2 S 1 D ˜ 1 + 2 c A ( 2 1 ) λ θ ε D ω , X D v , Y σ ω σ v = S 1 D ˜ 1 + c A Ω ω , X 1 2 Ω v , Y 1 2 ( 2 1 ) λ θ ε .
From the fact that Δ b j + 1 q Δ b j , 1 j n , we know that
Δ b j q j n Δ b n > ε q j n .
It follows from the monotonic of K about variables and Proposition 2 that D ˜ b j 2 D v , Y . In addition, we obtain that the total number of iterations performed in the non-significant phases can be bounded by
N 2 = j = 1 n K MUSL ( Δ b j , D ˜ b j ) j = 1 n K MUSL ( ε q j n , 2 D v , Y ) n + j = 1 n 2 c A λ θ ε q j n D ω , X 2 D v , Y σ ω σ v n + 2 c A λ θ ε 2 D ω , X D v , Y σ ω σ v j = 1 n q n j S 2 ( ε ) + 2 c A λ θ ε 2 D ω , X D v , Y σ ω σ v j = 1 S 2 ( ε ) q n j S 2 ( ε ) + 2 c A ( 1 q ) λ θ ε 2 D ω , X D v , Y σ ω σ v = S 2 ( ε ) + 2 c A Ω ω , X 1 2 Ω v , Y 1 2 ( 1 q ) λ θ ε .
Combining (49) and (50), we conclude that the total number of iterations of the MUSL method does not exceed
N 1 + N 2 = S 1 D ˜ 1 + c A Ω ω , X 1 2 Ω v , Y 1 2 ( 2 1 ) λ θ ε + S 2 ( ε ) + 2 c A Ω ω , X 1 2 Ω v , Y 1 2 ( 1 q ) λ θ ε = S 1 D ˜ 1 + S 2 ( ε ) + 1 2 1 + 2 1 q c A Ω ω , X 1 2 Ω v , Y 1 2 λ θ ε .
 □
In what follows, we apply the MUSL method to solve a special problem where f ^ in (32) is a smooth convex function. In addition, the complexity results of procedure G M U S L and MUSL method are established in this special case. f ^ is a smooth convex function means that there exists a constant L 1 > 0 such that f ^ satisfies
f ^ ( z ) f ^ ( x ) f ^ ( x ) , z x L 1 2 z x 2 , z , x X .
From the fact that F η has a Lipschitz continuous gradient with Lipschitz-constant L η > 0 , we have
F η ( z ) F η ( x ) F η ( x ) , z x L η 2 z x 2 , z , x X .
Thus, we know that f η : = f ^ ( x ) + F η ( x ) is a smooth function on X and satisfies
f η f η f η , z x L 2 z x 2 , z , x X ,
where L = L 1 + L η . From the fact that F is non-smooth, we konw that f is non-smooth as well.
We make the following changes to the outernal algorithm of the MUSL method and internal procedure G M U S L . Replace (47) in outernal algorithm with
p 1 argmin x X h 0 p 0 , x : = f η p 0 + f η p 0 , x p 0 ,
and (40) in procedure G M U S L with
h η ( x k l , x ) : = f η ( x ) + f η ( x k l ) + f η ( x k l ) , x x k l .
The function f η is fixed as a result of the fact that η is fixed in the internal procedure G M U S L . Similar to Propostion 2, we have the following results.
Theorem 3.
We assume that { α k } satisfies condition (24) and take γ = 1 . Let ℓ be the level parameter, d ω ( · ) denotes the prox-function and η is the smoothing parameter in procedure G M U S L . If f ^ is a smooth function, then the number of iterations of procedure G M U S L can be bounded by,
N ¯ : = c Ω ω , X 1 2 L 1 θ λ Δ 0 + 2 A θ λ Δ 0 D ˜ σ v , Y + 1 ,
where Δ 0 = ub 1 b = f ¯ 0 f ̲ 0 , and Ω ω , X is defined in (14).
Proof. 
Suppose that the procedure G MUSL does not stop in iteration K ¯ . Because of the features of f ^ and F η , we know that f η satisfies (52). Let L η = L . From (14), we have
f η ( x ˜ K ¯ u ) c 2 L Ω w , X 2 K ¯ 2 , K ¯ 1 .
Then according to the stopping test in Step 3b in procedure G M U S L , we have
f η x ˜ K u > λ θ Δ 0 2 .
Due to L = L 1 + L η , (35) and (38) we obtain that
K ¯ < c Ω ω , X 1 2 L λ θ Δ 0 < c Ω ω , X 1 2 ( L 1 λ θ Δ 0 + L η λ θ Δ 0 ) = c Ω ω , X 1 2 L 1 λ θ Δ 0 + 2 A λ θ Δ 0 D ˜ σ v , Y .
Thus, we have
N ¯ = c Ω ω , X 1 2 L 1 λ θ Δ 0 + 2 A λ θ Δ 0 D ˜ σ v , Y + 1 .
 □
Here we will also establish the complexity of the MUSL method under this special case where f ^ is smooth and convex.
Theorem 4.
Assume that { α k } satisfy condition (24) and take γ = 1 . If f ^ is a smooth convex function, then the following statements hold for the MUSL method.
(1) the number of non-significant phases can be bounded by
S ¯ 1 : = log 2 D v , Y D ˜ 1 ,
and the number of significant phases can be bounded by
S ¯ 2 : = max 0 , log 1 q 2 A Ω ω , X 1 2 Ω v , Y 1 2 ε .
(2) the total number of iterations performed by the MUSL method does not exceed
S ¯ 1 + S ¯ 2 + S ¯ 1 + 1 1 q c Ω w , X 1 2 L 1 λ θ ε + 1 2 1 + 2 1 q c A Ω ω , X 1 2 Ω v , Y 1 2 λ θ ε .
Proof. 
Let Δ s = ub s lb s , s 1 . Similar to Theorem 2, when non-significant phase occurs we have S ¯ < log 2 D v , Y D ˜ 1 . By (53), (48) and (51) we have
Δ 1 = f ( p 1 ) h 0 ( p 0 , p 1 ) [ F ( p 1 ) F ( p 0 ) F ( p 0 ) , p 1 p 0 ] + [ f ^ ( p 1 ) f ^ ( p 0 ) f ^ ( p 0 ) , p 1 p 0 ] 4 A D v , Y D ω , X σ v σ ω + L 1 2 D ω , X = 2 A Ω ω , X 1 2 Ω v , Y 1 2 + L 1 2 D ω , X .
From (e) in Lemma 1 and the definition of ub s and lb s , we know that ε < Δ S ¯ 2 + 1 q Δ S ¯ 2 q S ¯ 2 Δ 1 and ( 1 q ) S ¯ 2 Δ 1 ε . In conclusion, we obtain that
S ¯ 2 log 1 q Δ 1 ε log 1 q 2 A Ω ω , X 1 2 Ω v , Y 1 2 + L 1 2 D ω , X ε .
(2) Let K ¯ 1 = { c 1 , c 2 , , c m } and K ¯ 2 = { d 1 , d 2 , , d m } denote the index sets of non-significant and significant phases respectively. For any 1 i m in the non-significant phases, we have Q c i + 1 = 2 Q c i . Then we know that Q c i = 2 i 1 Q c 1 = 2 i 1 Q 1 and Δ c i > ε . By Proposition 2, we know that K ¯ decreases about the first variable and increases about the second variable monotonously. It follows that the total number of iterations performed by the significant phases can be bounded by
N ¯ 1 = i = 1 m K MUSL Δ c i , D ˜ c i i = 1 m K MUSL ε , D ˜ c i m + i = 1 m 2 c A λ θ ε D ω , X D ˜ c i σ ω σ v m + i = 1 m 2 c A λ θ ε D ω , X 2 i 1 D ˜ c 1 σ ω σ v S ¯ 1 1 + c 2 D ω , X L 1 σ ω λ θ ε + 2 c A λ θ ε D ω , X D ˜ 1 σ ω σ v i = 1 S ¯ 1 2 i 1 2 S ¯ 1 1 + c 2 D ω , X L 1 σ ω λ θ ε + 2 c A ( 2 1 ) λ θ ε D ω , X D v , Y σ w σ v = S ¯ 1 1 + c Ω ω , X 1 2 L 1 λ θ ε + c A ( 2 1 ) λ θ ε Ω ω , X 1 2 Ω v , Y 1 2 .
From the fact that Δ d j + 1 q Δ d j 1 j n , we know that
Δ d j q j n Δ d n > ε q j n .
By the monotonic of K ¯ about variables and the fact that D ˜ b j 2 D v , Y , we know that the total number of iterations performed in the non-significant phases can be bounded by
N ¯ 2 = j = 1 n K MUSL Δ d j , D ˜ d j j = 1 n K MUSL ε q j n , 2 D v , Y n + j = 1 n 2 c A λ θ ε q j n D ω , X 2 D v , Y σ ω σ v n + 2 c A λ θ ε 2 D ω , X D v , Y σ ω σ v j = 1 n q n j S ¯ 2 + c 2 D ω , X L 1 σ ω λ θ ε j = 1 S ¯ 2 q n j 2 + 2 c A λ θ ε 2 D ω , X D v , Y σ ω σ v j = 1 S ¯ 2 q n j S ¯ 2 + c 1 q 2 D ω , X L 1 σ ω λ θ ε + 2 c A ( 1 q ) λ θ ε 2 D ω , X D v , Y σ ω σ v = S ¯ 2 + c Ω ω , X 1 2 1 q L 1 λ θ ε + 2 c A Ω ω , X 1 2 Ω v , Y 1 2 ( 1 q ) λ θ ε .
This together with (54) and (55) shows that the total number of MUSL method does not exceed N ¯ 1 + N ¯ 2 .  □

5. Two-Stage Stochastic Programming and Numerical Experiments

In this section, the MAPL and MUSL methods are applied to solve the two-stage stochastic programming problems. Furthermore, we compare the two modified methods with the APL and USL methods. All algorithms are implemented in MATLAB (R2014a), and Mosek (8.0) is called for solving subproblems. The environment in which the program runs is Windows 7 (64 bite), Intel(R) Core(TM) i7-6700 CPU 3.40GHz and 16G memory.
Consider the following two-stage stochastic programming with recourse
min { f ( x ) = c T x + E [ Q ( x , ξ ) ] : x X } ,
where
Q ( x , ξ ) = min { q T y : W y = h + T x , y 0 } .
Here, x R n 1 and y R n 2 are the decision variables of the first-stage problem and the second-stage problem respectively. X is a compact convex set, and ξ = ( q , h , T ) R n 1 + m 2 + m 2 + n 1 is a random vector with a known probability distribution and support Ξ . In [33], it was pointed out that by strong convexity one has
Q ( x , ξ ) = max { π T ( h + T x ) : W T π q } .
Let Π ( q ) = { π R m 2 : W T q } be the feasible set of (58) and we assume that Π ( q ) , q Ξ . It is easy to know that Q ( x , ξ ) is generally non-smooth. Therefore, for general distributions of ξ , (56) is a non-smooth convex programming problem. In [33] Ahmed applied Nesterov’s smoothing technique to the two-stage stochastic programming problem (56), and established a proper smooth approximation to Q ( x , ξ ) as follows. For a given smoothing parameter μ , consider the function f μ ( x ) with the form
f μ ( x ) = c T x + E [ Q μ ( x , ξ ) ] ,
where
Q μ ( x , ξ ) = max { π T ( h + T x ) μ 2 π T π : π Π ( q ) } .
For ξ with discrete distribution, it is shown in [34,35] that f μ is differentiable and its gradient is Lipschitz continuous. Furthermore, when μ is sufficiently small, f μ can approximate f uniformly. Based on the above analysis, we next carry out the MAPL and MUSL methods to solve the problem (59) and (60) to illustrate the effectiveness of the methods and compare them with the APL and USL methods.
We perform numerical experiments on some existing SP instances in [1,36,37] including a telecommunication design (SSN) problem and a motor freight carrier routing problem (20Term). The SSN problem studied by Sen, Doverspike, and Cosares [1] comes from the telecommunications industry: the first-stage problem allocates capacity between network links, and the second-stage problem generates demands to connections requested between point-to-point pairs. The 20Term problem studied by Mak, Morton, and Wood [37] comes from a motor freight carrier’s model: the first-stage problem determines a program of carriers, and the second-stage problem adjusts the program according a multi-commodity network. The instances of SSN, 20Term and Storm are downloaded from the link: http://pwp.gatech.edu/guanghui-lan/computer-codes/. The dimensions related to these instances are shown in Table 1. n i denotes the number of constraints in the i-stage problem, and m i denotes the number of variable in the i-stage problem. In particular, we assume that the number of possible realizations of ξ is fixed, i.e. N = 50 or 100. In this case, a total of five instances are tested. The integers in the brackets are the number of possible realizations. For given parameters λ ( 0 , 1 ) , θ ( 0 , 1 ) , tolerance ε = 1.0 e 6 and stepsize α k = 2 / ( k + 1 ) such that (24), we compare the number of iterations and CPU time of different methods. The results are shown in Table 2, Table 3 and Table 4.
There are some observations on the results in Table 2, Table 3 and Table 4. When the initial gap Δ is large (up to 10 2 or 10 7 ), MAPL and APL algorithms are implemented 400 times. The results in Table 2 show that MAPL method has certain advantages over APL method in terms of CPU time and the number of iterations. The results in Table 3 show that in addition to the advantages in CPU time, MUSL algorithm can achieve higher accuracy. From the results in Table 4 we know that compared with the MUSL algorithm, the MAPL algorithm has less CPU time and fewer number of iterations. This is because that while the MAPL method solves 2 N linear programming problems, MUSL only needs to solve N smooth quadratic programming problems in the progress of algorithm.

6. Conclusions

In this paper, we presented two modified BL-type methods, the modified accelerated prox-level (MAPL) and modified uniform smoothing level (MUSL) methods, for uniformly solving the black-box CP problems and a class of structure non-smooth problems. In addition, both MAPL and MUSL methods can achieve optimal complexities respectively. To illustrate the effectiveness of the modified methods, they were then applied to solve the two-stage stochastic programming with recourse and numerical experiments were carried out. Finally, the numerical results shown that the MAPL and MUSL methods have certain advantages in algorithm efficiency and solution time.

Author Contributions

C.T. mainly contributed to the algorithm design and convergence analysis; B.H. mainly contributed to the convergence analysis and numerical results; and Z.W. mainly contributed to the algorithm design. All authors have read and agree to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation (11761013, 71861002) and Guangxi Natural Science Foundation (2018GXNSFFA281007) of China.

Acknowledgments

The authors would like to thank for the support funds.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Lemma 1.
(a) By using mathematical induction, we first show part a). It is obvious that X f ( ) X 0 . Assume that X k 1 is a localizer of X f ( ) i.e., X f ( ) X k 1 , for all k 1 . Then we have that y X k 1 , for all y X f ( ) . By the definition of h and convexity of f we know that h ( x k l , y ) f ( y ) , for any y X f ( ) . From the fact that X f ( ) X ̲ k and X ̲ k X k we have that X f ( ) X k .
(b) It follows from f ¯ k : = min f ¯ k 1 , f x ˜ k u f ¯ k 1 in Step 3 of procedure G M A P L that f * f ¯ k f ¯ k 1 f ¯ 1 f ¯ 0 , for all k 1 .
(c) Form a) we have that X f ( ) X k . By the optimal condition of (20) we have that f ( x k ) N ( x k ) . Then we obtain that Δ d ω ( x k ) , y x k 0 , for any y X ̲ k . By the definition of X ¯ k we have y X ¯ k and further X ̲ k X ¯ k , i.e., there exists a set X k such that X ̲ k X k X ¯ k . In conclusion, we have X f ( ) X ̲ k X k X ¯ k , for all k 1 .
(d) It follows from the fact that X is a compact convex set and the definition of X ̲ k that X ̲ k is a compact convex set. Together with the strong convexity of d ω , we know that there is a unique optimal solution for (20) if X ̲ k . Then we have < f ( x ) , for any x X . Furthermore, we know that f * .
(e) From b) we know that f ( p + ) f ¯ 0 = f ( p ) . We have lb + = if the procedure G M A P L stops at Step 2. From the fact that = λ lb + ( 1 λ ) f ( p ) we have
f ( p + ) lb + f ( p ) [ λ lb + ( 1 λ ) f ( p ) ] = λ ( f ( p ) lb ) .
We have lb + = lb and f ( P + ) = f ( x k u ) + θ ( f ( p ) ) if the procedure G M A P L stops at Step 3. From the fact that = λ lb + ( 1 λ ) f ( p ) , we have
f ( p + ) lb + + θ ( f ( p ) ) lb = ( 1 θ ) + θ f ( p ) lb = [ 1 λ ( 1 θ ) ] ( f ( p ) lb ) .
 □
Proof of Lemma 2.
(a) The proof of a) is the same as the proof of e) of Lemma 1.
(b) If the internal procedure G M U S L stops at Step 3, we have that f ¯ k + θ ( f ¯ 0 ) and f η ( x ˜ k u ) + θ 2 ( f ¯ 0 ) . Then we have that
f ( x ˜ k u ) f η ( x ˜ k u ) f ¯ k f η ( x ˜ k u ) > θ 2 ( f ¯ 0 ) .
Furthermore we know that
η D ˜ = θ 2 ( f ¯ 0 ) < f ( x ˜ k u ) f η ( x ˜ k u ) < η D v , Y .
Then we have D ˜ < D v , Y . Together with the definition D ˜ + in Step 3b, we know that D ˜ + = 2 D ˜ < 2 D v , Y . □

References

  1. Sen, S.; Doverspike, R.D.; Cosares, S. Network planning with random demand. Telecommun. Syst. 1994, 3, 11–30. [Google Scholar] [CrossRef]
  2. Yang, L.F.; Jian, J.B.; Wang, Y.Y.; Dong, Z.Y. Projected mixed integer programming formulations for unit commitment problem. Int. J. Electr. Power Energy Syst. 2015, 68, 195–202. [Google Scholar] [CrossRef]
  3. Yang, L.F.; Jian, J.B.; Zhu, Y.N.; Dong, Z.Y. Tight Relaxation Method for Unit Commitment Problem Using Reformulation and Lift-and-Project. IEEE Trans. Power Syst. 2015, 30, 13–23. [Google Scholar] [CrossRef]
  4. Yang, L.F.; Zhang, C.; Jian, J.B.; Meng, K.; Xu, Y.; Dong, Z.Y. A novel projected two-binary-variable formulation for unit commitment in power systems. Appl. Energy 2017, 187, 732–745. [Google Scholar] [CrossRef]
  5. Wallace, S.W.; Ziemba, W.T. Applications of Stochastic Programming; Society for Industrial and Applied Mathematics and the Mathematical Programming Society: Philadelphia, PA, USA, 2005. [Google Scholar]
  6. Kelley, J.E., Jr. The cutting-plane method for solving convex programs. J. Soc. Ind. Appl. Math. 1960, 8, 703–712. [Google Scholar] [CrossRef]
  7. Veinott, A.F. The Supporting Hyperplane Method for Unimodal Programming. Oper. Res. 1967, 15, 147–152. [Google Scholar] [CrossRef]
  8. Ruszczynski, A. A regularized decomposition method for minimizing a sum of polyhedral functions. Math. Program. 1986, 35, 309–333. [Google Scholar] [CrossRef]
  9. Linderoth, J.; Wright, S. Decomposition Algorithms for Stochastic Programming on a Computational Grid. Comput. Optim. Appl. 2003, 24, 207–250. [Google Scholar] [CrossRef]
  10. Lemaréchal, C. Nonsmooth optimization and descent methods. In Research Report 78-4; IIASA: Laxenburg, Austria, 1978. [Google Scholar]
  11. Mifflin, R. A modification and an extension of Lemaréchal’s algorithm for nonsmooth minimization. In Nondifferential and Variational Techniques in Optimization; Springer: Berlin, Germany, 1982; pp. 77–90. [Google Scholar]
  12. Kiwiel, K.C. An aggregate subgradient method for nonsmooth convex minimization. Math. Program. 1983, 27, 320–341. [Google Scholar] [CrossRef]
  13. Kiwiel, K.C. Proximity control in bundle methods for convex nondifferentiable minimization. Math. Program. 1990, 46, 105–122. [Google Scholar] [CrossRef]
  14. Ruszczyński, A.; Świȩtanowski, A. Accelerating the regularized decomposition method for two stage stochastic linear problems. Eur. J. Oper. Res. 1997, 101, 328–342. [Google Scholar] [CrossRef]
  15. Lemaréchal, C.; Nesterov, Y.; Nemirovskii, A. New variants of bundle methods. Math. Program. 1995, 69, 111–147. [Google Scholar] [CrossRef] [Green Version]
  16. Ben-Tal, A.; Nemirovski, A. Lectures on Modern Convex Optimization: Analysis, Algorithms, and Engineering Applications; SIAM: Philadelphia, PA, USA, 2001. [Google Scholar]
  17. Ben-Tal, A.; Nemirovski, A. Non-euclidean restricted memory level method for large-scale convex optimization. Math. Program. 2005, 102, 407–456. [Google Scholar] [CrossRef] [Green Version]
  18. Richtárik, P. Approximate Level Method for Nonsmooth Convex Minimization. J. Optim. Theory Appl. 2012, 152, 334–350. [Google Scholar] [CrossRef]
  19. Fischer, F.; Helmberg, C. A parallel bundle framework for asynchronous subspace optimization of nonsmooth convex functions. SIAM J. Optim. 2014, 24, 795–822. [Google Scholar] [CrossRef]
  20. Kim, K.; Petra, C.G.; Zavala, V.M. An asynchronous bundle-trust-region method for dual decomposition of stochastic mixed-integer programming. SIAM J. Optim. 2019, 29, 318–342. [Google Scholar] [CrossRef]
  21. Van Ackooij, W.; Frangioni, A. Incremental bundle methods using upper models. SIAM J. Optim. 2018, 28, 379–410. [Google Scholar] [CrossRef] [Green Version]
  22. Iutzeler, F.; Malick, J.; de Oliveira, W. Asynchronous level bundle methods. Math. Program. 2019, 49, 1–30. [Google Scholar] [CrossRef]
  23. Tang, C.; Jian, J.; Li, G. A proximal-projection partial bundle method for convex constrained minimax problems. J. Ind. Manag. Optim. 2019, 15, 757–774. [Google Scholar] [CrossRef] [Green Version]
  24. Lan, G. Bundle-level type methods uniformly optimal for smooth and nonsmooth convex optimization. Math. Program. 2015, 149, 1–45. [Google Scholar] [CrossRef] [Green Version]
  25. Nesterov, Y. A method for unconstrained convex minimization problem with the rate of convergence O(1/k2). Doklady AN USSR 1983, 269, 543–547. [Google Scholar]
  26. Auslender, A.; Teboulle, M. Interior gradient and proximal methods for convex and conic optimization. SIAM J. Optim. 2006, 16, 697–725. [Google Scholar] [CrossRef]
  27. Lan, G.; Lu, Z.; Monteiro, R.D. Primal-dual first-order methods with O(1/ϵ) iteration-complexity for cone programming. Math. Program. 2011, 126, 1–29. [Google Scholar] [CrossRef]
  28. Lan, G. An optimal method for stochastic composite optimization. Math. Program. 2012, 133, 365–397. [Google Scholar] [CrossRef] [Green Version]
  29. Nesterov, Y. Introductory Lectures on Convex Optimization a Basic Course; Springer Science & Business Media: New York, NY, USA, 2004; Volume 87. [Google Scholar]
  30. Nesterov, Y. Smooth minimization of non-smooth functions. Math. Program. 2005, 103, 127–152. [Google Scholar] [CrossRef]
  31. Chen, Y.; Lan, G.; Ouyang, Y.; Zhang, W. Fast Bundle-Level Type Methods for Unconstrained and Ball-Constrained Convex Optimization. Comput. Optim. Appl. 2019, 73, 159–199. [Google Scholar] [CrossRef]
  32. Nemirovsky, A.S.; Yudin, D. Problem Complexity and Method Efficiency in Optimization. In Wiley-Interscience Series in Discrete Mathematics; Wiley-Interscience: New York, NY, USA, 1983. [Google Scholar]
  33. Ahmed, S. Smooth Minimization of Two-Stage Stochastic Linear Programs; Georgia Institute of Technology: Atlanta, GA, USA, 2006. [Google Scholar]
  34. Chen, X.; Qi, L.; Womersley, R.S. Newton’s method for quadratic stochastic programs with recourse. J. Comput. Appl. Math. 1995, 60, 29–46. [Google Scholar] [CrossRef] [Green Version]
  35. Chen, X. A parallel BFGS-SQP method for stochastic linear programs. In Computational Techniques and Applications; World Scientific: Princeton, NJ, USA, 1995; pp. 67–74. [Google Scholar]
  36. Linderoth, J.; Shapiro, A.; Wright, S. The empirical behavior of sampling methods for stochastic programming. Ann. Oper. Res. 2006, 142, 215–241. [Google Scholar] [CrossRef]
  37. Mak, W.; Morton, D.P.; Wood, R.K. Monte Carlo bounding techniques for determining solution quality in stochastic programs. Oper. Res. Lett. 1999, 24, 47–56. [Google Scholar] [CrossRef]
Table 1. Data about instances.
Table 1. Data about instances.
n 1 m 1 n 2 m 2 dim ( ξ ) ub 1 Δ 1
SSN(50)89170617586 2.352586 × 10 2 1.008000 × 10 3
SSN(100)89170617586 2.407279 × 10 2 9.878400 × 10 2
20Term(50)63376412440 7.718617 × 10 5 1.804693 × 10 7
20Term(100)63376412440 7.664142 × 10 5 1.801832 × 10 7
Storm(50)1211851259528117 2.138447 × 10 7 6.370092 × 10 6
Table 2. MAPL and APL methods on SP instances.
Table 2. MAPL and APL methods on SP instances.
ProbAlgIterUb δ Time
SSN(50)MAPL135 4.838073 × 10 0 9.042063 × 10 7 63.71
APL140 4.838181 × 10 0 7.252755 × 10 7 66.39
SSN(100)MAPL215 7.352607 × 10 0 9.748218 × 10 6 233.95
APL263 7.352607 × 10 0 9.835275 × 10 7 292.95
20Term(50)MAPL281 2.549453 × 10 5 1.206150 × 10 6 193.9
APL401 2.549453 × 10 5 1.168752 × 10 6 282.83
20Term(100)MAPL386 2.549453 × 10 5 8.772477 × 10 7 519.40
APL401 2.532875 × 10 5 1.059493 × 10 4 536.20
Storm(50)MAPL119 1.562059 × 10 7 1.093373 × 10 6 87.59
APL401 1.547024 × 10 7 2.211075 × 10 3 518.62
Table 3. MUSL and USL methods on SP instances.
Table 3. MUSL and USL methods on SP instances.
ProbAlgIterUb δ TimeQ
SSN(50)MUSL208 4.838072 × 10 0 3.003666 × 10 6 206.81 3.200000 × 10 1
USL322 4.839140 × 10 0 9.337998 × 10 7 309.43 3.200000 × 10 1
SSN(100)MUSL181 7.352608 × 10 0 1.686893 × 10 6 351.11 3.200000 × 10 1
USL190 7.352699 × 10 0 8.517813 × 10 7 367.12 3.200000 × 10 1
20Term(50)MUSL401 2.549453 × 10 5 6.026059 × 10 5 547.93 3.022315 × 10 23
USL401 2.549453 × 10 5 8.844555 × 10 4 522.41 2.147484 × 10 9
20Term(100)MUSL380 2.532875 × 10 5 2.171117 × 10 6 1148.29 1.677722 × 10 7
USL401 2.532875 × 10 5 4.172471 × 10 6 1308.03 1.801440 × 10 16
Storm(50)MUSL299 2.138447 × 10 7 1.087785 × 10 6 437.11 5.242880 × 10 5
USL401 1.546949 × 10 7 2.866387 × 10 4 1084.46 1.511157 × 10 23
Table 4. MAPL and MUSL methods on SP instances.
Table 4. MAPL and MUSL methods on SP instances.
ProbAlgIterUb δ Time
SSN(50)MAPL159 4.838072 × 10 0 1.016148 × 10 6 90.68
MUSL175 4.838073 × 10 0 9.421333 × 10 6 201.82
SSN(100)MAPL190 7.352608 × 10 0 6.397336 × 10 6 217.63
MUSL275 7.352607 × 10 0 2.900934 × 10 6 638.96
20Term(50)MAPL347 2.549453 × 10 5 4.465430 × 10 6 271.94
MUSL401 2.549453 × 10 5 3.039380 × 10 5 663.75
20Term(100)MAPL401 2.532875 × 10 5 3.203178 × 10 2 607.93
MUSL401 2.532875 × 10 5 1.512163 × 10 4 1174.29
Storm(50)MAPL19 1.866006 × 10 7 1.028180 × 10 6 18.08
MUSL29 2.138447 × 10 7 1.799315 × 10 6 61.25

Share and Cite

MDPI and ACS Style

Tang, C.; He, B.; Wang, Z. Modified Accelerated Bundle-Level Methods and Their Application in Two-Stage Stochastic Programming. Mathematics 2020, 8, 265. https://doi.org/10.3390/math8020265

AMA Style

Tang C, He B, Wang Z. Modified Accelerated Bundle-Level Methods and Their Application in Two-Stage Stochastic Programming. Mathematics. 2020; 8(2):265. https://doi.org/10.3390/math8020265

Chicago/Turabian Style

Tang, Chunming, Bo He, and Zhenzhen Wang. 2020. "Modified Accelerated Bundle-Level Methods and Their Application in Two-Stage Stochastic Programming" Mathematics 8, no. 2: 265. https://doi.org/10.3390/math8020265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop