Next Article in Journal
Numerical Solution of Mathematical Model of Heat Conduction in Multi-Layered Nanoscale Solids
Previous Article in Journal
On Polynomials Associated with Finite Topologies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Parallel Primal-Dual Method with Linearization for Structured Convex Optimization

Nanjing Institute of Technology, Nanjing 211167, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(2), 104; https://doi.org/10.3390/axioms14020104
Submission received: 7 January 2025 / Revised: 26 January 2025 / Accepted: 27 January 2025 / Published: 29 January 2025

Abstract

:
This paper presents the Parallel Primal-Dual (PPD3) algorithm, an innovative approach to solving optimization problems characterized by the minimization of the sum of three convex functions, including a Lipschitz continuous term. The proposed algorithm operates in a parallel framework, simultaneously updating primal and dual variables, and offers potential computational advantages. This parallelization can greatly accelerate computation, particularly when run on parallel computing platforms. By departing from traditional primal-dual methods that necessitate strict parameter constraints, the PPD3 algorithm removes reliance on the spectral norm of the linear operator, significantly reducing the computational burden associated with its evaluation. As the problem size grows, calculating the spectral norm, which is essential for many primal-dual methods, becomes progressively more expensive. In addition, adaptive step sizes are computed to accelerate the convergence process. In contrast to most primal-dual approaches that employ a fixed step size constrained by a global upper limit throughout all iterations, the adaptive step size is typically greater and may result in faster convergence. An O(1/k) ergodic convergence rate is proved theoretically. Applications in Fused LASSO and image inpainting demonstrate the method’s efficiency in computation time and convergence rate compared to state-of-the-art algorithms.

1. Introduction

In this paper, we focus on the structured nonlinear optimization problem below:
min x X f ( x ) + g ( x ) + h ( A x ) .
Let A R m × n , and let X R n and Y R m be closed convex sets. Consider three proper lower semi-continuous convex functions f, g, and h; the function f is differentiable, and its gradient is β -Lipschitz continuous, meaning that for all x , y X , we have
f ( x ) f ( y ) β x y 2 .
This condition implies that the difference in function values of f at any two points in X is bounded by a constant β times the Euclidean distance between those points.
It is crucial to understand that this original primal optimization problem embodies a primal-dual representation as follows:
min x X max y Y f ( x ) + g ( x ) + y T A x h ( y ) ,
where h denotes the conjugate of a convex function h, defined as
h ( y ) = sup x dom ( h ) y T x h ( x ) .
The formulation in (1) encompasses a broad spectrum of applications, such as image restoration, machine and statistical learning, compressive sensing, and portfolio optimization. Several illustrative examples are provided below:
Elastic net [1]: The elastic net method incorporates both l 1 and l 2 regularization terms, leveraging the advantages of each. The elastic net estimator is defined as the solution to the following optimization problem:
min x X μ 1 x 1 + μ 2 x 2 2 + ( A x b ) ,
where x R p , A R n × p , b R n , and is a loss function. The second term possesses a gradient that is Lipschitz continuous. When μ 2 = 0 , the elastic net reduces to LASSO regression. Conversely, when μ 1 = 0 , the elastic net simplifies to ridge regression.
Matrix completion [2]: Consider a matrix Y R m × n with entries confined to the interval [ l , u ] , where l < u are positive real constants. The task is to recover a low-rank matrix, X, from the observed elements of Y. Let Ω be the set of Y s non-zero elements, and let P Ω be a linear operator that extracts the available entries of Y, leaving the missing entries as zeros. The problem, as defined in [2], is formulated as the following optimization problem:
min X m × n μ X + 1 2 P Ω ( X Y ) 2 2 s . t . l X u
where μ > 0 is a regularization constant, and · denotes the nuclear norm. By applying the Lagrange multiplier method, the aforementioned optimization task can be restated in the form presented in (2).
Support vector machine classification [3]: The constrained quadratic optimization problem in R d can likewise be reformulated as presented in (2).
min x d 1 2 x Q 2 + c T x s . t . x C 1 C 2
In the dual representation of the soft-margin kernelized support vector machine classifier, we encounter a situation involving a symmetric positive semi-definite matrix Q R d × d , a vector c R d , and two sets of constraints, C 1 and C 2 , within R d . In particular, C 1 represents a bounding constraint, and C 2 refers to a linear constraint.
Before introducing algorithms to solve (2), we begin by reviewing several established techniques for general saddle-point problems, particularly the case of (2), where f is not present. Three key splitting methods are typically used:
  • The forward-backward-forward splitting method [4];
  • The forward-backward splitting method [5];
  • The Douglas-Rachford splitting method [6].
In recent years, several primal-dual splitting algorithms have emerged, grounded in three primary foundational techniques. The primal-dual hybrid gradient (PDHG) method, first introduced in [7], demonstrates significant numerical efficiency when applied to total variation (TV) minimization tasks. However, its convergence is highly sensitive to the choice of parameters. To mitigate this issue, Chambolle and Pock introduced the CP method in [8], which enhances the PDHG approach by adjusting the update rule for the dual variables. This adjustment improves both convergence rates and numerical efficiency. In [9], He and Yuan performed an extensive investigation of the CP approach, offering a comprehensive analysis via the Proximal Point Algorithm (PPA) and introducing a relaxation parameter within the PPA framework to expedite convergence (the HeYuan method). More recently, Goldstein et al. introduced the Adaptive Primal-Dual Splitting (APD) method in [10], which dynamically modifies the step-size parameters of the CP approach. In [11], O’Connor and Vandenberghe presented a primal-dual decomposition strategy based on Douglas-Rachford splitting, which is utilized for various primal-dual optimality condition decompositions. Additionally, the adaptive parallel primal-dual (APPD) method presented in [12] is an enhanced version of the primal-dual approach, designed with parallel computing capabilities to optimize performance.
A comprehensive framework outlining various well-known primal-dual methods tackles the saddle-point problem by following these steps:
x k + 1 = arg min x X { g ( x ) + x T A T y k + 1 2 τ x x k 2 } , x ¯ k = x k + 1 + θ ( x k + 1 x k ) , y k + 1 = arg min y Y { h ( y ) y T A x ¯ k + 1 2 σ y y k 2 } ,
As summarized in Table 1, different values of θ correspond to distinct methods under the framework described in Section 4. It is noteworthy that a straightforward correction step is necessary for both the HeYuan method and the APPD method.
In the context of (2), Condat and Vu proposed a primal-dual splitting framework, the CV scheme, in their studies [13,14]. This approach, based on forward-backward iteration, extends the Chambolle-Pock algorithm (CP) by linearizing the function f, though it operates with a narrower parameter range than CP. Later, the Asymmetric Forward-Backward-Adjoint (AFBA) splitting method, introduced in [15], generalizes the CV scheme. Additionally, the Primal-Dual Fixed-Point (PDFP) algorithm, proposed in [16], incorporates two proximal operations of g in each iteration. Other notable methods for solving (2) include the Primal-Dual Three-Operator splitting method (PD3O) from [17] and DavisYin’s three-operator splitting technique, known as PDDY, introduced in [18]. More recently, a modified primal-dual algorithm (MPD3O) was studied in [19], which combines the idea of the balanced augmented Lagrangian method (BALM) in [20] and PD3O. The parameter conditions are relaxed in [21] for the AFBA/PDDY method.
A similar overarching framework exists for algorithms addressing (2), analogous to the one in (4):
x k + 1 = arg min x X { g ( x ) + x T A T y k + 1 2 τ x x k 2 + x T f ( x k ) } , x ¯ k = x ¯ ( x k , x k + 1 ) k , y k + 1 = arg min y Y { h ( y ) y T A x ¯ k + 1 2 σ y y k 2 } .
The corresponding parameter restrictions for these methods are provided in Table 2. The relationship between the aforementioned methods and the framework described in (5) is detailed in Table 3.
In the Discussion section of Chambolle’s work [8], a prominent challenge for existing primal-dual methods is emphasized: the handling of a linear operator, A, with a potentially large or unknown norm. Numerous methods, including the PDHG method described in [7], the CP algorithm presented in [8], the HeYuan method from [9], the CV scheme introduced in [13,14], the PDFP method proposed in [16], and the PD3O method outlined in [17], all necessitate the satisfaction of the condition τ σ 1 / A T A for convergence. Consequently, the norm of A must be computed prior to initializing the parameters. However, as the dimensionality of A escalates, computing this norm becomes increasingly costly, thereby introducing a substantial computational burden.
To tackle the aforementioned challenge, we introduce the PPD3 algorithm, aimed at minimizing the sum of three functions, one of which includes a Lipschitz continuous term. The method we propose consists of two separate phases. In the first phase, a parallel version of (5), where x ¯ is replaced by x k , functions as a prediction step to produce an intermediate estimate. Subsequently, in the subsequent phase, the step length α k is determined, after which a simple adjustment is applied to the predictor, ensuring the convergence of the PPD3 algorithm.
The key contributions presented in this paper are outlined as follows:
  • In contrast to other primal-dual methods, which usually update the primal and dual variables alternately, the proposed approach computes both variables simultaneously and in parallel. This parallelization can greatly accelerate computation, particularly when run on parallel computing platforms.
  • The proposed algorithm is independent of any prior knowledge about the linear operator A. The constraint τ σ 1 / A T A on the parameters is no longer required. In fact, as the problem size grows, calculating the spectral norm of A, which is essential for many primal-dual methods, becomes progressively more expensive.
  • The step size α k is adjusted dynamically at each iteration. This method tailors the step size for each specific iteration, in contrast to most primal-dual approaches that employ a fixed step size constrained by a global upper limit throughout all iterations. As a result, the adaptive step size, which is typically greater, may result in faster convergence.
The organization of this paper is structured as follows: In Section 2, we introduce the proposed algorithm. Section 3 presents a comprehensive convergence analysis of the algorithm, while Section 4 delves into an examination of its computational complexity. Section 5 reports the performance of the algorithm through diverse applications, accompanied by corresponding numerical results. Lastly, Section 6 summarizes the key conclusions and outlines potential directions for future research endeavors.

2. The Parallel Primal-Dual Algorithm

To facilitate understanding, we combine the primal and dual parameters and introduce them as follows:
u = x y , θ ( u ) = g ( x ) + h ( y ) , F ( u ) = A T y + f ( x ) A x , Ω = X × Y .
The subsequent symbols are employed as well:
Q ( x k , x ˜ k ) = τ 1 ( x k x ˜ k ) A T ( y k y ˜ k ) ( f ( x k ) f ( x ˜ k ) ) A ( x k x ˜ k ) + σ 1 ( y k y ˜ k ) ,
H = τ 1 I n 0 0 σ 1 I m , G = ( τ 1 β ) I n 0 0 σ 1 I m ,
and
M ( x k , x ˜ k ) = ( x k x ˜ k ) τ A T ( y k y ˜ k ) τ ( f ( x k ) f ( x ˜ k ) ) σ A ( x k x ˜ k ) + ( y k y ˜ k ) ,
where τ > β and σ > 0 represent positive parameters. It is clear that
M ( x k , x ˜ k ) = H 1 Q ( x k , x ˜ k ) .
The concurrent primal-dual approach is described in Algorithm 1.
The loop in Algorithm 1 follows a pattern similar to the CV approach, with one key difference: the prediction step (6) utilizes x k rather than x ¯ , as seen in (5). The incorporation of x k provides a notable benefit: it eliminates the requirement for any updated data related to x in the y-subproblem, thus enabling the parallel processing of both the x- and y-subproblems.
Subsequently, the residuals p k and d k are calculated as termination conditions. Once both residuals tend toward zero, we will show (in the subsequent section) that Q ( x k , x ˜ k ) also approaches zero. Consequently, (1) converges to
u ˜ k Ω , θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ˜ k ) 0 , u Ω .
By contrasting this variational inequality with (3), it is evident that u ˜ k meets the criteria of the initial problem (2). Consequently, the convergence of the primal-dual algorithm can be assessed by analyzing the residuals’ norms (see also [10,22,23]).
At each iteration, the step size α k is computed, which addresses the challenge of setting a global upper bound for the step size across all iterations. In traditional primal-dual algorithms, this upper bound is usually constrained by the condition τ σ A T A . In contrast, our approach bypasses this limitation on the parameters τ and σ by adaptively computing α k . Experimental observations indicate that, in certain cases, the step size in our method exceeds the upper bound A T A , which has been found to accelerate the convergence process.
Ultimately, a straightforward adjustment procedure guarantees the algorithm’s convergence.
Algorithm 1: Parallel Primal-Dual Algorithm
  • Initialize τ < 1 / β , σ > 0 , x 0 X , y 0 Y
  • While p k , d k > tolerance do
  • Prediction
    x ˜ k = arg min x X g ( x ) + x T A T y k + 1 2 τ x x k 2 + x T f ( x k ) , y ˜ k = arg min y Y h ( y ) y T A x k + 1 2 σ y y k 2 .
  • Compute residual norms
    p k = τ 1 ( x k x ˜ k ) A T ( y k y ˜ k ) ( f ( x k ) f ( x ˜ k ) ) , d k = σ 1 ( y k y ˜ k ) + A ( x k x ˜ k ) .
  • Calculate step-size
    α k = τ 1 x k x ˜ k 2 + σ 1 y k y ˜ k 2 ( x k x ˜ k ) T ( f ( x k ) f ( x ˜ k ) ) τ p k 2 + σ d k 2 .
  • Correction
    x k + 1 = x k α k x k x ˜ k τ A T ( y k y ˜ k ) τ ( f ( x k ) f ( x ˜ k ) ) , y k + 1 = y k α k y k y ˜ k + σ A ( x k x ˜ k ) .
  • End while
The prediction stage in (6) is a special case of (5), where x ¯ k = x k . Unlike other algorithms, such as the CV method, where x ¯ k = 2 x k + 1 x k , our method uses the previous iterate x k as x ¯ k . When f = 0 , the prediction stage in (6) coincides with (4) when θ = 1 .
Note that our algorithm does not require the condition τ σ A T A for theoretical analysis, and thus, no such constraint is necessary.

3. Convergence Analysis

We begin by providing the following reformulation of the variational inequality (VI) corresponding to (2):
Find u = ( x , y ) such that
VI ( Ω , F , θ ) : u Ω , θ ( u ) θ ( u ) + ( u u ) T F ( u ) 0 , u Ω = X × Y ,
The mapping F ( u ) is composed of the sum of an affine skew-symmetric matrix and a monotone operator, which ensures that F ( u ) itself is monotone. Let Ω represent the solution set to the variational inequality VI( Ω , F, θ ); we assume that this solution set is non-empty.
Lemma 1.
For a given u k = x k y k , let u ˜ k be given by the expression in (6). Then,
u ˜ k Ω , θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ˜ k ) ( u u ˜ k ) T Q ( u k , u ˜ k ) for all u Ω ,
Proof of Lemma 1.
By applying the optimality condition from (6), we derive
g ( x ) g ( x ˜ k ) + ( x x ˜ k ) T { A T y k + τ 1 ( x ˜ k x k ) + f ( x k ) } 0 , x X ,
and
h ( y ) h ( y ˜ k ) + ( y y ˜ k ) T { A x k + σ 1 ( y ˜ k y k ) } 0 , y Y .
By combining the two inequalities above, it follows that u ˜ k = ( x ˜ k , y ˜ k ) Ω , and
g ( x ) g ( x ˜ k ) + h ( y ) h ( y ˜ k ) + x x ˜ k y y ˜ k T A T y ˜ k A x ˜ k + τ 1 I A T A σ 1 I x ˜ k x k y ˜ k y k + f ( x k ) f ( x ˜ k ) 0 0 ,
For every u Ω , the assertion is established through the notation introduced in the preceding section. □
Lemma 2.
For a given u k = x k y k , let u ˜ k be defined by (6), then u ˜ k Ω ,
α k ( θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ˜ k ) ) 1 2 u u k + 1 H 2 1 2 u u k H 2 + α k 2 u u ˜ k G 2 for all u Ω ,
where · H , · G are semi-norm, defined as x H = x T H x .
Proof of Lemma 2.
Step 5 and 6 in Algorithm 1 are equivalent to
α k = ( u k u ˜ k ) T Q ( u k , u ˜ k ) / M ( u k , u ˜ k ) H 2 and
u k + 1 = u k α k M ( u k , u ˜ k ) .
By utilizing Q ( u k , u ˜ k ) = H M ( u k , u ˜ k ) and the relation in (11), the right-hand side of (1) simplifies to ( u u ˜ k ) T H ( u k u k + 1 ) , and therefore
α k ( θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ˜ k ) ) ( u u ˜ k ) T H ( u k u k + 1 ) = 1 2 u u k + 1 H 2 1 2 u u k H 2 + 1 2 u u ˜ k H 2 1 2 u u ˜ k + 1 H 2
For the last two terms, we have
1 2 u u ˜ k H 2 1 2 u u ˜ k + 1 H 2 = 1 2 [ 2 α k ( u k u ˜ k ) T H M ( u k , u ˜ k ) α k 2 M ( u k , u ˜ k ) H 2 ] = α k 2 ( u k u ˜ k ) T H M ( u k , u ˜ k ) = α k 2 x x ˜ k y y ˜ k T τ 1 I A T A σ 1 I x k x ˜ k y k y ˜ k + f ( x k ) f ( x ˜ k ) 0 α k 2 τ 1 x k x ˜ k 2 + σ 1 y k y ˜ k 2 β x k x ˜ k 2 = α k 2 u u ˜ k G 2 ,
The proof of the assertion is, thus, complete. □
Lemma 3.
For a given u k = x k y k , let u ˜ k and u k + 1 be defined by (6) and (8), respectively. If τ 1 σ 1 c A 2 2 , then there exists a positive constant α + such that α + α k for all k > 0 , and the following inequality is satisfied:
u k + 1 u H 2 u k u H 2 α + u k u ˜ k G 2 , u Ω .
Proof of Lemma 3.
By substituting u = u into (10), we derive
α k θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ˜ k ) 1 2 u u k + 1 H 2 1 2 u u k H 2 + α k 2 u u ˜ k G 2 , u Ω .
Since
( u ˜ k u ) T F ( u ˜ k ) ( u ˜ k u ) T F ( u )
and
θ ( u ˜ k ) θ ( u ) + ( u ˜ k u ) T F ( u ) 0 ,
these inequalities can be combined to yield:
u k + 1 u H 2 u k u H 2 α k u k u ˜ k G 2 , u Ω .
Next, let us recall the definition of α k :
α k = ( u k u ˜ k ) T Q ( u k , u ˜ k ) M ( u k , u ˜ k ) H 2 .
We proceed to compute the term ( u k u ˜ k ) T Q ( u k , u ˜ k ) :
( u k u ˜ k ) T Q ( u k , u ˜ k ) = x x ˜ k y y ˜ k T τ 1 I A T A σ 1 I x k x ˜ k y k y ˜ k + f ( x k ) f ( x ˜ k ) 0 τ 1 x k x ˜ k 2 + σ 1 y k y ˜ k 2 β x k x ˜ k 2 = ( 1 τ β ) u u ˜ k H 2 .
Next, we examine the term M ( u k , u ˜ k ) H 2 :
M ( u k , u ˜ k ) H 2 = I τ A T σ A I ( u k u ˜ k ) + τ f ( x k ) f ( x ˜ k ) 0 H 2 2 I τ A T σ A I ( u k u ˜ k ) H 2 + 2 τ 2 f ( x k ) f ( x ˜ k ) 0 H 2 2 I τ A T σ A I u k u ˜ k H 2 + 2 τ 2 β 2 u k u ˜ k H 2 .
Therefore, α k is bounded below by a positive constant α + , and the conclusion follows directly. □
Theorem 1.
Given the sequence u k = ( x k , y k ) , let u ˜ k and u k + 1 be defined by Equations (6) and (8), respectively. It follows that the sequence { u k } converges to a limit point u Ω .
Proof of Theorem 1.
From Equation (12), it follows that the sequence { u k } is bounded, and consequently, we obtain
lim k u k u ˜ k = 0 .
Thus, the sequence { u ˜ k } is also bounded. Let u denote a cluster point of { u ˜ k } , and assume that { u ˜ k j } is a subsequence of { u ˜ k } that converges to u . Furthermore, let { u k } and { u k j } represent the corresponding subsequences induced by { u ˜ k } and { u ˜ k j } , respectively. From Equation (1), it follows that
u ˜ k j Ω , θ ( u ) θ ( u ˜ k j ) + ( u u ˜ k j ) T F ( u ˜ k j ) ( u u k j ) T Q ( u k j u ˜ k j ) , u Ω .
By utilizing the continuity of θ ( u ) and F ( u ) , we derive
w Ω , θ ( u ) θ ( u ) + ( u w ) T F ( u ) 0 , w Ω .
This variational inequality implies that u satisfies the variational inequality VI( Ω , F ).
By applying Equation (16) and using the fact that lim j u ˜ k j = u , it follows that the subsequence { u k j } converges to u . Moreover, from Equation (12), we obtain
u k + 1 u H u k u H ,
This result demonstrates that the sequence { u k } approaches u , thus finalizing the proof. □

4. Computational Complexity

Lemma 4.
The solution set of the variational inequality VI ( Ω , F , θ ) is convex and can be described as
Ω = u Ω u ˜ Ω : θ ( u ) θ ( u ˜ ) + ( u u ˜ ) T F ( u ) 0 .
Proof of Lemma 4.
The proof can be derived as a natural extension of Theorem 2.3.5 in [24], or alternatively, one may consult the proof of Theorem 2.1 in [25] for a related approach. □
Lemma 4 demonstrates that u ˜ Ω constitutes an approximate solution to the variational inequality VI ( Ω , F , θ ) , with an accuracy of ϵ > 0 , provided that it satisfies
θ ( u ) θ ( u ˜ ) + ( u u ˜ ) T F ( u ) ϵ , u Ω D ( u ˜ ) ,
where D ( u ˜ ) = { u u u ˜ 1 } represents the neighborhood of u ˜ .
Theorem 2.
Algorithm 1 demonstrates convergence with an ergodic rate of O ( 1 / t ) . Specifically, let { u k } represent the sequence generated by Algorithm 1, starting from an arbitrary initial point u 0 . For the sequence { u ¯ t } defined by
u ¯ t = 1 k = 0 t 1 α k k = 0 t 1 α k u ˜ k ,
the subsequent convergence bound holds:
θ ( u ) θ ( u ¯ t ) + ( u u ¯ t ) T F ( u ) 1 2 t α + u u 0 H 2 .
Proof of Theorem 2.
By leveraging the monotonicity of F, we obtain
θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ) θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) F ( u ˜ ) .
By substituting this into (12), we derive
α k θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ) 1 2 u u k + 1 H 2 1 2 u u k H 2 u Ω .
It is important to highlight that the above assertion holds when G > 0 .
By summing the inequality in (21) over k = 0 , 1 , , t 1 , we obtain
k = 0 t 1 α k θ ( u ) θ ( u ˜ k ) + ( u u ˜ k ) T F ( u ) 1 2 u u 0 H 2 .
By applying the definition of u ¯ t , this can be expressed as
k = 0 t 1 α k θ ( u ) k = 0 t 1 α k θ ( u ˜ k ) + k = 0 t 1 α k ( u u ¯ t ) T F ( u ) 1 2 u u 0 H 2 .
Given that θ ( u ) is convex, it follows that
k = 0 t 1 α k θ ( u ˜ t ) k = 0 t 1 α k θ ( u ˜ k ) .
Substituting this into the preceding inequality yields
θ ( u ) θ ( u ¯ t ) + ( u u ¯ t ) T F ( u ) 1 2 k = 0 t 1 α k u u 0 H 2 1 2 t α + u u 0 H 2 .
Therefore, the result has been established. □

5. Numerical Experiments

Algorithm 1 was applied to the Fused LASSO and image inpainting problem to show its numerical performance. We compare our method with Condat and Vu’s primal-dual splitting scheme (CV), Yan’s primal-dual three-operator splitting scheme (PD3O), and the Asymmetric Forward-Backward-Adjoint method (AFBA)/Davis and Yin’s three-operator splitting method (PDDY). All codes were written and implemented in Matlab 2024 a , and all experiments were carried out on a computer with a 1.20 GHz Intel Core Ultra 5 125H processor and 32 GB of memory.

5.1. Fused LASSO

The formulation of Fused LASSO involves the following optimization problem:
x ^ = arg min i b i j a i j x j 2
s . t . j = 1 p | x j | μ 1 and j = 2 p | x j x j 1 | μ 2 ,
where μ 1 and μ 2 are tuning parameters that control the sparsity of the coefficients and the smoothness of their differences, respectively. This formulation captures both sparsity and smoothness, making it especially suited for high-dimensional settings where feature ordering is important. Examples of such problems include gene expression analysis and protein mass spectroscopy.
The optimization problem can be formulated as follows:
x = arg min x 1 2 A x b 2 2 + μ 1 x 1 + μ 2 D x 1 ,
where x R p , A R n × p , and b R n . The matrix D captures the differences between consecutive coefficients, and the 1 -penalty on D x promotes smoothness. We let f ( x ) = 1 2 A x b 2 , g ( x ) = μ 2 x 1 , and h ( D x ) = μ 1 D x 1 . The problem can also be reformulated in a primal-dual form for efficient computation:
min x R p max y R p 1 1 2 A x b 2 + μ 2 x 1 + D x , y μ 1 I B y μ 1
where B is the closed unit ball in the l norm and I B is its indicator function. We used the same problem setting as in [21].
We set the initial parameter values τ 0 = 0.95 / β = 0.95 / A T A and σ 0 = 1 / ( τ D T D ) , D T D 2 ( n ) . Figure 1 shows that a larger primal step size produces a better convergence rate for PPD3; thus, we chose τ close to its upper bound for the PPD3 method in this subsection. Since PPD3 has no restriction of τ σ 1 / D T D , we fixed the maximal primal step size at τ = 1 / β and chose a different dual step size σ so as to find the optimal dual step size (Figure 2). We see that removing the restriction on the dual step size, indeed, provides more choice regarding the step size parameters and has the potential to speed up algorithmic convergence. We chose σ = 1.1 / ( τ D T D ) for PPD3 to solve Fused LASSO, and the rest of the parameters are presented in Table 4. All τ values take their upper bound as the same as the experiment in [21], drawing the conclusion that a larger τ leads to faster convergence in Fused LASSO. θ was 1.5 for improved parameter conditions in AFBA/PDDY, as in [21] (Figure 3, Figure 4 and Figure 5).
We compared PPD3 with other methods for Fused LASSO problems of different dimensions. The results are consistent with the dimension of the problem, showing that PPD3 converges faster among all methods. PPD3 and AFBA/PDDY have close performance and are significantly better than CV or PD3O.

5.2. Image Inpainting

The classical Rudin-Osher-Fatemi (ROF) model, which minimizes total variation (TV), serves as a foundational approach to image restoration. A generic model for image restoration can be derived from the classical ROF model (in the discrete setting) as
min x X Φ x 1 + λ 2 x f 2 2 ,
where Φ is a linear transformation (e.g., wavelets and curvelets), and x represents the restored image. The first term enforces sparsity in the transformed domain, while the second term ensures the restored image is close to the observed image f, which is partially corrupted. λ controls the trade-off between sparsity and fidelity.
Image inpainting is the process of reconstructing missing or corrupted parts of an image. We consider a simple modification of (25). Let D = { ( i , j ) 1 i M , 1 j N } denote the set of indices of the image domain, and let I D denote the set of indices of the inpainting domain. The inpainting model can then be defined as
min x X Φ x 1 + λ 2 i , j D I ( x i , j f i , j ) 2 ,
This formulation minimizes the squared error between the reconstructed and observed image in non-missing regions. A large λ focuses solely on inpainting, while a small λ denoises the image. Here, we used a more compact form to present (26):
min l X u μ x 1 + 1 2 S x f 2 2 ,
where ∇ is the discrete gradient operator, S N × N is the mask operator (i.e., a diagonal matrix where zero entries denote missed information and identity entries indicate observed information), and f is the observed image. We let f ( x ) = 1 2 S x f 2 , g ( x ) = I [ l , u ] ( x ) , and h ( x ) = μ x 1 . By rewriting (27), the primal-dual form
min x max y 1 2 S x f 2 2 + I [ l , u ] ( x ) + x , y μ I B y μ
was obtained, which fits the form (2). This formulation balances sparsity, fidelity, and structure for effective image reconstruction.
The tested images are h o u s e . p n g ( 256 × 256 ) , l e n a . p n g ( 512 × 512 ) , and P e p p e r s . p n g ( 512 × 512 ) . The image of h o u s e . p n g ( 256 × 256 ) is degraded by a character mask, where about 15% of the pixels are missing. The degraded l e n a . p n g ( 512 × 512 ) retains one row of pixels in every eight rows, whereas the degraded Pepper.png (512 × 512) keeps one column of pixels in every eight columns, which means about 87% of the pixels are missing. For all images, a zero-mean Gaussian noise with a standard deviation of 0.02 was added. The original and degraded images are shown in Figure 6, Figure 7 and Figure 8.
Signal-to-Noise Ratio (SNR) is a measure used in science and engineering to quantify the level of a desired signal relative to the level of background noise. The formula for SNR in decibels (dB) is typically
SNR ( dB ) = 10 log 10 P signal P noise ,
where P signal is the power of the signal, and P noise is the power of the noise. In the realm of image processing, it measures the clarity of images, especially in applications like medical imaging or low-light photography. Higher SNR values indicate a clearer, stronger signal relative to noise.
We set the initial parameter values as τ 0 = 0.95 / β = 0.95 , σ 0 = 1 / ( τ T ) , and T 8 ( n ) . SNR was plotted against iteration number for the masked house image inpainting problems in Figure 9 and Figure 10. As shown below in Figure 9, the step size parameter is less sensitive in image inpainting than in Fused LASSO. Figure 10 indicates that a larger primal step size always leads to better performance; hence, we chose τ = 0.95 / S 2 , σ = 1.5 / τ 2 for PPD3, and we used the same settings listed in Table 5 for the rest of the parameters.
The recovered images are shown in Figure 11, Figure 12 and Figure 13, and the SNR was plotted against iteration number for all image inpainting in Figure 14.
The result in Figure 14 shows that in all three image inpainting missions, PPD3 is the best among the compared methods. The PPD3 and AFBA/PDDY methods are significantly better than the CV or PD3O methods.

5.3. Other Computational Advantages of PPD3

Beyond its rapid convergence in terms of iterations, the PPD3 method also demonstrates considerable computational efficiency in terms of time. By taking the image inpainting tasks of house.png ( 256 × 256 ) and pepper.png ( 512 × 512 ) as examples, the primary CPU time costs are detailed in Table 6, Table 7, Table 8 and Table 9. In particular, the parallel implementation of PPD3 achieves theoretical time savings of up to 46.6 % and 47.8 % , respectively, when communication overhead is ignored.
Furthermore, alternative methods typically require the computation of A T A , an operation for which the time cost is non-trivial. Although the linear operators in both experiments happen to have theoretical values, this is not guaranteed in all situations. We report the computational times of various MATLAB commands for A T A under scenarios where A is either dense or sparse. If A T A is unknown, we observe that even the fastest estimation of A T A accounts for approximately 10 % of the total computation time. These results underscore that PPD3 is particularly advantageous as a solver, especially when the linear operator A is dense and of significant size. In practice, PPD3 can take τ to its upper bound and choose a relatively large σ or a random one to avoid the calculation of A T A .

6. Discussion

This paper introduces the Parallel Primal-Dual (PPD3) algorithm, a novel framework for solving optimization problems involving the minimization of the sum of three convex functions, including a Lipschitz continuous term. By leveraging parallel updates of primal and dual variables and removing stringent parameter dependencies, PPD3 addresses the computational inefficiencies inherent in existing methods. The algorithm’s adaptability to various problem settings and its ability to perform without reliance on the spectral norm of the linear operator mark notable progressions in the field of convex optimization.
In our algorithm, parallelization has the potential to accelerate computation, particularly when running on parallel computing platforms. The constraint τ σ 1 A T A on the parameters is removed, eliminating the need to compute the spectral norm of A, which is essential for most primal-dual methods. This also broadens the choice of parameters. Additionally, the step size is adjusted dynamically at each iteration rather than using a fixed global upper limit. As a result, our step size is typically larger, leading to faster convergence.
Through applications in image inpainting and Fused LASSO, PPD3 has demonstrated satisfactory performance in terms of computational efficiency and convergence speed. These results underscore its potential to handle complex, large-scale optimization challenges across a variety of domains.
For future work, we could explore extending the PPD3 framework to non-convex settings. Additionally, we may develop a stochastic version of PPD3. Furthermore, the parameters could be further investigated to determine whether they can be relaxed.

Author Contributions

Conceptualization, X.Z.; methodology, X.Z.; software, S.Z. and J.W.; validation, X.Z. and K.Z.; formal analysis, X.Z. and W.T.; writing—original draft preparation, X.Z. and W.T.; writing—review and editing, X.Z. and K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated or analyzed during this study are available from the corresponding author.

Acknowledgments

The authors would like to thank Yuan Shen and Wenxing Zhang for their valuable advice, especially in the numerical experiment.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ROFRudin-Osher-Fatemi
CVCondat and Vu
PDFPPrimal-Dual Fixed-Point
PD3OPrimal-Dual Three-Operator splitting
AFBAAsymmetric Forward-Backward-Adjoint
PDDYDavisYin’s 82 three-operator splitting
PDHGPrimal-dual hybrid gradient
CPChambolle and Pock
APPDAdaptive parallel primal-dual
PPD3Parallel Primal-Dual

References

  1. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef]
  2. Candes, E.J.; Plan, Y. Matrix Completion With Noise. Proc. IEEE 2010, 98, 925–936. [Google Scholar] [CrossRef]
  3. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  4. Tseng, P. A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control Optim. 2000, 38, 431–446. [Google Scholar] [CrossRef]
  5. Passty, G.B. Ergodic convergence to a zero of the sum of monotone operators in Hilbert space. J. Math. Anal. Appl. 1979, 72, 383–390. [Google Scholar] [CrossRef]
  6. Lions, P.-L.; Mercier, B. Splitting algorithms for the sum of two nonlinear operators. SIAM J. Numer. Anal. 1979, 16, 964–979. [Google Scholar] [CrossRef]
  7. Zhu, M.; Chan, T.F. An Efficient Primal-Dual Hybrid Gradient Algorithm for Total Variation Image Restoration; CAM Report 08-34; UCLA: Los Angeles, CA, USA, 2008. [Google Scholar]
  8. Chambolle, A.; Pock, T. A first-order primal-dual algorithm for convex problems with applications to imaging. Math. Imaging Vis. 2011, 40, 120–145. [Google Scholar] [CrossRef]
  9. He, B.; Yuan, X. Convergence analysis of primal-dual algorithms for a saddle-point problem: From contraction perspective. SIAM J. Imaging Sci. 2012, 5, 119–149. [Google Scholar] [CrossRef]
  10. Goldstein, T.; Li, M.; Yuan, X.; Esser, E.; Baraniuk, R. Adaptive primal-dual hybrid gradient methods for saddle-point problems. arXiv 2015, arXiv:1305.0546v2. [Google Scholar]
  11. O’Connor, D.; Vandenberghe, L. Primal-dual decomposition by operator splitting and applications to image deblurring. SIAM J. Imaging Sci. 2014, 7, 1724–1754. [Google Scholar] [CrossRef]
  12. Zhang, X. Adaptive parallel primal-dual method for saddle point problems. Numer. Math. Theory, Method Appl. 2018, 11, 187–210. [Google Scholar]
  13. Condat, L. A primal-dual splitting method for convex optimization involving Lipschitzian, proximable and linear composite terms. J. Optim. Theory Appl. 2013, 158, 460–479. [Google Scholar] [CrossRef]
  14. Vũ, B.C. A splitting algorithm for dual monotone inclusions involving cocoercive operators. Adv. Comput. Math. 2013, 38, 667–681. [Google Scholar] [CrossRef]
  15. Latafat, P.; Patrinos, P. Asymmetric forward-backward-adjoint splitting for solving monotone inclusions involving three operators. arXiv 2016, arXiv:1602.08729. [Google Scholar] [CrossRef]
  16. Chen, P.; Huang, J.; Zhang, X. A primal-dual fixed point algorithm for minimization of the sum of three convex separable functions. Fixed Point Theory Appl. 2016, 54, 1–18. [Google Scholar] [CrossRef]
  17. Yan, M. A Primal-Dual Three-Operator Splitting Scheme. 2017. Available online: https://optimization-online.org/?p=14319 (accessed on 26 January 2025).
  18. Davis, D.; Yin, W. Convergence rate analysis of several splitting schemes. arXiv 2014, arXiv:1406.4834. [Google Scholar]
  19. Xu, C.Y.H.; Yang, J. A Modified Primal-Dual algorithm for Structured Convex Optimization with a Lipschitzian Term. J. Oper. Res. China 2024, 1–21. [Google Scholar]
  20. He, B.; Yuan, X. Balanced augmented Lagrangian method for convex programming. arXiv 2021, arXiv:2108.08554. [Google Scholar]
  21. Yan, M.; Li, Y. On the improved Conditions for Some Primal-Dual Algorithms. J. Sci. Comput. 2024, 99, 74. [Google Scholar] [CrossRef]
  22. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via admm. Found. Trends Machine Learn. 2010, 3, 1–122. [Google Scholar] [CrossRef]
  23. He, B.; Yuan, X.; Zhang, J.J.Z. Comparison of two kinds of prediction-correction methods for monotone variational inequalities. Comput. Optim. Appl. 2004, 27, 247–267. [Google Scholar] [CrossRef]
  24. Facchinei, F.; Pang, J.-S. Finite-Dimensional Variational Inequalities and Complementarity Problems; Springer Series in Operations Research; Springer: New York, NY, USA, 2003; Volume I. [Google Scholar]
  25. He, B.; Yuan, X. On the O(1/n) convergence rate of the Douglas-Rachford alternating direction method. SIAM J. Numer. Anal. 2012, 50, 700–709. [Google Scholar] [CrossRef]
Figure 1. Fused LASSO with various primal step sizes. (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Figure 1. Fused LASSO with various primal step sizes. (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Axioms 14 00104 g001
Figure 2. Fused LASSO with various dual step sizes. (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Figure 2. Fused LASSO with various dual step sizes. (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Axioms 14 00104 g002
Figure 3. Sample matrix A R 100 × 2000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Figure 3. Sample matrix A R 100 × 2000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Axioms 14 00104 g003
Figure 4. Sample matrix A R 500 × 10 , 000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Figure 4. Sample matrix A R 500 × 10 , 000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Axioms 14 00104 g004
Figure 5. Sample matrix A R 1000 × 20 , 000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Figure 5. Sample matrix A R 1000 × 20 , 000 . (a) relative objective error vs. iteration number. (b) relative variable error vs. iteration number.
Axioms 14 00104 g005
Figure 6. House original (a). House masked (b).
Figure 6. House original (a). House masked (b).
Axioms 14 00104 g006
Figure 7. Lena original (a). Lena masked (b).
Figure 7. Lena original (a). Lena masked (b).
Axioms 14 00104 g007
Figure 8. Pepper original (a). Pepper masked (b).
Figure 8. Pepper original (a). Pepper masked (b).
Axioms 14 00104 g008
Figure 9. Inpainting “house.png”, with various dual stepsize. (a) stepsize with large range. (b) stepsize with narrow range.
Figure 9. Inpainting “house.png”, with various dual stepsize. (a) stepsize with large range. (b) stepsize with narrow range.
Axioms 14 00104 g009
Figure 10. Inpainting “house.png”, with various primal stepsize. (a) fixed dual stepsize. (b) fixed primal-dual stepsize ratio.
Figure 10. Inpainting “house.png”, with various primal stepsize. (a) fixed dual stepsize. (b) fixed primal-dual stepsize ratio.
Axioms 14 00104 g010
Figure 11. House recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Figure 11. House recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Axioms 14 00104 g011
Figure 12. Lena recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Figure 12. Lena recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Axioms 14 00104 g012
Figure 13. Pepper recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Figure 13. Pepper recovered by CV (a), PD3O (b), AFBA/PDDY (c), and PDD3 (d).
Axioms 14 00104 g013
Figure 14. SNR vs. iteration number for House (a), Lena (b), and Pepper (c).
Figure 14. SNR vs. iteration number for House (a), Lena (b), and Pepper (c).
Axioms 14 00104 g014
Table 1. Comparison of algorithms with respect to θ .
Table 1. Comparison of algorithms with respect to θ .
Algorithm θ Correction Step Required
PDHG/Arrow-Hurwicz0No
CP1No
HeYuan [ 1 , 1 ] Yes
APPD 1 Yes
Table 2. Comparison of algorithms with respect to primal variable upgrade.
Table 2. Comparison of algorithms with respect to primal variable upgrade.
AlgorithmPrimal Variable Upgrade
CV x ¯ k = 2 x k + 1 x k
PDFP x ¯ k = prox τ g ( x k + 1 τ f ( x k + 1 ) τ A y k )
PD3O x ¯ k = 2 x k + 1 x k + τ f ( x k ) τ f ( x k + 1 )
AFBA/PDDY x k + 1 = x ¯ k 1 τ A ( y k y k 1 ) ;
x ¯ k + 1 = prox τ g ( x k + 1 τ f ( x k + 1 ) τ A y k )
Table 3. Comparison of algorithms with respect to parameter restrictions.
Table 3. Comparison of algorithms with respect to parameter restrictions.
AlgorithmParameter Restrictions
CV τ β / 2 + τ σ A T A 1
PDFP τ 2 β , τ σ 1 A T A
PD3O τ 2 β , τ σ 1 A T A
AFBA/PDDY τ 2 β , τ σ 1 A T A
Table 4. Parameter settings for different algorithms.
Table 4. Parameter settings for different algorithms.
AlgorithmPrimal Stepsize ( τ )Dual Stepsize ( σ ) σ · τ
CV 1.9 A T A 0.095 τ D T D 0.095 · D T D
PD3O 1.95 A T A 1 τ D T D 1 · D T D
AFBA/PDDY 3.8 A T A 2 3 τ D T D 2 3 · D T D
PPD3 0.95 A T A 1.1 τ D T D 1.1 · D T D
Table 5. Parameter settings for image inpainting.
Table 5. Parameter settings for image inpainting.
AlgorithmPrimal Stepsize ( τ )Dual Stepsize ( σ ) σ · τ
CV 1.9 S T S 0.095 τ T 0.095 · T
PD3O 1.95 S T S 1 τ T 1 · T
AFBA/PDDY 3.8 S T S 2 3 τ T 2 3 · T
PPD3 0.95 S T S 1.5 τ T 1.5 · T
Table 6. CPU time for Pepper.png.
Table 6. CPU time for Pepper.png.
AlgorithmTotal X-Update TimeTotal Y-Update Time
CV1.2911.622
PD3O1.2831.62
PDDY1.4351.63
PPD31.475 (47.8% )1.611
Table 7. CPU time for Pepper.png.
Table 7. CPU time for Pepper.png.
Method to Get Norm ( A T A )Time CostValue of Norm ( A T A )
Norm(.)full≥10007.9999
Normest(.)full≥10007.8816
Svds(.)sparse10.9895527.9999
Normest(.)sparse0.2779607.8816
Table 8. CPU time for House.png.
Table 8. CPU time for House.png.
AlgorithmTotal X-Update TimeTotal Y-Update Time
CV0.4000.557
PD3O0.4190.571
PDDY0.4870.609
PPD30.539 (46.6% )0.618
Table 9. CPU Time for House.png.
Table 9. CPU Time for House.png.
Method to get Norm ( A T A )Time CostValue of Norm ( A T A )
Norm(.)full≥10007.9997
Normest(.)full235.87557.8816
Svds(.)sparse2.1854567.9997
Normest(.)sparse0.0978477.8816
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Tang, W.; Wang, J.; Zhang, S.; Zhang, K. Parallel Primal-Dual Method with Linearization for Structured Convex Optimization. Axioms 2025, 14, 104. https://doi.org/10.3390/axioms14020104

AMA Style

Zhang X, Tang W, Wang J, Zhang S, Zhang K. Parallel Primal-Dual Method with Linearization for Structured Convex Optimization. Axioms. 2025; 14(2):104. https://doi.org/10.3390/axioms14020104

Chicago/Turabian Style

Zhang, Xiayang, Weiye Tang, Jiayue Wang, Shiyu Zhang, and Kangqun Zhang. 2025. "Parallel Primal-Dual Method with Linearization for Structured Convex Optimization" Axioms 14, no. 2: 104. https://doi.org/10.3390/axioms14020104

APA Style

Zhang, X., Tang, W., Wang, J., Zhang, S., & Zhang, K. (2025). Parallel Primal-Dual Method with Linearization for Structured Convex Optimization. Axioms, 14(2), 104. https://doi.org/10.3390/axioms14020104

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop