Next Article in Journal / Special Issue
Outlier Detection Based on Residual Histogram Preference for Geometric Multi-Model Fitting
Previous Article in Journal
MVDT-SI: A Multi-View Double-Triangle Algorithm for Star Identification
Previous Article in Special Issue
Deep Learning Based Switching Filter for Impulsive Noise Removal in Color Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Orthonormalization-Free Approach for Sparse Dictionary Learning and Dual Principal Component Pursuit

1
Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing 100190, China
2
School of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing 100049, China
3
State Key Laboratory of Scientific and Engineering Computing, Beijing 100190, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(11), 3041; https://doi.org/10.3390/s20113041
Submission received: 10 May 2020 / Revised: 23 May 2020 / Accepted: 25 May 2020 / Published: 27 May 2020
(This article belongs to the Special Issue Data, Signal and Image Processing and Applications in Sensors)

Abstract

:
Sparse dictionary learning (SDL) is a classic representation learning method and has been widely used in data analysis. Recently, the m -norm ( m 3 , m N ) maximization has been proposed to solve SDL, which reshapes the problem to an optimization problem with orthogonality constraints. In this paper, we first propose an m -norm maximization model for solving dual principal component pursuit (DPCP) based on the similarities between DPCP and SDL. Then, we propose a smooth unconstrained exact penalty model and show its equivalence with the m -norm maximization model. Based on our penalty model, we develop an efficient first-order algorithm for solving our penalty model (PenNMF) and show its global convergence. Extensive experiments illustrate the high efficiency of PenNMF when compared with the other state-of-the-art algorithms on solving the m -norm maximization with orthogonality constraints.

1. Introduction

In this paper, we focus on solving the optimization problem with orthogonality constraints:
min W R n × p f ( W ) : = 1 m W Y m m s . t . W W = I p ,
where W is the variable, Y R n × N is a given data matrix, and  I p denotes the identity matrix in  R p × p . Besides, the  m -norm is defined as Y m = i = 1 n j = 1 N ( Y i j ) m 1 m with constant m ( 2 , 4 ] . For brevity, the orthogonality constraints W W = I p in (1) can be expressed as W S n , p : = { W R n × p | W W = I p } . Here, S n , p denotes the Stiefel manifold in real matrix space, and we call it the Stiefel manifold for simplicity in the rest of our paper.
The sparse dictionary learning (SDL) exploits the low-dimensional features within a set of unlabeled data, and therefore plays an important role in unsupervised representative learning. More specifically, given a data set Y = [ y 1 , y 2 , , y N ] R n × N that contains N samples in  R n , SDL aims to compute a full-rank matrix D R n × p named as dictionary, and an associated sparse representation X = [ x 1 , , x N ] that satisfies
Y = D X ,
or equivalently, find a  W = D 1 such that
X = W Y .
As a result, the SDL can be solved by finding a  W R n × n , which leads to a sparse W Y . Some existing works introduce the  0 -norm or 1 -norm penalty term to promote the underlying sparsity of W Y and present various algorithms for solving the consequent optimization models, see the work in [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] for instance. Interested readers are referred to a recent paper [16] and the references therein. However, the  1 -norm minimization-based models are known to be sensitive with noise, and so far the existing approaches are not efficient enough for the purpose of solving real application problems which are often large-scale [17]. Consequently, a proper model with an efficient algorithm for SDL is desired, especially for the large-scale case.
Recently, an 4 -norm maximization model is proposed in [17], which can recover the entire dictionary in a single run. This new formulation is motivated by the fact that maximizing a higher-order norm promotes spikiness and sparsity at the same time. The authors of [17] demonstrate that the global minimizers of 4 -norm maximization with orthogonality constraints are very close to the true dictionary. Moreover, concaveness of the objective function in Equation (1) enables a fast fixed-point type algorithm, named matching, stretching, and projection (MSP). MSP achieves significant speedup compared with existing methods based on 0 -norm or 1 -norm penalty minimization. As maximizing any higher-order norm over a lower-order norm constraint leads to sparse and spiky solutions, Shen et al. [18] extend 4 -norm maximization technique to a generalized l m -norm maximization ( m 3 ). In addition, the authors propose a gradient projection method (GPM) for solving it with guaranteed global convergence.
However, both MSP and GPM invoke polar decomposition to keep the feasibility in each iteration. As illustrated in  [19,20,21], orthonormalization lacks concurrency, which results in low scalability in column-wise parallel computing, particularly when the number of columns is large.
Several infeasible approaches have been developed to avoid orthonormalization. Gao et al. [19] propose the proximal linearized augmented Lagrangian method (PLAM) as well as its enhanced version, PCAL. Based on the merit function used in Gao et al. [19], Xiao et al. [21] propose an exact penalty model with a convex and compact auxiliary constraint, named PenC, for optimization problems with orthogonality constraints. The authors propose an approximated gradient method named PenCF for solving PenC and showed its global convergence and local convergence rate under mild conditions. The above-mentioned infeasible approaches do not require orthonormalization in each iteration. Numerical experiments illustrate the promising performance of these infeasible approaches with the existing state-of-the-art algorithms.
Although PCAL and PenCF avoid the orthonormalization process by taking infeasible steps, these approaches require additional constraints to restrict the sequence in a compact set in  R n × p , which can undermine their overall efficiency. Therefore, to develop an efficient algorithm on solving SDL, an infeasible model without constraints is desired.
Similar to the  1 -norm penalty model for SDL, dual principal component pursuit (DPCP) aims to recover a tangent vector in  R n from samples Y = [ y 1 , , y n ] R n × N contaminated by outliers. Specifically, DPCP solves the following nonsmooth nonconvex optimization problem with a spherical constraint:
min W R n W Y 1 s . t . W W = 1 .
Due to the ability on recovering an  n 1 dimensional hyperplane from R n , DPCP has wide applications in 3D computer vision, such as detecting planar structures in 3D point clouds in KITTI dataset [22,23] and estimating relative poses in multiple-view [24].
Existing approaches [25,26,27,28] for solving convex problem (4) are not scalable and not competent in high dimensional cases [29]. On the other hand, the Random Sampling and Consensus (RANSAC) algorithm [30] has been one of the most popular methods in computer vision for the high relative dimension setting. RANSAC alternates between fitting a subspace to a randomly sampled minimal number of points ( n 1 in the case of DPCP) and measuring the quality of selected subspace by using the number of data-points close to the subspace. In particular, as described in [29], RANSAC can be extremely effective when the probability of sampling outlier-free samples inside the allocated time budget is large. Recently, Tsakiris and Vidal [31] introduce Denoised-DPCP (DPCP-d) by minimizing y W Y F 2 + γ y 1 over the constraints W W = 1 , y R N . In the same paper, Tsakiris and Vidal [31] propose an Iteratively-Reweighted-Least-Squares algorithm (DPCP-IRLS) for solving the non-convex DPCP problem (4). The authors illustrate that DPCP-IRLS can successfully handle 30 % to 50 % of outliers and showed its high efficiency compared with RANSAC. In addition, Zhu et al. [32] propose a projected subgradient-based algorithm named DPCP-PSGM, which exhibits great efficiency on reconstructing road-plane in the KITTA dataset. There are also some approaches using smoothing techniques to approximate the  1 -norm term such as Logcosh [8,33], Huber loss [34], pseudo-Huber [5], etc. Then, algorithms for minimizing a smooth objective function on a sphere can be applied. Nonetheless, these smoothing techniques often introduce approximation errors as the smooth objective functions usually lead to dense solutions. Qu et al. [35] and Sun et al. [8] propose a rounding step as postprocessing to achieve exact recovery [16] by solving a linear programming, which leads to addition computational cost.
The main difficulties in developing efficient algorithms are the nonsmoothness and nonconvexity in DPCP models. By observing the similarity between SDL and DPCP, we consider to adopt the  m -norm maximization to reformulate DPCP as a smooth optimization problem on sphere.

1.1. Contribution

In this paper, we first point out that the DPCP problem can be formulated as the m -norm ( m ( 2 , 4 ] ) maximization (1) with p = 1 . Therefore, both SDL and DPCP can be unified as a smooth optimization problem on the Stiefel manifold.
Motivated by PenC [21], we propose a novel penalty function as the following expression,
h ( W ) : = f ( W ) 1 2 W W I p , Φ ( W f ( W ) ) + β 6 W F 6 β 2 W F 2 ,
where β > 0 is the penalty-parameter and Φ is the operator that symmetrizes the square matrix, defined by Φ ( M ) = M + M 2 . We show that h ( W ) is bounded from below, then the convex compact constraint in PenC can be avoided. Therefore, we propose the following smooth unconstrained penalty model for m -norm maximization (PenNM),
min W R n × p h ( W ) .
We prove that Equation (6) with m ( 2 , 4 ] is an exact penalty function of Equation (1) under some mild conditions. Moreover, when p = 1 , we verify that PenNM does not introduce any first-order stationary point other than those of Equation (1) and x = 0 . Based on the new exact penalty model, we propose an efficient orthonormalization-free first-order algorithm named PenNMF with no additional constraint. In PenNMF, we adopt an approximate gradient in each iterate instead of the exact one in which the second-order derivative of the original objective involves. The global convergence of PenNMF under mild conditions can be established.
The numerical experiments on synthetic and real imaginary data demonstrate that PenNMF outperforms PenCF and MSP/GPM in solving SDL, especially in large-scale cases. As an infeasible method, PenNMF shows superior performance when compared with MSP and GPM, which invoke an orthonormalization process to keep the feasibility. Moreover, when compared with PenCF, PenNMF also shows better performance, implying the benefits of avoiding the constraints in PenC. In our numerical experiments on DPCP, our proposed model (1) with p = 1 shows comparable accuracy with 1 -norm based penalty model (4) on solving road-plane recovery in KITTA dataset. In some test examples, (1) can have even better accuracy than (4). Besides, PenNMF takes less CPU time while achieving comparable accuracy in reconstructing road-plane in KITTA dataset when compared with other state-of-the-art algorithms such as DPCP-PSGM and DPCP-d.

1.2. Notations and Terminologies

Norms: In this paper, · m denotes the element-wise m-th norm of a vector or matrix, i.e., A m = i = 1 j = 1 | A i j | m 1 / m . Besides, · F denotes the Frobenius norm and · denotes the 2-th operator norm, i.e., A equals the maximum singular value of A. Besides, we denote σ min ( A ) as the smallest singular value of a given matrix A. The operator A B stands for the Hadamard product of matrices A and B with the same size. | A | and A l represent the component-wise absolute value and l-th power of matrix A, respectively. Besides, for two symmetric matrices A and B, A B denotes that A B is semi-positive definite, and  A B denotes that A B is positive definite.
Optimality Condition:W is a first-order stationary point of (1) if and only if 
( I n W W ) f ( W ) = 0 ; W f ( W ) = f ( W ) W ; W W = I p .
Besides, W is a first-order stationary point of PenNM if and only if  h ( W ) = 0 .

2. Model Description

In this section, we first discuss how to reformulate DPCP as an  m -norm maximization with orthognoality constraints. To construct a orthonormalization-free algorithm, we minimize h ( W ) rather than directly solve (1). As an unconstrained penalty problem for (1), the model (6) may introduce additional infeasible first-order stationary points. Therefore, in this section, we characterize the equivalence between (1) and (6) to provide theoretical guarantees for our approach.

2.1. m -Norm Maximization for DPCP Problems

Based on the fact that maximization of a higher-order norm promotes spikiness and sparsity, we maximize the  m -norm of W ^ Y over the constraint W ^ Y Y W ^ = 1 . The model can be expressed as
min W ^ R n 1 m W ^ Y m m s . t . W ^ Y Y W ^ = 1 .
Although with different constraints to (1), (4) can be reshaped to the formulation of (1). Let  Y = R Z be the rank-revealing QR decomposition of Y, where Z R n × N is an orthogonal matrix and R R n × n is an upper-triangular matrix, and denote W = R T W ^ , then the optimization model can be reshaped as
min W R n 1 m W Z m m s . t . W W = 1 .
Clearly, problem (8) is a special case of (1) with p = 1 . Moreover, suppose W * is a global minimizer of (8), the solution for DPCP problem can be recovered by W ^ * = R T W * . The detailed framework for solving DPCP by m -norm maximization is presented in Algorithm 1.
Algorithm 1 Framework for Solving DPCP by m -Norm Maximization.
Require: Data matrix Y R n × N
 1: Perform QR-factorization for Y. Namely, Y = R Z where R is upper-triangular matrix and Z R n × N is orthogonal matrix;
 2: Compute the solution W ^ for (1);
 3: Return W * = R T W ^

2.2. Equivalence

In this subsection, we first derive the expression for f ( W ) and h ( W ) .
Proposition 1.
The gradient and the Hessian of f ( W ) can be expressed as
f ( W ) = Y | Y W | ( m 1 ) sign ( ( Y W ) ) ; 2 f ( W ) [ D ] = ( m 1 ) Y ( Y W ) ( m 2 ) ( Y D ) ,
respectively. Moreover, the gradient of h ( W ) can be formulated as
h ( W ) = f ( W ) 3 2 I p 1 2 W W W Φ ( W f ( W ) ) 1 2 2 f ( W ) [ W ( W W I p ) ] + 2 β W W W ( W W I p ) + β W ( W W W W I p ) .
Proof. 
From the work in [17] we have f ( W ) = Y | Y W | ( m 1 ) s i g n ( Y W ) . Based on the expression for f ( W ) , the Hessian of f can be expressed as 2 f ( W ) [ D ] = ( m 1 ) Y ( Y D ) ( Y W ) ( m 2 ) . As a result, 2 f ( W ) [ W ( W W I p ) ] = ( m 1 ) Y ( Y W ) ( m 1 ) Y ( W W I p ) .
Therefore, based on ([21], Equation 2.8), the gradient of h ( W ) can be formulated as
h ( W ) = f ( W ) 3 2 I p 1 2 W W W Φ ( W f ( W ) ) 1 2 2 f ( W ) [ W ( W W I p ) ] + 2 β W W W ( W W I p ) + β W ( W W W W I p )
 □
With the expression for h ( W ) , we can establish the equivalence between (1) and our proposed model, (6). The equivalence is illustrated in Theorem 4, and the main body of the proofs is presented in Appendix A.
Theorem 2.
(First-order equivalence)Suppose β ( 4 m + 8 ) Y F m and W ˜ is a first-order stationary point of (6), then either W ˜ W ˜ = I p holds, which further implies that W ˜ is a first-order stationary point of problem (1), or the inequality σ min ( W ˜ W ˜ ) ( 2 m + 4 ) Y F m β holds.
Theorem 2 characterizes the relationship between the first-order stationary points of (1) and those of (6). Namely, the penalty model only yields the first-order stationary points other than those of the original model (1) far away from the Stiefel manifold. When p = 1 , we can derive a stronger result on those additional first-order stationary points produced by the penalty model in Corollary 3.
Corollary 3. 
(Stronger first-order equivalence for p = 1 )Suppose p = 1 in (1), β ( 4 m + 8 ) Y F m , and W ˜ is a first-order stationary point of (6), then either W ˜ W ˜ = I p holds, which further implies that W ˜ is a first-order stationary point of problem (1), or  W ˜ = 0 .
Theorem 2 characterizes the equivalence between (1) and (6) in the sense that all the infeasible first-order stationary points of (6) is relatively far away from the constraint W W = I p . Besides, Corollary 3 shows that when p = 1 , the only infeasible first-order stationary point of (6) is 0. Therefore, when we achieve a solution near the constraint W W = I p by solving (1), we can conclude that W is a first-order stationary point of (1). Instead of directly solving (1), we can compute the first-order stationary point of (6) and thus avoid intensive orthonormalization in the computation.

3. Algorithm

3.1. Global Convergence

In this section, we focus on developing an infeasible approach for solving (6). The calculation of the gradient of h ( W ) is involved with the second-order derivative, which is typically even more expensive than the iterations in MSP/GPM. Therefore, we consider to solve (6) by an approximated gradient descent algorithm. Let D ( W ) : = f ( W ) W Φ ( W f ( W ) ) + β W ( W W W W I p ) be the approximation for the gradient of h ( W ) , we present the detailed algorithm as Algorithm 2.
Algorithm 2 First-Order Method for Solving (6). (PenNMF)
Require: f : R n × p R , β > 0 ;
 1: Randomly choose W 0 satisfies W 0 W 0 = I p , set k = 0 ;
 2: while not terminate do
 3:  Compute inexact gradient
D ( W k ) = f ( W k ) W k Φ ( W k f ( W k ) ) + β W k ( W k W k W k W k I p ) ;
 4:  Compute stepsize η k ;
 5:   W k + 1 = W k η k D ( W k ) ;
 6:   k + + ;
 7: end while
 8: Return W k
Next, we establish the convergence of PenNMFin Theorem 4, which illustrates the global convergence and worst-case convergence rate of PenNMF under mild conditions. The main body of the proof is presented in Appendix B.
Theorem 4.
(Global convergence)Suppose δ 0 , 1 3 and β max 228 m Y F m , 32 δ Y F m . Let { W k } be the iterate sequence generated by PenNMF, starting from any initial point W 0 satisfying | | W 0 W 0 I p | | F 2 1 8 δ , and the stepsize η k 1 2 η ¯ , η ¯ , where η ¯ 1 2 M 1 . Then, W k weakly converges to a first-order stationary point of (1). Moreover, for any k = 1 , 2 , , the convergence rate of PenNMF can be estimated by
min 0 i k D ( W i ) F 8 m Y F m + 2 β δ η ¯ ( k + 1 ) .

3.2. Some Practical Settings

As illustrated in Algorithm 2, the hyperparameters in PenNMF are the penalty parameter β and stepsize η k . In the theoretical analysis for PenNMF, the upper bound of η k adopted in Theorem 4 is too restrictive in practice. There are many adaptive stepsize for first-order algorithms, and here we consider the Barzilai–Borwein (BB) stepsize [36],
η B B 1 , k = S k , Y k Y k , Y k , η B B 2 , k = S k , S k S k , Y k ,
and alternating Barzilai–Borwein (ABB) stepsize [37],
η A B B k = η B B 1 k m o d ( k , 2 ) = 1 η B B 2 k m o d ( k , 2 ) = 0 ,
where and S k = W k W k 1 , Y k = h ( W k ) h ( W k 1 ) . We suggest to choose the stepsize η k as ABB stepsize in PenNMF, and we test PenNMF with ABB stepsize in our numerical experiments.
Another parameter is β , which controls the smooth penalty term in h ( W ) . Similarly, the lower-bound for β in Theorem 4 is too large to be practical. In our numerical examples, we uses the constant s : = f ( W 0 ) F , which is an approximation to 2 f ( W 0 ) F , as an upper-bound for β . According to the work in [21], we suggest to choose the penalty parameter by β = 0 . 01 2 f ( W 0 ) F .
Additionally, to achieve high accuracy in feasibility, we perform the polar factorization to the final solution generated by PenCF and PenNMF as the default postprocess. More precisely, when we compute the final solution W k by PenNMF, we can compute its rank-revealing singular-value decomposition W k = U k Σ k V k and return W ^ k : = U k V k . Using the same proof techniques in [21], our postprocess leads to decrease in feasibility as well as the functional value. Moreover, the numerical experiments in [19] show that the introduced orthonormalization process results in little changes in h ( W ) . Therefore, we suggest to perform the described postprocess for PenNMF.

4. Numerical Examples

In this section, we present our preliminary numerical examples. We compare our algorithm with some state-of-the-art algorithms on SDL and DPCP problems, which are formulated as (1) and (8), respectively. Then, we observe the performance of our algorithm under different selections of parameters, and then choose the default setting. All the numerical experiments in this section are tested on an Intel(R) Core(R) Silver 4110 CPU @ 2.1 GHz, with 32 cores and 394 GB of memory running under Ubuntu 18.04 and MATLAB R2018a.

4.1. Numerical Results on Sparse Dictionary Earning

In this subsection, we mainly compare the numerical performance of PenNMF with some state-of-the-art algorithms on SDL. As illustrated in Table 2 in [17], MSP is significantly faster than the Riemannian subgradient [3] and Riemannian trust-region method [8]. Therefore, to have a better illustration on the performance of PenNMF, we compare PenNMF with state-of-the-art algorithms on solving (1), which is a smooth optimization problem with orthogonality constraints. We first select two state-of-the-art algorithms on solving optimization problems with orthogonality constraints. One is Manopt [38,39], a projection-based feasible method. In our numerical test, we choose nonlinear conjugate gradient with inexact linear-search strategy to accelerate Manopt. Another one is PenCF [21], which is an infeasible approach for optimization problems with orthogonality constraints. In our algorithms we choose to apply Alternating Bzarzilar–Borwein stepsize to accelerate PenNMF, and uses all parameters as default setting described in [21]. Besides, we test the MSP algorithm [17] and GPM algorithm [18]. It is worth to mention that when m = 4 , the MSP and GPM are actually the same. According to the numerical examples in [18], m = 3 has better recovery quality than the case m = 4 . Therefore, in our numerical experiments, we test the mentioned algorithms on the case where m = 3 .
The stopping criteria for Manopt, MSP/GPM is f ( W k ) W k Λ ( W k ) F 10 2 , while the stopping criteria for PenCF and PenNMF is h ( W k ) F 10 2 . Besides, the max iteration for all compared algorithms is set as 200.
In all test examples, we randomly generate the sparse representation X by X * = randn ( n , N ) . * ( randn ( n , N ) < 0.3 ) and the dictionary W * by randomly selecting a point on Stiefel manifold. Then, the original data matrix Y is constructed by Y o = W * X * . To test the performance of all compared algorithms, we add different types of noise to Y o . We first fix the level of noise θ = 0 . 3 and choose n from 20 to 100. Then, we test the performance of compared algorithms with different types of noisy while fix n = 50 . In our numerical tests, the “Noise” denotes the Gaussian Noise, where Y is constructed by Y = Y o + θ · randn ( n , N ) . Besides, the term “Outliers” denotes the Gaussian outliers, where outliers = randn ( n , round ( θ m ) ) ; Y = cat ( 2 , Y o , outliers ) ; . Additionally, the term “Corruption” refers to the Gaussian corruption to Y o , which is achieved by rademacher = ( rand ( n , m ) < 0.5 ) * 2 1 ; Y = Y o + ( rand ( n , m ) < θ ) . * rademacher . Besides, the term ’CPU time’ denotes the averaged run-time, while the term ’Error’ denotes the 1 W ^ W * 4 4 , where W ^ denotes the final output of all the compared algorithm.
The numerical results are listed in Figure 1. From Figure 1d–f, j–l we conclude that all these compared algorithms achieve almost the same accuracy in all the cases. Besides, for Gaussian noise, the performance of PenNMF is comparable to MSP/GPM algorithm and outperforms Manopt. Moreover, with Gasuuain outliers and Gaussian corruption, the performance of PenNMF is better than PenCF, MSP/GPM, and Manopt. One possible explanation is that for Manopt invokes computing the Riemannian gradient, line-search in each iteration, resulting in higher computational complexity than MSP/GPM. Besides, the infeasible approaches overcome the bottleneck in the orthonormalization process in Manopt and MSP/GPM, and thus achieve comparable performance to MSP/GPM. Additionally, PenCF solves a constrained model by taking approximated gradient descent steps, while in PenNMF the model is an unconstrained one. The absence of constraint helps to improve the performance of PenNMF.
Besides testing on synthetic datasets, we also perform extensive experiments to verify the performance of PenNMF on real imagery data. A classic application of dictionary learning involves learning sparse representations of image patches [40]. In this paper, we extend the experiments in [17] to learn patches from grayscale and color images. Based on the 512 × 512 grayscale image “Barbara”, we construct the clean data matrix Y o by vectorizing each 16 × 16 patches from it. Then, we use the same approach to construct the clean data matrix Y from 512 × 512 grayscale images “Boat” and “Lena”, together with a 256 × 256 grayscale image ”House”. In “Barbara”, “Boat”, and “Lena”, the clean data matrix Y R 256 × 247 , 009 , and the data matrix from “House” satisfies Y R 256 × 58 , 081 . Besides, we construct the matrix Y R 192 × 62 , 001 by vectorizing the 8 × 8 × 3 patches from the 256 × 256 RGB image “Duck”. In such setting, all the compared algorithms recover the dictionary for all three channels simultaneously rather than learn them once for each channel in “Duck”. Such approach is aslo applied to generate the data matrix in R 192 × 146 , 633 from 338 × 450 RGB image “Chateau”. We run MSP/GPM, PenNMF, PenCF, and Manopt with m = 3 to compute the dictionary from Y = Y o + θ · randn ( n , N ) with different level of noise, where Y o is generated in the same manner as our first numerical experiment and has the same size as these patched figures. The numerical results are presented in Figure 2 and Figure A1. In all experiments, PenNMF takes less time than PenCF, MSP/GPM, and Manopt, which further illustrate the high efficiency of PenNMF in tackling the real imagery data, especially in the large-scale case.

4.2. Dual Principal Component Pursuit

In this subsection, we first verify the recovery property of our proposed model (8), which is a special case of (1) by fixing p = 1 . We first compare the distance between global minimizer of (8) and the ground-truth for DPCP problem. We first fix n = 30 and randomly select W * R n . Then, we randomly generate N 1 inliers in the hyperplane whose normal vector is W * . Besides, we randomly generate N 2 outliers in R n following Gaussian distribution. Additionally, the data is corrupted by Gaussian noise by adding θ n · randn ( n , N ) to Y. Then, we normalize each sample in Y. The range of N 1 is [ 10 , 500 ] , whereas the range of N 2 is [ 10 , 3000 ] . We run each test problem for 5 instances. Moreover, in each instance, we run DPCP-PSGM to solve (4) and PenNMF to solve (8) with m = 3 and 4, and get the solution W ˜ for each model. We plot the principal angle between W ˜ and W * in Figure 3. From Figure 3a,b we can conclude that (4) can tolerate O ( N 1 2 ) outliers while achieve exact recovery, which coincides the theoretical results presented in [32]. For model (8), numerical experiments do not show the exact recovery ability of (8) for m = 3 and 4. However, with some tolerance on the principal angle, we also observe that (4) can tolerate O ( N 1 2 ) outliers. Moreover, we conclude that with m = 3 , (8) has better ability to recover the normal vector than m = 4 . As a result, in the rest of this subsection, we only test (8) with m = 3 . In addition, we analyze the number of successfully recovered instances, where the 1 W ˜ , W * 2 is less than 0.1 or 0.2 . The results are presented in Figure 4. From Figure 4, we can conclude that, with tolerance on the errors, the m -norm maximization model can successfully recover the normal vector. Moreover, in model (8), m = 3 has better performance than m = 4 , which coincides with the numerical experiments in [18]. Therefore, when applying m -norm maximization model to solving the DPCP problems, we suggest to choose m = 3 in (8).
In the rest of this subsection, we test the numerical performance of PenNMF on solving DPCP problem, which plays an important role in autonomous driving applications. DPCP is applied to recover the road-plane, which can be regarded as inliers, from the 3d point clouds in KITTA dataset [22], which is recorded from a moving platform while driving in and around Karlsruhe, Germany. This dataset consists of image data together with corresponding 3D points collected by a rotating 3D laser scanner [32]. Moreover, DPCP only uses the 3D point clouds with the objective of determining the 3D points that lie on the road plane (inliers) and those off that plane (outliers): Given a 3D point cloud of a road scene, the DPCP problem focuses on reconstructing an affine plane { x R 3 | a x b = 0 } as a representation for the road. Equivalently, this task can be converted to a linear subspace learning problem by embedding the affine plane into the linear hyperplane H R 4 with normal vector b ˜ = [ a , b ] , through the mapping x [ x , 1 ] [29]. We use the experimental set-up in [29,32] to further compare Equations (4) and (8), RANSAC, and other alternative methods in the task of 3D road plane detection in KITTA dataset. Each point cloud contains over 10 5 samples with approximately 50 % outliers. Besides, the samples are homogenized and normalized to unit 2 -norm.
We use 11 frames annotated in [29,32] from KITTA dataset. We compare DPCP-PSGM [29], DPCP-IRLS, and DPCP-d [31], which focus on solving the 1 -norm minimization model (4). Besides, we test RANSAC and 2 , 1 -RPCA [25]. Additionally, we test PenNMF and MSP/GPM on solving our proposed model (8), which is a special case of (1). For DPCP-PSGM, DPCP-d, DPCP-IRLS, and 2 , 1 -RPCA, all parameters are set by following the suggestions in [32].
Figure 5 illustrates the numerical performance of all the compared algorithms. We present the numerical results in Figure 5d–f. Moreover, we draw the performance profiles proposed by Dolan and Moré [41] in Figure 5a–c to present an illustrative comparison on the performance of all compared algorithms. The performance profiles can be regarded as distribution functions for a performance metric for benchmarking and comparing optimization algorithms. Besides, we draw the recovery results of frames 328 and 441 in KITTA-CITY-71, which is presented in Figure 6. Here the term “AUC” denotes the area under the AUC curve, and “iterations” denotes the total iterations taken by these compared algorithms. Besides, “Prob” in Figure 5d–f denotes the indexes of tested frames, which are presented in Table 1.
From Figure 5a, we can conclude that PenNMF and MSP/GPM successfully recover the hyperplanes with comparable accuracy. Moreover, in problems 3 , 7 , and 9, PenNMF and MSP produce better classification accuracy than other approaches. Besides, in the aspect of CPU time, PenNMF and MSP cost much less time than other compared algorithms in most cases. Moreover, from Figure 5c, we can conclude that PenNMF takes less time than MSP as well as other compared algorithms in almost all the cases. As a result, we can conclude that our proposed model (1) is easy to be solved and PenNMF shows better efficiency than MSP in our test examples.

5. Conclusions

Sparse dictionary learning (SDL) and dual principal pursuit (DPCP) are two powerful tools in data science. In this paper, we formulate DPCP as a special case of the m -norm maximization on the Stiefel manifold proposed for SDL. Then, we propose a novel smooth unconstrained penalty model PenNM for the original optimization problem with orthogonality constraints. We show PenNM is an exact penalty function of (1) under mild assumptions. We develop an novel approximate gradient approach PenNMF for solving PenNM. The global convergence of PenNMF as well as its sublinear convergence rate are established. Numerical experiments illustrate that our proposed approach enjoys better performance than MSP/GPM [17,18] on various testing problems.

Author Contributions

Conceptualization, methodology, software, validation, formal analysis, investigation, resources, data curation, writing—original draft preparation, writing—review and editing, visualization, supervision, project administration, X.H. and X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

Research is supported in part by the National Natural Science Foundation of China (No. 11971466, 11991021, and 11991020); Key Research Program of Frontier Sciences, Chinese Academy of Sciences (No. ZDBS-LY-7022); the National Center for Mathematics and Interdisciplinary Sciences, Chinese Academy of Sciences; and the Youth Innovation Promotion Association, Chinese Academy of Sciences.

Acknowledgments

We gratefully thank Yuexiang Zhai, Hermish Mehta, Zhengyuan Zhou, and Yi Ma for sharing their codes on MSP. Besides, we also gratefully thank Tianyu Ding, Zhihui Zhu, Tianjiao Ding, Yunchen Yang, Rene Vidal, Manolis Tsakiris, and Daniel Robinson for sharing their codes on DPCP problems.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DPCPdual principal pursuit
DPCP-dDenoised-DPCP
DPCP-IRLSIteratively-Reweighted-Least-Squares algorithm
DPCP-PSGMProjected subgradient-based algorithm for solving DPCP
GPMgradient projection method
MSPmatching, stretching, and projection
PenNMpenalty model for m -norm maximization
PenNMFfirst-order algorithm for solving our penalty model
RANSACRandom Sampling and Consensus
SDLSparse dictionary learning

Appendix A. Proof for Theorem 2 and Corollary 3

In this section, we present the proof for Theorem 2 and Corollary 3. As (6) is an unconstrained optimization problem, the upper-bound for h ( W ) F should be estimated. Before estimating the upper bounds of f ( W ) F and 2 f ( W ) [ W ( W W I p ) ] F , we first present two linear algebraic inequalities:
Lemma A1.
For any A , B R n × N , A B F A F B F .
Proof. 
A B F 2 = i j A i j 2 B i j 2 i j A i j 2 i j B i j 2 = A F 2 B F 2 .
 □
Lemma A2.
For any A R n × N and any m 3 , we have A ( m ) F A F m .
Proof. 
This lemma directly follows the fact that
A ( m ) F 2 = i = 1 n j = 1 N A i , j 2 m i = 1 n j = 1 N A i , j 2 m = A F 2 m .
 □
Now, we present the upper bound estimation for f ( W ) F and 2 f ( W ) [ W ( W W I p ) ] F .
Lemma A3.
For any W R n × p , f ( W ) F Y F m W m 1 .
Proof. 
Due to the fact that f ( W ) = Y ( Y W ) ( m 1 ) sign ( Y W ) ,
f ( W ) F Y F ( Y W ) ( m 1 ) sign ( Y W ) F = Y F ( Y W ) ( m 1 ) F Y F Y W F m 1 Y F m W m 1 .
Here, the last inequality follows the fact that A B F A F B A F B F . □
Lemma A4.
For any W R n × p ,
2 f ( W ) [ W ( W W I p ) ] F ( m 1 ) W m 2 W ( W W I p ) Y F m .
Proof. 
From the expression of 2 f ( W ) in Proposition 1,
2 f ( W ) [ W ( W W ) ] F = ( m 1 ) Y ( Y W ) ( m 2 ) ( Y W ( W W I p ) ) F ( m 1 ) Y F ( Y W ) ( m 2 ) ( Y W ( W W I p ) ) F ( m 1 ) Y F ( Y W ) ( m 2 ) F Y W ( W W I p ) F ( m 1 ) W m 1 W ( W W I p ) Y F m .
Here, the second inequality directly uses Lemma A1 and the last inequality follows Lemma A2. □
In the rest of this section, we consider the equivalence between (1) and (6). We first establish the relationships between the first-order stationary points of (6) and problem (1).
From the optimality condition of (6), we derive an important equality in Lemma A5.
Lemma A5.
For any first-order stationary point W ˜ of (6) and any symmetric matrix T R p × p that satisfies T W ˜ W ˜ = W ˜ W ˜ T , we have
0 = tr T ( W ˜ W ˜ I p ) β W ˜ W ˜ ( W ˜ W ˜ + I p ) 3 2 Λ ( W ˜ ) 1 2 tr T W ˜ 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] .
Proof. 
Suppose W ˜ is a first-order stationary point of (6), by the first-order optimality condition, h ( W ˜ ) = 0 . Then, for any symmetric matrix T R p × p that satisfies T W ˜ W ˜ = W ˜ W ˜ T , W ˜ T , h ( W ˜ ) = 0 .
As described in Proposition 1, f ( W ) can be separated into three parts, we estimate their inner-product with W ˜ T respectively.
First,
W ˜ T , f ( W ˜ ) 3 2 I p 1 2 W ˜ W ˜ W ˜ Φ ( W ˜ f ( W ˜ ) ) = tr T W ˜ f ( W ˜ ) 3 2 I p 1 2 W ˜ W ˜ T W ˜ W ˜ Λ ( W ˜ ) = tr T 3 2 I p 1 2 W ˜ W ˜ Λ ( W ˜ ) T W ˜ W ˜ Λ ( W ˜ ) = 3 2 tr T ( I p W ˜ W ˜ ) Λ ( W ˜ ) .
Here, the second equality follows the fact that that tr B C = tr B C B = 0 holds for any symmetric B and skew-symmetric C. Besides, the last inequality follows that ( W ˜ W ˜ I p ) T = T ( W ˜ W ˜ I p ) . As a result, we achieve the following equality:
W ˜ T , f ( W ˜ ) 3 2 I p 1 2 W ˜ W ˜ W ˜ Φ ( W ˜ f ( W ˜ ) ) 1 2 W ˜ T , 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] = 3 2 tr T ( W ˜ W ˜ I p ) Λ ( W ˜ ) 1 2 T W ˜ , 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] .
Additionally, we estimate their inner-product of W ˜ T and β W ˜ ( W ˜ W ˜ W ˜ W ˜ I p ) and achieve the following equality
W ˜ T , β ( W ˜ W ˜ W ˜ W ˜ I p ) = W ˜ T , β W ˜ ( W ˜ W ˜ + I p ) ( W ˜ W ˜ I p ) = β tr T W ˜ W ˜ ( W ˜ W ˜ + I p ) ( W ˜ W ˜ I p ) .
Based on the above two equations, multiplying ( W ˜ W ˜ I p ) W ˜ on both sides of 0 = h ( W ˜ ) results in
0 = W ˜ T , h ( W ˜ ) = tr T ( W ˜ W ˜ I p ) β W ˜ W ˜ ( W ˜ W ˜ + I p ) 3 2 Λ ( W ˜ ) 1 2 tr T W ˜ 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] ,
and thus we complete the proof. □
Then based on the equality in Lemma A5, the following proposition shows that all first-order stationary point of (6) is uniformly bounded.
Proposition A6.
For any first-order stationary point W ˜ of (6), suppose β ( 4 m + 8 ) Y F m , then W ˜ 2 1 + ( m + 2 ) Y F m β .
Proof. 
Let u denotes the top eigenvector of W ˜ W ˜ , i.e. W ˜ W ˜ u = W ˜ W ˜ u .
Suppose W ˜ is a first-order stationary point that satisfies W ˜ 2 > 1 + ( m + 2 ) Y F m β . By Lemma A5 we first have
0 = u u W ˜ , f ( W ˜ ) = tr β u u ( W ˜ W ˜ I p ) W ˜ W ˜ ( W ˜ W ˜ + I p ) 3 2 tr u u ( W ˜ W ˜ I p ) Λ ( W ˜ ) 1 2 tr u u W ˜ 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] β W ˜ 6 W ˜ 2 3 2 W ˜ 2 u Λ ( W ˜ ) u m 1 2 Y F m W ˜ m + 2 β W ˜ 6 W ˜ 2 ( m + 2 ) Y F m W ˜ m + 2 2 β ( m + 2 ) Y F m 2 W ˜ 6 β W ˜ 4 = W ˜ 4 β ( m + 2 ) Y F m 2 W ˜ 2 β > 0 ,
which leads to the contradictory and shows that W ˜ 2 1 + ( m + 2 ) Y F m β . Here, the second equality directly follows Lemma A5. The first inequality uses Lemma A4 and the fact that W ˜ ( W ˜ W ˜ I p ) W ˜ 3 . Besides, the second inequality follows the fact that W ˜ W ˜ W ˜ 2 u u . The second inequality uses the fact that u Λ ( W ˜ ) u Λ ( W ˜ ) F W ˜ f ( W ˜ ) F W ˜ 2 f ( W ˜ ) F . The fourth inequality uses the fact that tr W ˜ W ˜ I p W ˜ F 2 , and the last inequality follows the fact that W ˜ 2 1 ( m + 2 ) Y F m β . □
Combine Lemma A5 and Proposition A6, we restate Theorem 2 as Theorem A7 and achieve the equivalence between (1) and (6).
Theorem A7.
Suppose β ( 4 m + 8 ) Y F m , and W ˜ is a first-order stationary point of (6), then either W ˜ W ˜ = I p holds, which further implies that W ˜ is a first-order stationary point of problem (1), or the inequality σ min ( W ˜ W ˜ ) ( 2 m + 4 ) Y F m β holds.
Proof. 
When β 4 ( m + 2 ) Y F m , any first-order stationary point W ˜ of (6) satisfies that W ˜ 2 2 .
Suppose W ˜ satisfies W ˜ W ˜ ( 2 m + 4 ) Y F m β I p , then β 2 W ˜ W ˜ ( W ˜ W ˜ + I p ) ( m + 2 ) W ˜ m Y F m · I p . Then from Lemma A5, we have
0 = tr ( W ˜ W ˜ I p ) 2 β W ˜ W ˜ ( W ˜ W ˜ + I p ) 3 2 Λ ( W ˜ ) 1 2 tr ( W ˜ W ˜ I p ) W ˜ 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] tr ( W ˜ W ˜ I p ) 2 β W ˜ W ˜ ( W ˜ W ˜ + I p ) m + 2 2 W m Y F m · I p tr β ( W ˜ W ˜ I p ) 2 W ˜ W ˜ ( W ˜ W ˜ + I p ) 0 ,
showing that tr β ( W ˜ W ˜ I p ) 2 W ˜ W ˜ ( W ˜ W ˜ + I p ) = 0 . Then by the positive-definiteness of W ˜ W ˜ , we can conclude that W ˜ W ˜ I p = 0 .
As a result, we have that either W ˜ W ˜ I p = 0 or σ min ( W ˜ W ˜ ) ( 2 m + 4 ) Y F m β , and completes the proof. □
Corollary A8.
Suppose p = 1 in (1), β ( 4 m + 8 ) Y F m , and W ˜ is a first-order stationary point of (6), then either W ˜ W ˜ = I p holds, which further implies that W ˜ is a first-order stationary point of problem (1), or W ˜ = 0 .
Proof. 
By the same routine of Theorem 2, when β 4 ( m + 2 ) Y F m , any first-order stationary point W ˜ of (6) satisfies that W ˜ 2 2 . Then, following the same proof routine in Lemma A5 and Theorem 2, we have
0 tr ( W ˜ W ˜ I p ) 2 β W ˜ W ˜ ( W ˜ W ˜ + I p ) m + 2 2 W m Y F m · I p = W ˜ 2 1 2 β W ˜ 2 ( W ˜ 2 + I p ) m + 2 2 W m Y F m 0 .
When β 4 m + 8 2 Y F , we have
β 2 W F 4 + β 2 W F 2 > m + 2 2 Y F W ˜ 2 m
holds for any W R n 0 .
Then, for any W ˜ 0 , we can conclude that
β W ˜ 2 ( W ˜ 2 + I p ) m + 2 2 W m Y F m > 0 ,
and thus W ˜ 2 1 = 0 . As a result, from (A1), when W ˜ is a first-order stationary point of (6), either W ˜ = 0 or W ˜ W ˜ I p = 0 . □

Appendix B. Proof for Theorem 4

In this section, we present the main body of the proof for Theorem 4. To show the convergence of PenNMF, we first present some preliminary lemmas. Then, we show that the updating direction D ( W k ) is a descending direction and thus h ( W k + 1 ) h ( W k ) , as illustrated in Lemma A12. Together with Lemma A10, we show that the sequence is restricted in the neighborhood of the constraints, and we achieve the global convergence property of PenNMF in Theorem 4. We first estimate the upper-bound of the term f ( W ) 1 2 W W I p , Λ ( W ) in h ( W ) .
Lemma A9.
For any W R n × p ,
f ( W ) 1 2 W W I p , Λ ( W ) m 2 Y F m max { W 2 + 1 , 2 } W m .
Proof. 
We first estimate the upper-bound for | f ( W ) | , which can be achieved by
| f ( W ) | = 1 m W Y m m 1 m W Y F m 1 m Y F m W m .
Besides, from Lemma A3, we have
1 2 W W I p , Λ ( W ) = 1 2 W ( W W I p ) , f ( W ) 1 2 W ( W W I p ) F f ( W ) F m 2 W ( W W I p ) 2 Y F m W m 1 m 2 Y F m max { 1 , W 2 } W m .
Combine the above two equations, we achieve
| f ( W ) 1 2 W W I p , Λ ( W ) | m 2 Y F m max { W 2 + 1 , 2 } W m ,
and complete the proof. □
We then show that the penalty term ψ ( W ) : = 1 6 W F 6 1 2 W F 2 builds a barrier around S n , p , i.e., those points that are sufficiently far from S n , p have higher functional value than those points that are close to S n , p .
Lemma A10.
Suppose for any δ 0 , 1 3 and β max { 228 m Y F m , 32 m δ Y F m } , we have
max W W I p F 2 δ 8 h ( W ) min W W I p F 2 δ h ( W ) .
Proof. 
Let ψ ( W ) : = 1 6 W 6 W 2 . For any W 1 satisfies ψ ( W 1 ) δ and W 2 satisfies W 2 2 2 and ψ ( W 2 ) 2 δ , then
h ( W 2 ) h ( W 1 ) β δ | f ( W 1 ) | 1 2 W 1 W 1 I p , Λ ( W 1 ) | | f ( W 2 ) | 1 2 W 2 W 2 I p , Λ ( W 2 ) | β δ m Y F m ( 1 + δ ) m + 2 m Y F m 2 m + 2 β δ 19 m Y F m 0 .
Here, the second inequality uses the fact that W 1 2 2 1 + W 1 W 1 I p F 1 + δ , and W 2 2 2 .
Moreover, when W 2 F 2 2 , we have ψ ( W 2 ) β 6 W ˜ 6 . Then, 1 2 ψ ( W 2 ) 1 3 β δ β ψ ( W 1 ) .
h ( W 2 ) h ( W 1 ) m Y F m ( 1 + δ ) m + 2 + ψ ( W 2 ) 16 m Y F m W m + 2 δ β 19 m Y F m W 2 m + 2 + β 12 W ˜ 6 0 .
Here, the second inequality follows the fact that ( 1 + δ ) m ( 1 + δ ) 4 < 3 .
Besides, as W W I p F 2 δ implies
1 6 W F 6 1 2 W F 2 + p 3 = 1 6 tr ( W W I p ) 2 ( W W + 2 I p ) .
As a result, ψ ( W ) 3 δ implies W W I p F δ . Besides, W W I p F 2 δ implies ψ ( W ) 2 3 δ . Therefore,
max W W I p F 2 δ 8 h ( W ) min W W I p F 2 δ h ( W ) .
 □
Lemma A10 shows that the smooth penalty term builds a barrier around S n , p . Moreover, we characterize the relations between D ( W ) F and W W I p F in the following lemma.
Lemma A11.
Suppose δ 0 , 1 3 , set | | W W I p | | F 2 1 8 δ , and β ( 4 m + 8 ) Y F m . Then,
D ( W ) F 3 β 6 · W W I p F ,
where D ( W ) : = f ( W ) W Φ ( W f ( W ) ) + β W ( W W W W I p ) .
Proof. 
First, we present two linear algebra relationships: The first is the inequality | | A | | F A + A 2 F holds for any square matrix A, which is quite obvious and the proof is omitted. The second is the equality | | A B + B A | | F = 2 | | A B | | F holds for any symmetric matrices A and B, which results from the fact | | A B + B A | | F 2 = 2 | | A B | | F 2 + 2 tr ( A B A B ) = 2 | | A B | | F 2 + 2 tr ( A 1 2 B A 1 2 A 1 2 B A 1 2 ) = 4 | | A B | | F 2 .
It follows from the above facts that
W D ( W ) F 1 2 W D ( W ) + D ( W ) W F = 1 2 β W W ( W W + I p ) Λ ( W ) ( W W I p ) + ( W W I p ) β W W ( W W + I p ) Λ ( W ) F β 3 · | | W W I p | | F ,
where the last equality uses the fact that σ min ( W W ( W W + I p ) ) σ min ( W W ) 1 δ 2 3 .
Together with the facts that W D ( W ) F W 2 D ( W ) F and σ max ( W W ) 1 + δ 4 3 , we have
D ( W ) F 1 W F W D ( W ) F 3 2 W D ( W ) F 3 β 6 W W I p F .
 □
Let M 1 : = sup h ( W 1 ) h ( W 2 ) F W 1 W 2 F , then we have that the following illustrating that PenNMF generates a descending sequence { W k } .
Lemma A12.
Suppose δ 0 , 1 3 and β max 228 m Y F m , 32 δ Y F m . Let { W k } be the iterate sequence generated by PenNMF, starting from any initial point W 0 satisfying | | W 0 W 0 I p | | F 2 1 8 δ , and the stepsize η k 1 2 η ¯ , η ¯ , where η ¯ 1 2 M 1 . Then, it holds that
h ( W k + 1 ) h ( W k ) η ¯ 4 D ( W k ) F 2
for any k = 0 , 1 , .
Proof. 
By the explicit expression of h ( W k ) , we first have
h ( W k ) D ( W k ) F = 1 2 f ( W k ) ( W k W k I p ) + 1 2 2 f ( W k ) [ W k ( W k W k I p ) ] F 1 2 f ( W k ) ( W k W k I p ) F + 2 f ( W k ) [ W k ( W k W k I p ) ] F m 2 Y F m W k m 1 W k W k I p F 2 m Y F m W k W k I p F .
Here, the first inequality follows Lemma A3 and Lemma A4.
Besides, by the definition of M 1 , we have
h ( W k + 1 ) h ( W k ) + W k + 1 W k , h ( W k ) + M 1 2 W k + 1 W k F 2 .
Suppose W k W k I p F 2 δ , then by Lemma A11 we can conclude that
h ( W k ) D ( W k ) F 2 m Y F m W k W k I p F 4 3 m Y F m β D ( W k ) F .
Substitute W k + 1 W k = η D ( W k ) and (A14) into (A13), we have
h ( W k + 1 ) h ( W k ) W k + 1 W k , h ( W k ) + M 1 2 W k + 1 W k F 2 η k D ( W k ) , D ( W k ) + M 1 2 η k D ( W k ) F 2 + h ( W k ) D ( W k ) , D ( W k ) η k + 4 3 m Y F m β η k + M 1 2 η k 2 D ( W k ) F 2 η k 2 D ( W k ) F 2 η ¯ 4 D ( W k ) F 2 .
Then by Lemma A10, as h ( W k + 1 ) h ( W k ) , we can conclude that W k + 1 W k + 1 I p F 2 δ . Then, by induction we can conclude that W k W k I p F 2 δ holds for k = 1 , 2 , 3 , . Then, by (A15) again we conclude that
h ( W k + 1 ) h ( W k ) η ¯ 4 D ( W k ) F 2
holds for k = 1 , 2 , 3 , and completes our proof. □
The following lemma shows that when our algorithm stops at W ˜ , then W ˜ is a first-order stationary point of (1).
Lemma A13.
Suppose δ 0 , 1 3 and β max 228 m Y F m , 32 δ Y F m . For any W ˜ satisfying W ˜ W ˜ I p F 2 δ and D ( W ˜ ) = 0 , we have that W ˜ is a first-order stationary point of (1).
Proof. 
Suppose D ( W ˜ ) = 0 , then by the same proof routine in Theorem 2, we consider the inner-product of D ( W ˜ ) and W ˜ ( W ˜ W ˜ I p ) :
0 = D ( W ˜ ) , W ˜ ( W ˜ W ˜ I p ) = f ( W ˜ ) W ˜ Λ ( W ˜ ) + β W ˜ ( W ˜ W ˜ W ˜ W ˜ I p ) , W ˜ ( W ˜ W ˜ I p ) = tr ( W ˜ W ˜ I p ) W ˜ f ( W ˜ ) ( W ˜ W ˜ I p ) W ˜ W ˜ Λ ( W ˜ ) + β W ˜ W ˜ ( W ˜ W ˜ + I p ) ( W ˜ W ˜ I p ) 2 = tr ( W ˜ W ˜ I p ) 2 ( β W ˜ W ˜ ( W ˜ W ˜ + I p ) Λ ( W ˜ ) tr ( W ˜ W ˜ I p ) 2 ( β W ˜ W ˜ ( W ˜ W ˜ + I p ) ( 4 m + 8 ) Y F m · I p β 2 tr ( W ˜ W ˜ I p ) 2 0 .
Here, the fourth equation follows the definition of Λ ( W ˜ ) : = Φ ( W ˜ f ( W ˜ ) ) and the first inequality uses the fact that W ˜ 2 , then together with Lemma A3, we can conclude that Λ ( W ˜ ) 2 ( 4 m + 8 ) Y F m · I p . Besides, the last inequality uses the fact that β 2 W ˜ W ˜ ( W ˜ W ˜ + I p ) ( 4 m + 8 ) Y F m · I p .
Then we can conclude that that W ˜ W ˜ = I p . By the definition of h ( W ˜ ) , we have
h ( W ˜ ) D ( W ˜ ) = 1 2 f ( W ˜ ) ( W ˜ W ˜ I p ) 1 2 2 f ( W ˜ ) [ W ˜ ( W ˜ W ˜ I p ) ] = 0 ,
showing that f ( W ˜ ) = 0 . Together with Theorem 2 we can conclude that W ˜ is a first-order stationary point of (1). □
Based on the Lemmas A10–A13, we restate Theorem 4 as Theorem A14 and show the global convergence property of PenNMF in the following theorem.
Theorem A14.
Suppose δ 0 , 1 3 and β max 228 m Y F m , 32 δ Y F m . Let { W k } be the iterate sequence generated by PenNMF, starting from any initial point W 0 satisfying | | W 0 W 0 I p | | F 2 1 8 δ , and the stepsize η k 1 2 η ¯ , η ¯ , where η ¯ 1 2 M 1 . Then, W k weakly converges to a first-order stationary point of (1). Moreover, for any k = 1 , 2 , , the convergence rate of PenNMF can be estimated by
min 0 i k D ( W i ) F 8 m Y F m + 2 β δ η ¯ ( k + 1 ) .
Proof. 
By Lemma A12, it holds that
h ( W k + 1 ) h ( W k ) η ¯ 4 D ( W k ) F 2 .
If W * is a cluster point of { W k } , we have W k + 1 W k = 0 . Together with W * W * = I p implied by Lemma A11, we can conclude that W * is a first-order stationary point of problem (1).
Calculating the summation of the above inequalities from k = 0 to N 1 , we have
i = 0 k η ¯ 4 D ( W i ) F 2 h ( W 0 ) h ( W k ) < h ( W 0 ) inf | | W W I p | | F 2 δ h ( W ) < sup | | W W I p | | F 2 δ h ˜ ( W ) inf | | W W I p | | F 2 δ h ˜ ( W ) + β 4 | | W 0 W 0 I p | | F 2 | | W k W k I p | | F 2 2 m Y F m + β δ 2 ,
showing that lim inf k + D ( W k ) F = 0 , which further implies that D ( W * ) = 0 . By Lemma A13, D ( W * ) = 0 implies that W * is a first-order stationary point of (1).
Moreover, by (A19), we have that
min 0 i k D ( W i ) F 2 1 k + 1 i = 0 k D ( W i ) F 2 8 m Y F m + 2 β δ η ¯ ( k + 1 ) ,
and complete the proof. □

Appendix C. Additional Experimental Results

In this section, we propose some additional numerical experiments. Figure A1 shows the top 12 basis computed from the testing instances in Section 4.1 by PenNMF. As described in [17], the top bases are those with the largest coefficients in terms of 1 -norm.
Figure A1. The top 12 bases computed from all patches of the test images without noise by PenNMF. (a) “Barbara“; (b) “Boat”; (c) “Duck“; (d) “House”; (e) “Lena“; (f) “Chateau”.
Figure A1. The top 12 bases computed from all patches of the test images without noise by PenNMF. (a) “Barbara“; (b) “Boat”; (c) “Duck“; (d) “House”; (e) “Lena“; (f) “Chateau”.
Sensors 20 03041 g0a1

References

  1. Hansen, T.L.; Badiu, M.A.; Fleury, B.H.; Rao, B.D. A sparse Bayesian learning algorithm with dictionary parameter estimation. In Proceedings of the Sensor Array and Multichannel Signal Processing Workshop (SAM), A Coruña, Spain, 22–25 June 2014; pp. 385–388. [Google Scholar]
  2. Shen, H.; Li, X.; Zhang, L.; Tao, D.; Zeng, C. Compressed Sensing-Based Inpainting of Aqua Moderate Resolution Imaging Spectroradiometer Band 6 Using Adaptive Spectrum-Weighted Sparse Bayesian Dictionary Learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 894–906. [Google Scholar] [CrossRef]
  3. Bai, Y.; Jiang, Q.; Sun, J. Subgradient descent learns orthogonal dictionaries. arXiv 2018, arXiv:1810.10702. [Google Scholar]
  4. Gilboa, D.; Buchanan, S.; Wright, J. Efficient dictionary learning with gradient descent. arXiv 2018, arXiv:1809.10313. [Google Scholar]
  5. Kuo, H.W.; Zhang, Y.; Lau, Y.; Wright, J. Geometry and symmetry in short-and-sparse deconvolution. SIAM J. Math. Data Sci. 2020, 2, 216–245. [Google Scholar] [CrossRef]
  6. Rambhatla, S.; Li, X.; Haupt, J. NOODL: Provable Online Dictionary Learning and Sparse Coding. arXiv 2019, arXiv:1902.11261. [Google Scholar]
  7. Song, X.; Wu, L. A Novel Hyperspectral Endmember Extraction Algorithm Based on Online Robust Dictionary Learning. Remote Sens. 2019, 11, 1792. [Google Scholar] [CrossRef] [Green Version]
  8. Sun, J.; Qu, Q.; Wright, J. Complete dictionary recovery over the sphere I: Overview and the geometric picture. IEEE Trans. Inf. Theory 2016, 63, 853–884. [Google Scholar] [CrossRef]
  9. Wang, D.; Wan, J.; Chen, J.; Zhang, Q. An Online Dictionary Learning-Based Compressive Data Gathering Algorithm in Wireless Sensor Networks. Sensors 2016, 16, 1547. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, L.; Fang, J.; Cheng, H.; Li, H. Sparse Bayesian dictionary learning with a Gaussian hierarchical model. Signal Process. 2017, 130, 93–104. [Google Scholar] [CrossRef] [Green Version]
  11. Wang, Y.; Wu, S.; Yu, B. Unique Sharp Local Minimum in 1-minimization Complete Dictionary Learning. arXiv 2019, arXiv:1902.08380. [Google Scholar]
  12. Zhang, Y.; Kuo, H.W.; Wright, J. Structured local optima in sparse blind deconvolution. IEEE Trans. Inf. Theory 2019, 66, 419–452. [Google Scholar] [CrossRef]
  13. Zhou, Q.; Feng, Z.; Benetos, E. Adaptive Noise Reduction for Sound Event Detection Using Subband-Weighted NMF. Sensors 2019, 19, 3206. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Ling, Y.; Gao, H.; Zhou, S.; Yang, L.; Ren, F. Robust Sparse Bayesian Learning-Based Off-Grid DOA Estimation Method for Vehicle Localization. Sensors 2020, 20, 302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Liu, S.; Huang, Y.; Wu, H.; Tan, C.; Jia, J. Efficient Multi-Task Structure-Aware Sparse Bayesian Learning for Frequency-Difference Electrical Impedance Tomography. IEEE Trans. Industr. Inform. 2020. [Google Scholar] [CrossRef]
  16. Qu, Q.; Zhu, Z.; Li, X.; Tsakiris, M.C.; Wright, J.; Vidal, R. Finding the Sparsest Vectors in a Subspace: Theory, Algorithms, and Applications. arXiv 2020, arXiv:2001.06970. [Google Scholar]
  17. Zhai, Y.; Yang, Z.; Liao, Z.; Wright, J.; Ma, Y. Complete Dictionary Learning via 4-Norm Maximization over the Orthogonal Group. arXiv 2019, arXiv:1906.02435. [Google Scholar]
  18. Shen, Y.; Xue, Y.; Zhang, J.; Letaief, K.B.; Lau, V. Complete Dictionary Learning via p-norm Maximization. arXiv 2020, arXiv:2002.10043. [Google Scholar]
  19. Gao, B.; Liu, X.; Yuan, Y.x. Parallelizable Algorithms for Optimization Problems with Orthogonality Constraints. SIAM J. Sci. Comput. 2019, 41, A1949–A1983. [Google Scholar] [CrossRef] [Green Version]
  20. Wen, Z.; Yang, C.; Liu, X.; Zhang, Y. Trace-penalty minimization for large-scale eigenspace computation. J. Sci. Comput. 2016, 66, 1175–1203. [Google Scholar] [CrossRef] [Green Version]
  21. Xiao, N.; Liu, X.; Yuan, X. A Class of Smooth Exact Penalty Function Methods for Optimization Problems with Orthogonality Constraints. Available online: http://www.optimization-online.org/DB_HTML/2020/02/7607.html (accessed on 26 May 2020).
  22. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Rob. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  23. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from RGBD images. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  24. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  25. Xu, H.; Caramanis, C.; Sanghavi, S. Robust PCA via outlier pursuit. IEEE Trans. Inf. Theory 2010, 58, 3047–3064. [Google Scholar] [CrossRef] [Green Version]
  26. Soltanolkotabi, M.; Candes, E.J. A geometric analysis of subspace clustering with outliers. Ann. Stat. 2012, 40, 2195–2238. [Google Scholar] [CrossRef]
  27. Rahmani, M.; Atia, G.K. Coherence pursuit: Fast, simple, and robust principal component analysis. IEEE Trans. Signal Process. 2017, 65, 6260–6275. [Google Scholar] [CrossRef]
  28. You, C.; Robinson, D.P.; Vidal, R. Provable self-representation based outlier detection in a union of subspaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3395–3404. [Google Scholar]
  29. Ding, T.; Zhu, Z.; Ding, T.; Yang, Y.; Robinson, D.; Vidal, R.; Tsakiris, M. Noisy dual principal component pursuit. In Proceedings of the International Conference on Machine learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  30. Fischler, M.A.; Bolles, R.C. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  31. Tsakiris, M.C.; Vidal, R. Dual principal component pursuit. J. Mach. Learn. Res. 2018, 19, 684–732. [Google Scholar]
  32. Zhu, Z.; Wang, Y.; Robinson, D.P.; Naiman, D.Q.; Vidal, R.; Tsakiris, M.C. Dual principal component pursuit: probability analysis and efficient algorithms. arXiv 2018, arXiv:1812.09924. [Google Scholar]
  33. Shi, L.; Chi, Y. Manifold gradient descent solves multi-channel sparse blind deconvolution provably and efficiently. arXiv 2019, arXiv:1911.11167. [Google Scholar]
  34. Qu, Q.; Li, X.; Zhu, Z. A nonconvex approach for exact and efficient multichannel sparse blind deconvolution. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, CB, Canada, 8–14 December 2019; pp. 4017–4028. [Google Scholar]
  35. Qu, Q.; Sun, J.; Wright, J. Finding a sparse vector in a subspace: Linear sparsity using alternating directions. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QB, Canada, 8–13 December 2014; pp. 3401–3409. [Google Scholar]
  36. Barzilai, J.; Borwein, J.M. Two-point step size gradient methods. IMA J. Numer. Anal. 1988, 8, 141–148. [Google Scholar] [CrossRef]
  37. Dai, Y.H.; Fletcher, R. Projected Barzilai-Borwein methods for large-scale box-constrained quadratic programming. Numer. Math. 2005, 100, 21–47. [Google Scholar] [CrossRef]
  38. Absil, P.A.; Mahony, R.; Sepulchre, R. Optimization Algorithms on Matrix Manifolds; Princeton University Press: Princeton, NJ, USA, 2009. [Google Scholar]
  39. Boumal, N.; Mishra, B.; Absil, P.A.; Sepulchre, R. Manopt, a Matlab toolbox for optimization on manifolds. J. Mach. Learn. Res. 2014, 15, 1455–1459. [Google Scholar]
  40. Mairal, J.; Elad, M.; Sapiro, G. Sparse Representation for Color Image Restoration. IEEE Trans. Image Process. 2008, 17, 53–69. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Dolan, E.D.; Moré, J.J. Benchmarking optimization software with performance profiles. Math. Program. 2002, 91, 201–213. [Google Scholar] [CrossRef]
Figure 1. A detailed comparison among MSP, Manopt, PenCF, and PenNMF. (a)–(c) a comparison with different level of noisy on CPU time; (d)–(f) a comparison with different level of noisy on errors; (g)–(i) a comparison with different n on their CPU time; (j)–(l) a comparison with different n on their errors. The errors are evaluated by 1 W ^ W * 4 4 , where W ^ denotes the final output of all the compared algorithm.
Figure 1. A detailed comparison among MSP, Manopt, PenCF, and PenNMF. (a)–(c) a comparison with different level of noisy on CPU time; (d)–(f) a comparison with different level of noisy on errors; (g)–(i) a comparison with different n on their CPU time; (j)–(l) a comparison with different n on their errors. The errors are evaluated by 1 W ^ W * 4 4 , where W ^ denotes the final output of all the compared algorithm.
Sensors 20 03041 g001
Figure 2. The CPU time of PenCF, PenNMF, MSP/GPM, and Manopt on computing the dictionary. (a) Barbara, Y R 256 × 247 , 009 ; (b) Boat, Y R 256 × 247 , 009 ; (c) Duck, Y R 192 × 62 , 001 ; (d) House, Y R 256 × 58 , 081 ; (e) Lena, Y R 256 × 247 , 009 ; (f) Chateau, Y R 192 × 146 , 633 .
Figure 2. The CPU time of PenCF, PenNMF, MSP/GPM, and Manopt on computing the dictionary. (a) Barbara, Y R 256 × 247 , 009 ; (b) Boat, Y R 256 × 247 , 009 ; (c) Duck, Y R 192 × 62 , 001 ; (d) House, Y R 256 × 58 , 081 ; (e) Lena, Y R 256 × 247 , 009 ; (f) Chateau, Y R 192 × 146 , 633 .
Sensors 20 03041 g002
Figure 3. A comparison between the models (8) and (4) on the average recovery error 1 W ˜ , W * 2 of 5 random trials. (a)–(c) average recovery errors with θ = 0 ; (d)–(g) average recovery errors with θ = 0.1 .
Figure 3. A comparison between the models (8) and (4) on the average recovery error 1 W ˜ , W * 2 of 5 random trials. (a)–(c) average recovery errors with θ = 0 ; (d)–(g) average recovery errors with θ = 0.1 .
Sensors 20 03041 g003
Figure 4. A comparison on the number of successfully recovered instances on the different level of noise. (a) 1 W ˜ , W * 2 is less than 0.1 ; (b) 1 W ˜ , W * 2 is less than 0.2 .
Figure 4. A comparison on the number of successfully recovered instances on the different level of noise. (a) 1 W ˜ , W * 2 is less than 0.1 ; (b) 1 W ˜ , W * 2 is less than 0.2 .
Sensors 20 03041 g004
Figure 5. A comparison between PenNMF, MSP, DPCP-PSGM, DPCP-D, and Random Sampling and Consensus (RANSAC). (a)–(c) performance profile [41] of AUC, iterations and CPU time; (d)–(f) the numerical results of AUC, iterations and CPU time.
Figure 5. A comparison between PenNMF, MSP, DPCP-PSGM, DPCP-D, and Random Sampling and Consensus (RANSAC). (a)–(c) performance profile [41] of AUC, iterations and CPU time; (d)–(f) the numerical results of AUC, iterations and CPU time.
Sensors 20 03041 g005
Figure 6. Illustrations to some results in our numerical tests, with inliers in blue and outliers in red. (a) Frame 328 from KITTI-CITY-71, N = 121766 ; (b) Frame 441 from KITTI-CITY-71, N = 119428 . Inliers/outliers are detected by using a ground-truth thresholding on the distance to the hyperplane recovered by each compared method. The results are represented by projecting 3D point clouds onto the image.
Figure 6. Illustrations to some results in our numerical tests, with inliers in blue and outliers in red. (a) Frame 328 from KITTI-CITY-71, N = 121766 ; (b) Frame 441 from KITTI-CITY-71, N = 119428 . Inliers/outliers are detected by using a ground-truth thresholding on the distance to the hyperplane recovered by each compared method. The results are represented by projecting 3D point clouds onto the image.
Sensors 20 03041 g006
Table 1. The testing instances and their corresponding frames in KITTA dataset.
Table 1. The testing instances and their corresponding frames in KITTA dataset.
DatasetKITTI-CITY-71KITTI-CITY-5KITTI-CITY-48
Frame id.221328441881145120137153021
Test id.1234567891011

Share and Cite

MDPI and ACS Style

Hu, X.; Liu, X. An Efficient Orthonormalization-Free Approach for Sparse Dictionary Learning and Dual Principal Component Pursuit. Sensors 2020, 20, 3041. https://doi.org/10.3390/s20113041

AMA Style

Hu X, Liu X. An Efficient Orthonormalization-Free Approach for Sparse Dictionary Learning and Dual Principal Component Pursuit. Sensors. 2020; 20(11):3041. https://doi.org/10.3390/s20113041

Chicago/Turabian Style

Hu, Xiaoyin, and Xin Liu. 2020. "An Efficient Orthonormalization-Free Approach for Sparse Dictionary Learning and Dual Principal Component Pursuit" Sensors 20, no. 11: 3041. https://doi.org/10.3390/s20113041

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop