Next Article in Journal
A Multiscale Approach to the Smart Deployment of Micro-Sensors over Lightweight Structures
Next Article in Special Issue
Constrained Active Learning for Anchor Link Prediction Across Multiple Heterogeneous Social Networks
Previous Article in Journal
Design and Implementation of a Smart Home System Using Multisensor Data Fusion Technology
Previous Article in Special Issue
Node Scheduling Strategies for Achieving Full-View Area Coverage in Camera Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Non-Convex Sparse and Low-Rank Based Robust Subspace Segmentation for Data Mining

1
School of Information Science & Technology, Donghua University, Shanghai 200051, China
2
Department of Electronic Engineering, City University of Hong Kong, Kowloon 999077, Hong Kong, China
3
School of Mathematics and Computer Science, Northeastern State University, Tahlequah, OK 74464, USA
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(7), 1633; https://doi.org/10.3390/s17071633
Submission received: 16 June 2017 / Revised: 8 July 2017 / Accepted: 10 July 2017 / Published: 15 July 2017

Abstract

:
Parsimony, including sparsity and low-rank, has shown great importance for data mining in social networks, particularly in tasks such as segmentation and recognition. Traditionally, such modeling approaches rely on an iterative algorithm that minimizes an objective function with convex l1-norm or nuclear norm constraints. However, the obtained results by convex optimization are usually suboptimal to solutions of original sparse or low-rank problems. In this paper, a novel robust subspace segmentation algorithm has been proposed by integrating lp-norm and Schatten p-norm constraints. Our so-obtained affinity graph can better capture local geometrical structure and the global information of the data. As a consequence, our algorithm is more generative, discriminative and robust. An efficient linearized alternating direction method is derived to realize our model. Extensive segmentation experiments are conducted on public datasets. The proposed algorithm is revealed to be more effective and robust compared to five existing algorithms.

1. Introduction

High dimensionality research for data mining is an essential topic in modern imaging applications, such as social networks and the Internet of Things (IoT). It is worth noting that data of high dimension is often supposed to reside in several subspaces of lower dimension. For instance, facial images with various lightning conditions and expressions lie in a union of a nine-dimensional linear subspace [1]. Moreover, moving motions in videos [2] and hand-written digits [3] can also be approximated by multiple low-dimensional subspaces. Inspiringly, these characteristics enable effective segmentation, recognition, and classification to be carried out. The problem of subspace segmentation [4] is formulated as determining the number of subspaces and partitions the data according to the intrinsic structure.
Many subspace segmentation algorithms have emerged in the past decades. Some of these methods are algebraic or statistical. Among the algebraic methods, generalized principal component analysis (GPCA) [5] is mostly widely used. GPCA characterizes the data subspace with the gradient of a polynomial, and segmentation is obtained by fitting the data with polynomials. However, the performance drops quickly in the presences of noise, and polynomial fitting computation is time consuming. As to the statistical algorithms, including random sample consensus (RANSAC) [6], factorization-based methods [7,8], and probabilistic principal component analysis (PPCA) [9], the estimation of exact subspace models exquisitely changed the performance of this type of methods. Recently, spectral-type methods [10] like sparse representation (SR) [7,11,12], low-rank representation (LRR) [13,14,15], and the extensions base on SR or LRR [16,17,18,19,20,21,22] have attracted much attention, as they are robust to noise [11,12,13,14], with strong theoretical backgrounds [16] and are easy to implement. Once an affinity graph is learned from SR or LRR, segmentation results can be obtained by means of spectral clustering. Therefore, building an affinity graph that accurately captures relevant data structures is a key point for SR- and LRR-based spectral-type models.
A good affinity graph should preserve the local geometrical structure, as well as the global information [11]. Recently, Elhamifar et al. [23] presented the sparse subspace clustering (SSC) based on an l1-graph. The l1-graph is constructed by using SR coefficients deduced in the l1-norm minimization to represent the relationships among samples. Attributed to SR, the l1-graph is sparse, capable of finding neighborhood according to the data and robust to noisy data. Inspired by l1-graph, various SR-graphs have been proposed [24]. Wang et al. [25] provided l2-Graph based subspace clustering to eliminate errors from various types of projection spaces. Peng et al. [26] proposed a unified framework for representation-based subspace clustering methods to cluster both the out-of-sample and the large-scale data. Peng et al. [27] introduced principal coefficients embedding to automatically identify the number of features, as well as learn the underlying subspace in the presence of Gaussian noise. Wang et al. [28] proposed a race lasso-based regularizer for multi-view data while keeping individual views well encapsulated. However, SR-based methods seek the sparsest representation coefficients of each data point individually, without a global structure regularizer. This drawback alters the robustness of these methods in the presence of outliers [15], when the data is not “clean” enough. To account for the underlying global information, Liu et al. [15] presented a two-step algorithm which firstly computes low-rank representation coefficients for data and then uses the coefficient matrix to build an affinity graph (LRR-graph). The LRR-graph jointly represents all the data by solving a nuclear norm optimization problem and, thus, becomes a better option at capturing global information. Numerous LRR-graph-based methods have also been proposed for spectral clustering. For example, Wang et al. [28] proposed a multi-graph Laplacian-regularized LRR to characterize the non-linear spectral graph structure from each view. Both SSC and LRR are based on convex relaxations of the initial problems, SSC uses l1-norm to approximate the number of non-zero elements, while LRR applies nuclear norm to compute the number of non-vanishing singular values. These convex relaxations yield solutions that deviate far from solutions to the original problems. Hence, it is desirable to rather use non-convex surrogates, without causing any significant increase in computational complexity.
A large number of studies on non-convex surrogates for l0-norm problem have been addressed recently. Xu et al. [29] introduced the l1/2-norm in noisy signal recovery with an efficient iterative half-thresholding algorithm. Similarly, Zhang et al. [30] proposed the lp-norm minimization with a generalized iterative shrinkage and thresholding method (GIST). Their study shows lp-norm based model is more effective on image denoising and image deburring. Numerous researchers also considered the other non-convex surrogate functions, such as homotopic L0-minimization [31], smoothly clipped absolute deviation (SCAD) [32], multi-stage convex relaxation [17], logarithm [33], half-quadratic regularization (HQR) [34], exponential-type penalty (ETP) [35], and minimax concave penalty (MCP) [36]. Recent years have also witnessed the progress in non-convex rank minimization functions. Mohan et al. [37] developed an efficient IRLS-p to minimize rank function, and improved the recovery performance for matrix completion. Xu et al. [18] introduced the S1/2-norm with an efficient ADMM solver for video background modeling. Kong et al. [38] proposed a Schatten p-norm constrained model to recover noisy data. Another popular non-convex rank minimization is the truncated nuclear norm [39]. They all compete with the state-of-the-art algorithms to some extent.
Combining the non-convex lp-norm regularizer with the Schatten p-norm (0 < p ≤ 1), in this study, we propose a robust method named non-convex sparse and low-rank-based robust subspace segmentation (lpSpSS). Our lp-norm error function can better predict errors, which further improves the robustness of subspace segmentation. Meanwhile, the Schatten p-norm-regularized objective function shows a better ability to approximate the rank of coefficient matrix compared with the nuclear norm. Our method can provide a more accurate description of the global information and better measurement of data redundancy. Thus, our new objective is to solve joint the lp-norm and Schatten p-norm (0 < p ≤ 1) minimization together. When p→0, our proposed lpSpSS turns to be more robust and effective than SR- and LRR-based subspace segmentation algorithms. In addition, we enforce non-negative constraint to the reconstruction coefficients, which aids interpretability and allows better solutions in numerous application areas such as text mining, computer vision, and bioinformatics. Traditionally, an alternating direction method (ADM) [40] can solve this optimization problem efficiently. However, to increase the speed and scalability of the algorithm, we choose an efficient solver commonly named the linearized alternating direction method with adaptive penalty (LADMAP) [19]. As it is based on fewer auxiliary parameters and without an inverse of its matrix, it is more efficient than ADM. Numerical experimental results verify our proposed method, which consistently obtains better segmentation results.
The rest of this paper is structured as follows: In Section 2, the notations, as well as the overview of SSC and LRR, will be presented. Section 3 is dedicated to introducing our novel non-convex sparse and low-rank based robust subspace segmentation. Section 4 conducts multiple numerical experiments to examine the effectiveness and robustness of lpSpSS. Section 5 concludes this work.

2. Background

This section is divided into three parts. First, the notation and definition are illustrated. The background of two algorithms, SSC and LRR, will be discussed in Section 2.2 and Section 2.3, respectively.

2.1. Notations and Definitions

Suppose X = [ x 1 , , x N ] is an image matrix consists of N sufficiently dense data points { x i R m } i = 1 N , { S i } i = 1 n is an arrangement of n subspaces. Let {xi} be drawn from n subspaces { S 1 S 2 S n } of lower dimension. Given X, the goal of subspace segmentation is to partition the data points into the underlying low-dimensional subspaces.
The lp-norm (0 < p < ∞) of vector xR1 can be expressed as x p = ( 1 n x i p ) 1 / p , in which xi is the i-th element. Therefore, the p-norm of xR1 to the power p can be expressed as x p p = 1 n x i p .
The Schatten p-norm of a matrix xRn×m is expressed as:
X S p = ( i = 1 min ( m , n ) σ i p ) 1 / p
in which 0 < p ≤ 1, and σi is the i-th largest singular value. Thus, it can be deduced that:
X S p p = ( i = 1 min ( m , n ) σ i p )
The Schatten 1-norm is just nuclear norm | X | * , while the Schatten 0-norm is the approximation of the rank of X. Compared with | X | * , | X | S p is a better approximation of the rank of X.

2.2. Sparse Subspace Clustering

Recently, SSC [16] has grabbed considerable attention. The hypothesis states that data are drawn from several subspaces of lower dimension, and can be sparsely self-expressed. More formally, SSC aims to solve the following program:
min Z i Z i 0 s . t . X = X Z a n d d i a g ( Z ) = 0
where Z i = [ z i 1 , z i 2 , , z i N ] T R N are the reconstruction coefficients, and z 0 refers to the number of nonzero values. As it is difficult to solve this non-convex objective, a convex l1 minimization problem is proposed by solving the following program:
min Z Z 1 s . t . X = X Z a n d d i a g ( Z ) = 0
The minimization problem in Equation (4) can be concluded using the alternating direction method of multipliers (ADMM) [19]. Afterwards, the coefficient matrix Z can be utilized to construct the affinity matrix as W = | Z | + | Z T | . Finally, W is performed via spectral clustering and the segmentation result is drawn. While SSC works well in practice, the model is invalid when the obtained similarity graph is poorly connected (we refer readers to Soltanolkotabi et al. [41] for very recent results in this direction).

2.3. Low-Rank Representation-Based Subspace Segmentation

The difference between LRR and SSC is that, LRR seeks the lowest rank representation Z but not the sparsest representation. LRR is based on the assumption that for observed data X R d × N drawn from n low-dimensional subspaces, and the rank of coefficient matrix r = r a n k ( Z ) = i = 1 n z i is assumed to be much smaller than min { d , N } . The LRR is formulated as:
min z   r a n k ( Z ) s . t . X = X Z
As the rank function minimization is non-convex, Equation (5) can be reformulated as the following convex minimization problem:
min z Z * s . t . X = X Z
in which Z * is the nuclear norm, which yields a good approximation to the matrix rank of Z. Singular value threshold (SVD) can be used to efficiently solve Equation (6), when there is no error present in X.
When the data X is noisy, an extension of LRR is proposed as follows:
min Z , E Z * + λ E p s . t . X = X Z + E
in which λ ≥ 0 is the tradeoff parameter, trading off low rankness between reconstruction error. E p R d × N is the noise term with different regularization strategies, which depends on the property of Ep. When the noise term is Gaussian noise, E p = E F 2 , in which F refers to the Frobenius norm. When the noise term are entry-wise corruptions, E p = E 1 , in which 1 refers to the l1 norm. When the noise term are sample-specific corruption and outliers, E p = E 2 , 1 , in which E 2 , 1 = i = 1 N i = 1 d E i j 2 = i = 1 N E j 2 refers to the l2,1 norm. Equation (7) can be solved by ADMM [19] to obtain the coefficient matrix Z. Afterwards, the coefficient matrix Z can be utilized for the construction of affinity matrix W = | Z | + | Z T | . Finally, spectral clustering can be applied to W for segmentation results.

3. Non-Convex Sparse and Low-Rank Based Robust Subspace Segmentation

In this section, we first propose the non-convex sparse and low-rank-ased robust subspace segmentation model, in which we combine the lp-norm with the Schatten p-norm together for clustering, and then use LADMAP to solve lpSpSS. Finally, we analyze of the time complexity of lpSpSS.

3.1. Model of lpSpSS

We consider the non-convex sparse and low-rank-based subspace segmentation for data contaminated by noise and corruption. Notice that the nuclear norm is replaced by the Schatten p-norm, when p is smaller than 1, the underlying global information can be captured more effectively. Additionally, the lp norm (0 < p ≤ 1) of the coefficient matrix is also introduced as an error function, in order to harvest stronger robustness to noise [42]. It has been demonstrated in some recent research [43] that the Schatten p-norm is more powerful than the nuclear norm in matrix completion, and the recovery performance of the lp-norm is also superior to the convex l1-norm [36]. Our lpSpSS will surely be more effective than the convex methods.
We begin by considering the relaxed low-rank subspace segmentation problem, which is equivalent to:
min Z , E Z S p p + β Z p p + λ E 2 , 1 s . t . X = X Z + E , Z 0
In which, the first term Z p p is the lp-norm, which improves the integration of the local geometrical structure. Meanwhile, the second term Z S p p is the Schatten p-norm, which can better approximate the rank of Z. Moreover, the third term reconstruction error E 2 , 1 is the l2,1 norm, which can better characterize errors like corruption and outliers. β and λ are trade-off parameters. Regarding the widely-used non-negative constraint (Z ≥ 0), which is to ensure direct use of the reconstruction coefficients in the affinity graph construction.

3.2. Solution to lpSpSS

3.2.1. Brief Description of LADMAP

We adopt LADMAP [19] to solve the objective function (Equation (8)) constrained by the lp-norm norm and Schatten-p regularizers. An auxiliary variable W is introduced and the optimization problem becomes separable. Thus, Equation (8) is rewritten as:
min Z , E Z S p p + β Z p p + λ E 2 , 1 s . t . X = X Z + E , Z = W , W 0
To remove two linear constraints in Equation (9), we introduce two Lagrange multipliers Y1 and Y2, hence, the optimization problem is defined using the following Lagrangian function:
L ( Z , W , E , Y 1 , Y 2 , u ) = Z S p p + β W p p + λ E 2 , 1 + Y 1 , X X Z E + Y 2 , Z W + μ 2 ( X X Z E F 2 ) + ( Z W F 2 ) = Z S p p + β W p p + λ E 2 , 1 + q ( Z , W , E , Y 1 , Y 2 , μ ) 1 2 μ ( Y 1 F 2 + Y 2 F 2 )
where q ( Z , W , E , Y 1 , Y 2 , μ ) = μ 2 ( X X Z E + Y 1 / μ F 2 + Z W + Y 2 / μ F 2 ) , Y1 and Y2 are the Lagrange multipliers, and μ ≥ 0 is a trade-off parameter. We solve Equation (10) by minimizing L to update each variable with the other variables fixed. The updating schemes at each iteration can be designed as follows:
W ( k + 1 ) = arg min W 0 β W p p + μ ( k ) 2 Z ( k + 1 ) W + Y 2 ( k ) / μ ( k ) F 2
Z ( k + 1 ) = arg min Z Z S p p + Z q ( Z ( k ) , W ( k ) , E ( k ) , Y 1 ( k ) , Y 2 ( k ) , μ ( k ) , Z Z ( k ) + μ ( k ) θ 2 Z Z ( k ) F 2 = arg min Z Z S p P +   θ μ ( k ) 2 | Z Z ( k ) + [ X T ( X X Z ( k ) E ( k ) + Y 1 ( k ) μ ( k ) ) + ( Z ( k ) W ( k ) + Y 2 ( k ) μ ( k ) ) ] / θ | F 2
E ( k + 1 ) = arg min E λ E 2 , 1 + u ( k ) 2 X X Z ( k + 1 ) + Y 1 ( k ) u ( k ) E F 2 = Ω λ u ( k ) 1 ( X X Z ( k + 1 ) + Y 1 ( k ) μ ( k ) )
In Equation (12), Z q is the partial differential of q with respect to Z, θ = X F 2 . In particular, the detailed procedures of LADMAP are shown in Algorithm 1. The first, Equation (11), and the second, Equation (12), are solved using the following subsections. The last convex problem (Equation (13)) can solved by the l2,1-norm minimization operator [15].
Algorithm 1. LADMAP for solving Equation (9).
Input: Data X, tradeoff variables β > 0, λ > 0
Initialize: Z(0) = W(0) = E(0), Y1(0) = Y1 (0) = 0, µ(0) = 0.1, µmax = 107, ρ0 = 1.1, ε1 = 107, ε2 = 106, k = 0, θ = X F 2 , and the number of maximum iteration MaxIter = 1000.
While not converged and k ≤ MaxIter do
(1)
Compute W(k+1) by solving Equation (11) with Z(k+1), E(k+1) are fixed;
(2)
Compute Z(k+1) by solving Equation (12) with W(k+1), E(k+1) are fixed;
(3)
Compute E(k+1) by solving Equation (13) with W(k+1), Z(k+1) are fixed;
(4)
Compute the multipliers Y1 (k + 1)and Y2 (k + 1) as follows:
{ Y 1 ( k + 1 ) = Y 1 ( k ) + μ ( k ) ( X X Z ( k + 1 ) E ( k + 1 ) ) Y 2 ( k + 1 ) = Y 2 ( k ) + μ ( k ) ( Z ( k + 1 ) W ( k + 1 ) )
(5)
Compute µ(k + 1) as follows:
μ ( k + 1 ) = min ( μ max , ρ μ ( k ) ) , in which
ρ = { ρ 0 , i f μ ( k ) Ω / X F ε 2 1 , o t h e r w i s e
(6)
Converges when
{ X X Z ( k ) E ( k ) F / X F ε 1 o r μ ( k ) Ω / X F ε 2
In which Ω = max ( θ Z ( k ) Z ( k + 1 ) F , W ( k ) W ( k + 1 ) F , E ( k ) E ( k + 1 ) F )
(7)
Increment of iteration: k = k + 1.
End while
Output: The optimal solution W, Z and E.

3.2.2. Solving the Non-Convex lp-Norm Minimization Subproblem (Equation (11))

For each element in { X i j | ( i ,   j ) Ω } , we can decouple Equation (11) into a simplified formula:
min 1 2 ( x y ) 2 + λ | x | p
Recently, Zhang et al. solved this lp-norm optimization problem via the proposed GIST [30]. For lp-norm minimization, the thresholding function is:
τ p ( λ ) = ( 2 λ ( 1 p ) ) 1 / ( 2 p ) + λ p ( 2 λ ( 1 p ) ) ( p 1 ) / ( 2 p )
Meanwhile, the generalized soft-thresholding function is:
T p ( y ; λ ) = { 0 , i f | y | τ p ( λ ) sgn ( y ) S p ( | y | ; λ ) , i f | y | > τ p ( λ )
The corresponding thresholding rule in the generalized soft-thresholding function is T p G S T ( y ; λ ) = 0 when | y | τ p G S T ( λ ) , and corresponding shrinkage rule is T p G S T ( y ; λ ) = sgn ( y ) S p G S T ( | y | ; λ ) when | y | > τ p G S T ( λ ) .

3.2.3. Solving the Non-Convex Schatten p-Norm Minimization Subproblem (Equation (12))

We can reformulate Equation (12) as the following simplified notation:
min X 1 2 X A F 2 + λ X s p p
After applying SVD on X, X is decomposed into summation of r rank-p matrices X = UΔVT. Here, U is the left singular vector, Δ is the non-zero singular diagonal matrix, and V is the right singular vector. The i-th singular value δi is solved by:
min δ i 0 1 2 ( δ i σ i ) 2 + λ δ i p
Equation (16) can be used to solve Equation (18) again. For p = 1, we can obtain the same solution with nuclear norm minimization [44].

3.3. Convergence and Computational Complexity Analysis

Although Algorithm 1 is described in three major alternative steps for solving W, Z, and E, we can actually combine steps for Z and E easily into one larger block step by simultaneously solving for (Z, E). Thus, the convergence conclusion of two variables LADMAP in [45] can be applied to our case. Finally, the convergence of the algorithm is ensured.
Suppose the size of X are d × n, k is number of total iterations, and r is the lowest rank for X. The major time consumption of Algorithm 1 is mainly determined by Step 2, as it involves time-consuming SVDs. In Step 1, each component of Z k q can be computed in O(rn2) by using the skinny SVD to update W. In Step 2, the complexity of the SVD to update Z is approximately O(d2n). In Step 3, the computation complexity of l2,1 minimization operator is about O(dn). The total complexity is, thus, O(krn2 + kd2n). Since rmin(d, n), the time cost is, at most, O(krn2).

3.4. Affinity Graph for Subspace Segmentation

Once Equation (9) was solved by LADMAP, we can obtain the optimal coefficient matrix Z*. Since every sample is reconstructed by its neighbors, Z* naturally characterizes the relationships among samples. Such information is a good indicator of similarity among samples, we use the reconstruction coefficients to build the affinity graph. The non-convex lp-norm ensures that each sample only connects to few samples. As a result, the weights of the graph tend to be sparse. While the non-convex Schatten p-norm ensures samples lying in the same subspace are highly correlated and tend to be assigned into the same cluster, Z* is theoretically able to capture the global information, and the graph weights are constrained with non-negativity, as they reflect similarities between data points.
After obtaining the coefficient matrix Z*, the reconstruction coefficients of each sample are normalized and thresholded to zero. Therefore, the obtained normalized sparse Z ^ * can be used to compute the affinity graph W = ( Z ^ * + ( Z ^ * ) T ) / 2 . Finally, W carries out spectral clustering to obtain the segmentation results. Our proposed non-convex sparse and low-rank based subspace segmentation is outlined in Algorithm 2.
Algorithm 2. lpSpSS.
Input: Matrix X = [ x 1 , x 2 , , x n ] R d × n , tradeoff parameters β, λ
  • Normalize each sample xi to obtain X ^ = [ x ^ 1 , x ^ 2 , , x ^ n ] .
  • Solve the non-convex sparse and low-rank constrained program by Algorithm 1:
    min Z S p p + β Z p p + λ E 2 p , s . t . X = X Z + E , Z 0
    and get (Z*, E*).
  • Normalize coefficient matrix Z * , and threshold small values by θ to obtain Z ^ * .
  • Set weights for affinity graph by W = ( Z ^ * + ( ( Z ^ * ) T ) / 2 .
  • The data is segmented by spectral clustering.
Output: The segmentation results.

4. Experimental Evaluation and Discussion

In the following, we will discuss the performance of our proposed lpSpSS model. Firstly, the experimental setting is detailed in Section 4.1. From Section 4.2Section 4.3Section 4.4, we will test the segmentation performance of lpSpSS on CMU-PIE, COIL20, and USPS. In Section 4.5, we will examine the robustness of lpSpSS to block occlusions and pixel corruptions. Finally, the discussion of experimental results will be given in Section 4.6.

4.1. Experimental Settings

The proposed lpSpSS approach will be evaluated on realistic images and compared with five related works. We use four publicly-available datasets, including CMU-PIE [46], COIL20 [47], USPS [48], and Extended Yale B [49]. Among them, datasets [46] and [49] contain face images with various poses/illuminations/facial expressions, COIL20 consists of different general objects, and USPS includes handwritten digit images. Our proposed lpSpSS will be compared with five segmentation methods, including PCA, SSC [16], LRR [15], and NNLRS [50], while K-means serves as a baseline for comparison.
We adopt the same experimental settings as Zhuang’s work [50]. For the compared methods, a grid search strategy is used for selecting model parameters, and the optimal segmentation is achieved by tuning the parameters carefully. As to our lpSpSS, there are two regularized parameters, β and λ, affecting its performance. We take a stepwise selection strategy to search the best parameters. For example, we search the possible candidate interval λ may exit, with β fixed, and alternatively search λ’s most possible candidate interval, with β fixed. Finally, the best values are found in a two-dimensional candidate space of (β, λ).
To quantitatively and effectively measure the segmentation performance, two quantity metrics, namely accuracy (AC) and normalized mutual information (NMI) [51], are used in our experiments. All the experiments are implemented by MATLAB, on a MacBook Pro with a 2.6 GHz Intel Core i7 CPU and 16 GB memory.

4.2. Segmentation Results on CMU-PIE Database

In this experiment, we compare lpSpSS with the other five methods on the CMU-PIE facial images dataset. It includes 41,368 pictures of 68 persons, acquired with various postures and lighting scenarios. The resolution of each image is 32 × 32 = 1024 pixels. Typical examples of CMU-PIE are shown in Figure 1. For each given cluster number K = 4,..., 68 in the whole dataset, the segmentation results with different K were averaged on the twenty tests. The averaged segmentation performance of proposed and existing algorithms on the CMU-PIE dataset [46] are reported in Table 1.
We can see that our proposed lpSpSS achieves the best segmentation AC and NMI on CMU-PIE dataset, which proves the effectiveness of our lpSpSS. For example, the average segmentation accuracy of NNLRS and lpSpSS are 84.1% and 89.8%, respectively. lpSpSS improves the segmentation accuracy by 5.7% compared with NNLRS (the second best algorithm). The improvement of lpSpSS indicates the importance of the non-convex SR and LRR affinity graph.

4.3. Segmentation Results on COIL20 Database

When it comes to the evaluation using second dataset COIL20 [49], the proposed lpSpSS is compared with five existing algorithms. This dataset contains 1440 images of 20 objects, with 72 different views. The resolution of each picture is 32 × 32 = 1024 pixels. Typical examples of COIL20 are shown in Figure 2. For each given cluster number K = 2,...,20 in the whole dataset, the segmentation results with different K were averaged on the twenty tests. The averaged segmentation performances of the proposed and existing algorithms on the COIL20 dataset [47] are reported in Table 2.
Experimental results on COIL20 indicates that our lpSpSS outperforms the other five existing algorithms. For example, the average AC and NMI of lpSpSS are 90.2% and 91.0%, which are higher than for NNLRS (the second best algorithm) by 2.0% and 1.5%, respectively. Especially, when the cluster number is large, the superiority of lpSpSS is very obvious.

4.4. Segmentation Results on USPS Handwritten Digit Dataset

In this experiment, we compare lpSpSS with the other five methods on the USPS handwritten digit dataset. It contains 9298 images of 10 classes, with a variety of orientations. The resolution of each picture is 16 × 16 = 256 pixels. Figure 3 shows typical sample images in the USPS dataset. For each given cluster number K = 2,...,10, the segmentation results with different K were averaged on the twenty tests. The averaged segmentation performances of proposed and existing algorithms on the USPS handwritten digit dataset are reported in Table 3.
Table 3 shows that our proposed lpSpSS still obtains the best segmentation performance. This result demonstrates that a non-convex sparse and low-rank graph is better to model complex related data than traditional SR and LRR based graphs. Experimental results have demonstrated that our proposed lpSpSS model can not only represent the global information, but also preserves the local geometrical structures in the data by incorporating the non-convex lp-norm regularizer.

4.5. Segmentation Results on Dataset with Block Occlusions and Pixel Corruptions

Finally, we evaluate the robustness of each model on the more challenging Extended Yale B face dataset. It has 38 × 64 facial images, with various lighting scenarios. To reduce the resources and budget, the resolution of each picture is downsized to 96 × 84. This dataset more challenging for subspace segmentation, as 50% of the samples with hard shadows or specularities. Figure 4 shows typical sample images of Extended Yale B.
We select 1134 images from first 18 individuals to evaluate the different methods. Two types of corruptions are introduced into this experiment. For Type I block occlusions, different block sizes (from 5 × 5 to 20 × 20) are added to randomly selected locations of the images. For Type II random pixel corruption, randomly-chosen pixels on each images are substituted with equally-distributed random values. The proportion of corrupted pixels per image is from 0 to 20%. Some examples from corrupted Extended Yale B face images are shown in Figure 5. The averaged segmentation AC of proposed and existing algorithms with multiple block occlusions and pixel corruptions are tabulated in Table 4 as well as Table 5.
Both experimental results show that our lpSpSS achieves the best segmentation results again. Segmentation results suggest that proposed lpSpSS is more robust than compared methods, especially when a significant portion of the realistic samples are corrupted.

4.6. Discussions

Our lpSpSS outperforms five existing subspace segmentation algorithms. Especially, in the case of CMU-PIE face dataset, the improvement by lpSpSS is largest. Our affinity graph can capture the local geometrical structure, as well as the global information of the data, hence, is both generative and discriminative.
lpSpSS is more robust than the other compared methods, which can properly deal with multiple noises. Images in Extended Yale B dataset contain different errors, including block occlusions, pixel corruptions, illuminations, partition them is challenging. However, lp-norm and Schatten p-norm are introduced for lpSpSS affinity graph construction. Therefore, our model can better predict errors and is a better measurement for data redundancy.
The segmentation performance of SSC and LRR are almost the same. For example, the segmentation accuracy of SSC on CMU-PIE is 0.9% better than LRR, while the performance of LRR on USPS is 0.7% higher than SSC. The segmentation results heavily depends on the intrinsic structure of the testing dataset, it is difficult to determine which one is better.
The LRR-based algorithm is robust in handling noisy data. It aims at obtaining the low rankness of coefficient matrix, thus, the LRR-based methods can better model the global information. Furthermore, LRR can find similar clusters which measure the data redundancy, ensuring high quality and stability of the segmentation results. For data heavily contaminated with corruptions or outliers, the model can find lower ranks that will be more robust to noise.
However, SVD computation for Schatten p-norm minimization is performed in each iteration, which is very time consuming, and the best segmentation results are not achieved at the lowest value of p. Hence, we will be focused on the study of speeding up the Schatten p-norm solver and the selection of best p values in our future work.

5. Conclusions

This paper presents an accurate and robust for subspace segmentation, named lpSpSS, by introducing the non-convex lp-norm and Schatten p-norm minimization. Taking advantages from the original sparsity and low rankness of data of high dimension, both local geometrical structure and the global information of the data can be learnt. A linearized alternating direction method with adaptive penalty (LADMAP) is also introduced to search for optimal solutions. Numerous experiments on CMU-PIE, COIL20, USPS, and Extended Yale B verify the effectiveness and robustness of our lpSpSS compared to five existing works.

Acknowledgments

This research is supported by the National Natural Science Foundation of China (grant no. 61601112), the Fundamental Research Funds for the Central Universities, and the DHU Distinguished Young Professor Program.

Author Contributions

This paper is prepared by all authors. W.C., M.Z., N.X., and K.T.C. proposed the idea. W.C. designed the algorithm, performed the numerical experiments, and analyzed the experimental results. W.C., M.Z., N.X., and K.T.C. wrote and revised this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shi, X.; Yang, Y.; Guo, Z.; Lai, Z. Face recognition by sparse discriminant analysis via joint l2,1-norm minimization. Pattern Recogn. 2014, 47, 2447–2453. [Google Scholar] [CrossRef]
  2. Costeira, J.P.; Kanade, T. A multibody factorization method for independently moving objects. Int. J. Comput. Vis. 1998, 29, 159–179. [Google Scholar] [CrossRef]
  3. Hastie, T.; Simard, P.Y. Metrics and models for handwritten character recognition. Stat. Sci. 1998, 13, 54–65. [Google Scholar] [CrossRef]
  4. Parsons, L.; Haque, E.; Liu, H. Subspace clustering for high dimensional data: A review. SIGKDD Explor. 2004, 6, 90–105. [Google Scholar] [CrossRef]
  5. Vidal, R.; Ma, Y.; Sastry, S. Generalized principal component analysis (GPCA). IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1945–1959. [Google Scholar] [CrossRef] [PubMed]
  6. Fischler, M.A.; Bolles, R.C. Random sample consensus-A paradigm for model-fitting with applications to image-analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  7. Lu, Y.; Lai, Z.; Xu, Y.; Li, X.; Zhang, D.; Yuan, C. Non-negative Discriminant Matrix Factorization. IEEE Trans. Circuits Syst. Video Technol. 2012, 45, 4080–4091. [Google Scholar] [CrossRef]
  8. Bailey, T.L.; Elkan, C. Fitting a mixture model by expectation maximization to discover motifs in bipolymers. In Proceedings of the International Conference on Intelligent Systems for Molecular Biology, Menlo Park, CA, USA, 14–17 August 1994; pp. 28–36. [Google Scholar]
  9. Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 1999, 61, 611–622. [Google Scholar] [CrossRef]
  10. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2006; pp. 849–856. [Google Scholar]
  11. Wright, J.; Yang, A.Y.; Ganesh, A.; Sastry, S.S.; Ma, Y. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar] [CrossRef] [PubMed]
  12. Wright, J.; Ma, Y.; Mairal, J.; Sapiro, G.; Huang, T.S.; Yan, S.C. Sparse representation for computer vision and pattern recognition. Proc. IEEE 2010, 98, 1031–1044. [Google Scholar] [CrossRef]
  13. Fang, X.; Xu, Y.; Li, X.; Lai, Z.; Wong, W.K. Robust semi-supervised subspace clustering via non-negative low-rank representation. IEEE Trans. Cybern. 2016, 46, 1828–1838. [Google Scholar] [CrossRef] [PubMed]
  14. Peng, Y.; Ganesh, A.; Wright, J.; Xu, W.; Ma, Y. Rasl: Robust alignment by sparse and low-rank decomposition for linearly correlated images. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2233–2246. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, G.; Lin, Z.; Yu, Y. Robust subspace segmentation by low-rank representation. In Proceedings of the International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 663–670. [Google Scholar]
  16. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [PubMed]
  17. Zhang, T. Analysis of multi-stage convex relaxation for sparse regularization. J. Am. Stat. Assoc. 2010, 11, 1081–1107. [Google Scholar] [CrossRef]
  18. Rao, G.; Peng, Y.; Xu, Z. Robust sparse and low-rank matrix decomposition based on S1/2 modeling. Sci. China Inform. Sci. 2013, 43, 733–748. [Google Scholar] [CrossRef]
  19. Lin, Z.; Liu, R.; Su, Z. Linearized alternating direction method with adaptive penalty for low-rank representation. In Proceedings of the Annual Conference on Neural Information Processing Systems, Granada, Spain, 12–17 December 2011; pp. 612–620. [Google Scholar]
  20. Zhuang, L.; Gao, H.; Lin, Z.; Ma, Y.; Zhang, X.; Yu, N. Non-negative low-rank and sparse graph for semi-supervised learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2328–2335. [Google Scholar]
  21. Lu, C.; Tang, J.; Yan, S.; Lin, Z. Generalized non-convex non-smooth low-rank minimization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4130–4137. [Google Scholar]
  22. Peng, X.; Yu, Z.; Yi, Z.; Tang, H. Constructing the l2-graph for robust subspace learning and subspace clustering. IEEE Trans. Cybern. 2017, 47, 1053–1066. [Google Scholar] [CrossRef] [PubMed]
  23. Elhamifar, E.; Vidal, R. Sparse subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 2790–2797. [Google Scholar]
  24. Tang, J.; Hong, R.; Yan, S.; Chua, T.-S.; Qi, G.-J.; Jain, R. Image annotation by kNN-sparse graph-based label propagation over noisily tagged web images. IEEE Trans. Intell. Trans. Syst. 2011, 2, 14. [Google Scholar] [CrossRef]
  25. Wang, Y.; Zhang, W.; Wu, L.; Lin, X.; Fang, M.; Pan, S. In Iterative views agreement: An iterative low-rank based structured optimization method to multi-view spectral clustering. In Proceedings of the International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 2153–2159. [Google Scholar]
  26. Peng, X.; Tang, H.; Zhang, L.; Yi, Z.; Xiao, S. A unified framework for representation-based subspace clustering of out-of-sample and large-scale data. IEEE Trans. Neural Netw. Learn Syst. 2016, 27, 2499–2512. [Google Scholar] [CrossRef] [PubMed]
  27. Peng, X.; Lu, J.; Yi, Z.; Yan, R. Automatic subspace learning via principal coefficients embedding. IEEE Trans. Cybern. 2016, PP, 1–14. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, Y.; Lin, X.; Wu, L.; Zhang, W.; Zhang, Q.; Huang, X. Robust subspace clustering for multi-view data by exploiting correlation consensus. IEEE Trans. Image Process. 2015, 24, 3939–3949. [Google Scholar] [CrossRef] [PubMed]
  29. Xu, Z.B.; Chang, X.Y.; Xu, F.M.; Zhang, H. l1/2 regularization: A thresholding representation theory and a fast solver. IEEE Trans. Neural Netw. 2012, 23, 1013–1027. [Google Scholar] [CrossRef]
  30. Zuo, W.M.; Meng, D.Y.; Zhang, L.; Feng, X.C.; Zhang, D. A generalized iterated shrinkage algorithm for non-convex sparse coding. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 217–224. [Google Scholar]
  31. Trzasko, J.; Manduca, A. Highly undersampled magnetic resonance image reconstruction via homotopic l0-minimization. IEEE Trans. Med. Image 2009, 28, 106–121. [Google Scholar] [CrossRef] [PubMed]
  32. Fan, J.Q.; Li, R.Z. Variable selection via non-concave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  33. Friedman, J.H. Fast sparse regression and classification. Int. J. Forecast. 2012, 28, 722–738. [Google Scholar] [CrossRef]
  34. Geman, D.; Yang, C. Nonlinear image recovery with half-quadratic regularization. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 4, 932–946. [Google Scholar] [CrossRef] [PubMed]
  35. Gao, C.; Wang, N.; Yu, Q.; Zhang, Z. A feasible nonconvex relaxation approach to feature selection. In Proceedings of the Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011; pp. 356–361. [Google Scholar]
  36. Zhang, C.H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef]
  37. Mohan, K.; Fazel, M. Iterative reweighted algorithms for matrix rank minimization. J. Mach. Learn. Res. 2012, 13, 3441–3473. [Google Scholar] [CrossRef]
  38. Kong, D.; Zhang, M.; Ding, C. Minimal shrinkage for noisy data recovery using Schatten p-norm objective. In Proceedings of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, Prague, Czech Republic, 23–27 September 2013; pp. 177–193. [Google Scholar]
  39. Hu, Y.; Zhang, D.; Ye, J.; Li, X.; He, X. Fast and accurate matrix completion via truncated nuclear norm regularization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2117–2130. [Google Scholar] [CrossRef] [PubMed]
  40. Shen, Y.; Wen, Z.; Zhang, Y. Augmented lagrangian alternating direction method for matrix separation based on low-rank factorization. Optim. Methods Softw. 2014, 29, 239–263. [Google Scholar] [CrossRef]
  41. Soltanolkotabi, M.; Candes, E.J. A geometric analysis of subspace clustering with outliers. Ann. Stat. 2012, 40, 2195–2238. [Google Scholar] [CrossRef]
  42. Zhou, W.; Li, P.; Wang, X.; Li, F.; Liu, H.; Zhang, R.; Ma, T.; Liu, T.; Guo, D.; Yao, D. lp norm spectral regression for feature extraction in outlier conditions. In Proceedings of the IEEE Conference on Digital Signal Processing, Singapore, 21–24 July 2015; pp. 535–538. [Google Scholar]
  43. Nie, F.; Wang, H.; Cai, X.; Huang, H.; Ding, C. Robust matrix completion via joint Schatten p-norm and lp-norm minimization. In Proceedings of the IEEE International Conference on Data Mining, Brussels, Belgium, 10–13 December 2012; pp. 566–574. [Google Scholar]
  44. Cai, J.F.; Candes, E.J.; Shen, Z.W. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  45. Favaro, P.; Vidal, R.; Ravichandran, A. A closed form solution to robust subspace estimation and clustering. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1801–1807. [Google Scholar]
  46. Baker, S.; Sim, T.; Bsat, M. The CMU pose, illumination, and expression database. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1615–1618. [Google Scholar] [CrossRef]
  47. Geusebroek, J.-M.; Burghouts, G.J.; Smeulders, A.W. The Amsterdam library of object images. Int. J. Comput. Vis. 2005, 61, 103–112. [Google Scholar] [CrossRef]
  48. Hull, J.J. A database for handwritten text recognition research. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 550–554. [Google Scholar] [CrossRef]
  49. Georghiades, A.S.; Belhumeur, P.N.; Kriegman, D.J. From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 643–660. [Google Scholar] [CrossRef]
  50. Zhuang, L.; Gao, S.; Tang, J.; Wang, J.; Lin, Z.; Ma, Y.; Yu, N. Constructing a non-negative low-rank and sparse graph with data-adaptive features. IEEE Trans. Image Process. 2015, 24, 3717–3728. [Google Scholar] [CrossRef] [PubMed]
  51. Cai, D.; He, X.F.; Han, J.W. Document clustering using locality preserving indexing. IEEE Trans. Knowl. Data Eng. 2005, 17, 1624–1637. [Google Scholar] [CrossRef]
Figure 1. Typical examples of the CMU-PIE dataset.
Figure 1. Typical examples of the CMU-PIE dataset.
Sensors 17 01633 g001
Figure 2. Typical examples of the COIL20 dataset.
Figure 2. Typical examples of the COIL20 dataset.
Sensors 17 01633 g002
Figure 3. Typical examples of USPS dataset.
Figure 3. Typical examples of USPS dataset.
Sensors 17 01633 g003
Figure 4. Typical sample images of Extended Yale B dataset.
Figure 4. Typical sample images of Extended Yale B dataset.
Sensors 17 01633 g004
Figure 5. Typical sample images of the corrupted dataset.
Figure 5. Typical sample images of the corrupted dataset.
Sensors 17 01633 g005
Table 1. Segmentation results of proposed and existing algorithms on the CMU-PIE dataset.
Table 1. Segmentation results of proposed and existing algorithms on the CMU-PIE dataset.
Clusters KAC (%)NMI (%)
K-MeansPCASSCLRRNNLRSlpSpSSK-MeansPCASSCLRRNNLRSlpSpSS
448.552.410010010010064.166.8100100100100
1241.947.381.589.57991.163.462.984.695.596.796.8
2038.836.780.681.388.392.262.358.885.990.894.398.5
2835.734.978.277.487.991.961.761.186.489.99496.9
3634.334.777.168.878.787.660.560.685.782.393.696.5
4433.833.775.171.781.384.259.162.684.784.293.895.7
5233.133.769.971.175.488.258.161.685.384.993.196
603333.268.165.679.684.752.753.684.980.293.595.4
683132.866.765.186.288.146.846.785.579.387.996.6
Average36.737.774.773.884.189.858.759.485.487.493.496.9
Table 2. Segmentation results of proposed and existing algorithms on the COIL20 dataset.
Table 2. Segmentation results of proposed and existing algorithms on the COIL20 dataset.
Clusters KAC (%)NMI (%)
K-meansPCASSCLRRNNLRSlpSpSSK-MeansPCASSCLRRNNLRSlpSpSS
288.388.196.290.398.398.780.481.488.790.892.292.6
484.78481.588.896.19877.578.784.687.387.791.2
674.583.480.683.994.695.172.373.885.986.289.190.6
873.871.178.27786.992.575.374.886.486.890.188.5
1071.269.477.174.987.289.174.174.885.785.989.591.9
1268.868.575.170.386.887.975.475.684.786.89089.5
1465.266.369.966.584.986.674.17585.385.688.692.1
1666.467.368.16785.587.774.874.684.984.188.991.7
1863.565.867.165.883.184.674.974.784.484.688.891.1
2062.864.366.76478.881.775.874.185.586.490.191
Average71.972.876.174.988.290.275.575.885.686.589.591
Table 3. Segmentation results of proposed and existing algorithms USPS handwritten digit dataset.
Table 3. Segmentation results of proposed and existing algorithms USPS handwritten digit dataset.
Clusters KAC (%)NMI (%)
K-MeansPCASSCLRRNNLRSlpSpSSK-MeansPCASSCLRRNNLRSlpSpSS
294.194.394.294.696.698.371.972.281.473.879.382.9
388.188.889.389.394.796.471.171.479.875.980.383.4
482.279.283.38490.191.267.168.279.472.377.380.3
579.178.279.180.887.388.36566.777.970.679.281.1
677.474.375.275.188.390.265.166.776.373.675.779.1
774.873.374.275.682.784.362.763.281.469.874.575
871.571.874.376.380.682.461.363.479.868.973.275.1
968.769.275.37579.480.659.960.279.467.372.774.7
1065.463.374.274.375.177.459.460.777.366.671.672.6
Average77.976.979.980.686.187.764.865.979.2717678.2
Table 4. Segmentation results of proposed and existing algorithms on Extended Yale B dataset with multiple block occlusions.
Table 4. Segmentation results of proposed and existing algorithms on Extended Yale B dataset with multiple block occlusions.
Block SizeAC (%)NMI (%)
K-MeansPCASSCLRRNNLRSlpSpSSK-MeansPCASSCLRRNNLRSlpSpSS
5 × 513.615.178.588.389.590.614.216.179.090.491.592.8
10 × 1011.713.475.486.787.088.712.614.477.388.989.291.2
15 × 159.811.372.784.585.186.810.812.875.486.387.488.6
20 × 207.59.670.282.182.884.68.510.672.784.485.886.2
Table 5. Segmentation results of proposed and existing methods on Extended Yale B with pixel corruptions.
Table 5. Segmentation results of proposed and existing methods on Extended Yale B with pixel corruptions.
Corruption RateAC (%)NMI (%)
K-MeansPCASSCLRRNNLRSlpSpSSK-MeansPCASSCLRRNNLRSlpSpSS
0.0511.213.168.883.986.988.314.217.174.389.990.892.3
0.17.69.464.578.381.483.611.613.468.387.487.689.1
0.155.88.862.472.276.279.27.89.865.482.383.984.5
0.23.55.660.768.472.875.24.56.662.777.478.879.2

Share and Cite

MDPI and ACS Style

Cheng, W.; Zhao, M.; Xiong, N.; Chui, K.T. Non-Convex Sparse and Low-Rank Based Robust Subspace Segmentation for Data Mining. Sensors 2017, 17, 1633. https://doi.org/10.3390/s17071633

AMA Style

Cheng W, Zhao M, Xiong N, Chui KT. Non-Convex Sparse and Low-Rank Based Robust Subspace Segmentation for Data Mining. Sensors. 2017; 17(7):1633. https://doi.org/10.3390/s17071633

Chicago/Turabian Style

Cheng, Wenlong, Mingbo Zhao, Naixue Xiong, and Kwok Tai Chui. 2017. "Non-Convex Sparse and Low-Rank Based Robust Subspace Segmentation for Data Mining" Sensors 17, no. 7: 1633. https://doi.org/10.3390/s17071633

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop