Next Article in Journal
Stop Band Continuous Profile Filter in Empty Substrate Integrated Coaxial Line
Previous Article in Journal
Experimental Verification of the Dynamic Model of Turbine Blades Coupled by a Sealing Strip
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sparse and Low-Rank Subspace Data Clustering with Manifold Regularization Learned by Local Linear Embedding

1
Business Administration School, North China Electric Power University, Beijing 102206, China
2
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
3
Traffic Control Technology Co., Ltd., Beijing 100070, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2018, 8(11), 2175; https://doi.org/10.3390/app8112175
Submission received: 4 October 2018 / Revised: 21 October 2018 / Accepted: 2 November 2018 / Published: 6 November 2018

Abstract

:
Data clustering is an important research topic in data mining and signal processing communications. In all the data clustering methods, the subspace spectral clustering methods based on self expression model, e.g., the Sparse Subspace Clustering (SSC) and the Low Rank Representation (LRR) methods, have attracted a lot of attention and shown good performance. The key step of SSC and LRR is to construct a proper affinity or similarity matrix of data for spectral clustering. Recently, Laplacian graph constraint was introduced into the basic SSC and LRR and obtained considerable improvement. However, the current graph construction methods do not well exploit and reveal the non-linear properties of the clustering data, which is common for high dimensional data. In this paper, we introduce the classic manifold learning method, the Local Linear Embedding (LLE), to learn the non-linear structure underlying the data and use the learned local geometry of manifold as a regularization for SSC and LRR, which results the proposed LLE-SSC and LLE-LRR clustering methods. Additionally, to solve the complex optimization problem involved in the proposed models, an efficient algorithm is also proposed. We test the proposed data clustering methods on several types of public databases. The experimental results show that our methods outperform typical subspace clustering methods with Laplacian graph constraint.

1. Introduction

Data clustering or segmentation is an active research topic in data mining, signal processing and unsupervised learning [1,2]. In practice, many data are generated from a space which has intrinsic subspace structure, i.e., the space is composed of union of multiple subspaces. In this case, it is necessary to exploit and reveal the subspace structure, and so different subspace clustering methods were proposed in recent years [3,4]. In all of the subspace clustering methods, the spectral clustering methods based on the affinity matrix of data are considered to have good prospects. The typical methods are the Sparse Subspace Clustering (SSC) [5] and the Low Rank Representation (LRR) methods [6]. The key problems of the subspace spectral clustering are how to find the proper data representation and how to construct the affinity or similarity matrix of the data for spectral clustering. In SSC and LRR methods, the original data are firstly represented by the self expression model with sparse or low-rank constraints. Then the affinity matrix is constructed from the representation coefficients. Using the affinity matrix as input, the final clustering results can be obtained by common spectral clustering algorithms, such as K-means or NCut methods.
It is critical for the spectral subspace clustering methods to construct a proper affinity matrix; one should exploit the data intrinsic properties to form necessary regularization for the data representation model. Based on the assumption that the data sample could only be represented by the samples coming from the same subspace it belongs to, SSC adds sparse constraint onto the representation coefficients matrix by l 1 norm [7]. However, the sparse constraint in SSC is an individual constraint for each datum representation. There is no consideration of correlation among the data. For this purpose, LRR adds the holistic constraint of low rank onto the representation coefficients matrix by using nuclear norm · * , which has been proved to be more helpful to reveal the subspace structure of the data drawn from the true multiple subspaces combined space [6].
Although SSC and LRR clustering methods have shown good clustering performance, the data representation is just formulated by sparse or low rank constraints. The other properties of the data or its representation are not well exploited. For example, in the real world, many high dimensional data are often considered to reside in low-dimensional manifolds space and to have non-linear geometrical structure. The sparse manifold clustering and embedding method [8] showed that introducing the local manifold structure can enhance the clustering performance for the data with non-linear property. However, the basic LRR and SSC methods do not take into account this manifold structure for high dimensional data. As a result, some non-linear metric related properties within data may be corrupted in the procedure of clustering. Observing that, numerous researchers have revised the basic SSC and LRR methods to model the data intrinsic properties by introducing extra constraint or regularization. Zheng et al. [9] proposed a graph regularized sparse coding method to obtain data sparse representations with the constraint of local manifold structure. Gao et al. [10] proposed two types of Laplacian regularizer by incorporating a similarity preserving term into the sparse representation model. Cai et al. [11] developed a graph-based approach for non-negative matrix factorization in order to represent the naive geometric structure in data. He et al. [12] proposed a Laplacian regularized Gaussian mixture model for data clustering by exploiting the probability distribution of the intrinsic manifold structure of data. Incorporating manifold regularization, Zhang et al. [13] proposed a novel low rank matrix factorization model for data representation. Hu et al. [14] proposed a SMooth Representation (SMR) model to enforce grouping effect in the data representation model. Liu et al. [15] proposed a Laplacian regularized LRR (LapLRR) to enhance low-rank subspace clustering by utilizing manifold property of data. Peng at el. [16] improved the low-rank representation via sparse manifold adaption for semi-supervised learning. Yin et al. [17] proposed a Non-negative Sparse Laplacian regularized LRR model (NSLLRR) to improve the performance of LRR. Recently, Wang et al. [18] proposed a LRR model on Grassmann manifold with Laplacian graph constraint. Yin et al. [19] proposed an adaptive way to construct the affinity matrix for eliminating the interference of noise in data. Except for the above Laplacian graph constraints, other constraints like the block-diagonal form of the affinity matrix [20,21,22] and the covariance of the data [23] are also exploited to improve subspace clustering performance.
In spite of obtaining improvement by utilizing the non-linear property hiding in the high dimensional data, most of the current methods justly represent this non-linear property by the Laplacian graph constructed directly from the similarity of the raw data under Euclidean distance, which is hard to fulfill modeling the intrinsic manifold structure. Additionally, these methods are easily biased by the noise and outliers existing in the raw data. Facing these problems, we think that the manifold learning methods, such as Local Linear Embedding (LLE) [24], Isomap [25] and Laplacian Eigenmaps (LE) [26], etc. give a possible solution, as they adopt a data-driven manner to learn and represent the non-linear structure of data and show robustness to noise or outliers. So we try to integrate manifold learning methods and the subspace clustering methods, and construct a subspace clustering method for high dimensional data based on manifold learning. Here, we adopt the classic LLE manifold learning method to learn the non-linear structure in the data and embed it into SSC and LRR subspace clustering procedure, thus we propose two new subspace clustering methods, namely LLE-SSC and LLE-LRR. The main contributions of this paper are listed as follows:
  • A new framework of clustering is constructed by integrating the manifold learning methods with subspace clustering methods.
  • Instead of building the Laplacian graph from the raw data, the manifold learning method, LLE, is adopted to reveal the non-linear property of the data for clustering, and this strategy is considered superior in dealing with noise and outliers.
  • To solve the complicated optimization problem involved in the proposed LLE-SSC and LLE-LRR models, an efficient algorithm is determined as the solution.
The rest of the paper is organized as follows. In Section 2, we summarize the SSC and LRR subspace clustering methods and review the related works of our model. In Section 3, our proposed models, LLE-SSC and LLE-LRR, are described in detail and their solutions are given. In Section 4, the performance of the proposed method is evaluated on several public datasets. Finally, conclusions and future work are provided in Section 5.

2. Related Works

Our work is based on the classic SSC and LRR subspace methods, which make impressive progression and have state of the art clustering performance. In the following, we give a review of these two methods and discuss some related subspace clustering works.
The SSC and LRR subspace clustering methods are typical spectral clustering methods, which generally have two steps to implement clustering. First, the raw data are represented over a dictionary for obtaining the data representation coefficients and constructing the affinity matrix; Second, based on the affinity matrix, the clustering is implemented by using common spectral clustering methods, e.g., K-mean or NCut. As the dictionary training procedure usually has a high cost of computation for large datasets, an alternative scheme is to use the data itself as the dictionary, known as the self-expression property of data [5]. The first step in SSC and LRR is the core step as the affinity matrix has significant influence for the clustering result. So, many works focus on constructing an optimal affinity matrix by exploiting different constraints or regularizations. In SSC and LRR methods, the sparse and low rank constraints are added for the data representation matrix, respectively.
For convenience, we denote the data for clustering by X = [ x 1 , , x n ] , which are assumed being generated from k subspaces, i.e., X = j = 1 k X j s.t. j = 1 k X j = ϕ and x i X , X j * , x i X j * . The objective of subspace clustering is to find the proper subspace for each datum. For this purpose, SSC implements clustering by representing the data as the sparse coefficients. The original SSC model is usually formulated as follows [5]:
min Z X X Z F 2 + λ Z 1 , s . t . d i a g ( Z ) = 0 ,
where Z is the sparse representation coefficient matrix and λ is the weight to balance the data self representation error and the sparse representation term.
In SSC, the sparse constraint is just applied for the coefficient of each datum. However, the correlation among the whole dataset or its sparse representation coefficients are not described. Based on this observation, to further reveal the holistic sparsity of the data, LRR proposed using low rank constraint for the coefficient matrix. Additionally, to solve the NP-hard optimization problem with low rank constraint, the nuclear norm · * is generally used to replace the low rank constraint. Thus the LRR model can be formulated as the following optimization model [6]:
min Z X X Z 2 , 1 + λ Z * ,
where · 2 , 1 norm is used to obtain robustness for noises and outliers.
The above SSC and LRR models only use sparse and low rank constraint to the data coefficients, while the other data intrinsic property, such as the non-linear structure of high dimensional data, is not considered. To reveal the non-linear property of high dimensional data, some researchers proposed to use the Laplacian graph constraint to model the local manifold structure of the data. The SMR method [14] enforces the grouping effect on the data representation matrix by using Laplacian graph constraint as follows:
min Z X X Z F 2 + λ t r ( Z L W Z T ) ,
where Z is the data self representation, and L W is the Laplacian graph matrix generally constructed from the local neighbours of each data sample. For example, from the local neighbourhood with K nearest neighbours of each data sample, a weight matrix W R n × n can be constructed with the element W i j defined as follows:
W i j = 1 if x i N K ( x j ) o r x j N K ( x i ) ; 0 otherwise ,
where N K ( x i ) denotes the set of K nearest neighbors of x i . From the weight matrix W , it is easy to get its Laplacian matrix L W with the following element:
L W ( i , j ) = W ( i , j ) if i j ; l i W ( i , l ) otherwise .
The Laplacian matrix is looked at as a graph which represents the pairwise relation of the data. Intuitively, if two data points x i and x j are close in the intrinsic geometry of raw data, their data representations should preserve this property. So it is natural to combine the sparse in (1) or low rank constraint in (2) with the Laplacian graph constraint in (3) to obtain new clustering models. Thus, the LapLRR method [15] presented a manifold regularization for LRR characterized by adopting the Laplacian graph constraint as follows:
min Z X X Z F 2 + λ 1 Z * + λ 2 2 t r ( Z L W Z T ) s . t . Z 0 ,
where Z 0 demands the elements of Z being nonnegative.
Considered that there usually exist noise and outliers in the raw data, the NSLLRR method [17] adopted sparse constraints both for the data reconstruction error and the data representation. Additionally, the method exploits the Laplacian matrix construction methods in different cases, which results a hyper Laplacian matrix even when the data sampling is insufficient. The NSLLRR model is shown as follows:
min Z X XZ 1 + γ Z * + λ Z 1 + β t r ( ZL h Z T ) ,
where L h is the hyper Laplacian matrix and λ , β and γ are penalty parameters for balancing the regularization terms.
In the above methods, the Laplacian graph constraint is generally constructed from the raw data samples under certain metric, which is easily interfered by the outliers or noise exiting in the original dataset. In this paper, we adopt a new strategy by using the manifold learning method to learn the underlining non-linear structure of the data and use the learned relation of the data to regularize the sparse and low rank representations. The manifold learning methods have been proved successful in representation of high dimensional data and robust to noise [24,25], thus it is considered more suitable to reveal the non-linear property of high dimensional data for clustering, and expected to obtain better clustering performance. In the following section, we will give the proposed methods in detail.

3. SSC and LRR Clustering with LLE Regularization

In this section, we first construct our LRR and SSC models with LLE regularization, i.e., LLE-SSC and LLE-LRR, then we give the models solution in detail.

3.1. The Proposed LLE-SSC and LLE-LRR Models

LLE is a classic manifold learning method for dimension reduction, which can discover non-linear structure in high dimensional data by exploiting the data local geometry by linear reconstruction. Additionally, LLE maps its input data into a single global coordinate system of lower dimensionality and does not involve local minima by generating highly non-linear embedding operation. For the given dataset X , LLE is to find the optimal weight matrix W to minimize the following sum error of the data samples which are linearly represented by their neighbours:
min W i j i = 1 n x i x j N ( x i ) W i j x j 2 s . t . x j N ( x i ) W i j = 1 ,
where N ( x i ) is the neighbourhood of x i , and W i j = 0 if x j N ( x i ) . For convenience, the above objective function can be formulated as matrix form as follows:
i = 1 n x i x j N ( x i ) W i j x j 2 = t r ( X ( I W ) T ( I W ) X T ) .
To this minimization problem, there exist several efficient solutions, such as the method in [24]. Once obtaining the solution, denoted by W * , which is regarded as the matrix encoding the manifold structure of the data, we can use W * to construct the regularization for SSC and LRR. The argument is that the learned manifold property of the data X is also valid to its sparse or low rank representations Z . In other words, if we replace X with Z , the error in (8) will be still minimal as they share the same manifold structure.
If we assume
L M = ( I W * ) T ( I W * ) ,
then we can obtain the LLE regularization as follows:
min Z t r ( Z L M Z T ) .
Integrating the above regularization with SSC and LRR models, we naturally obtain our final LLE-SSC and LLE-LRR models as follows:
min Z Z 1 + λ 1 2 X X Z F 2 + λ 2 t r ( Z L M Z T ) ,
min Z Z * + λ 1 2 X X Z F 2 + λ 2 t r ( Z L M Z T ) .
In practice, to avoid numerical instability, L M is generally added a disturbed value to enforce L M to be strictly positive definite, i.e., L M is replaced by L M = L M + ε I where I is the identity matrix and 0 < ε 1 , the same as that in [14].

3.2. Solutions to LLE-SSC and LLE-LRR

For the problems in (12) and (13), we use the Alternating Direction Method of Multipliers (ADMM) [27] to find the solutions. To deal with the sparse item and the nuclear norm item of Z in (12) and (13), we separate the term of Z by introducing an auxiliary S . Let S = Z , then the problems in (12) and (13) can be re-written as the following problems with the Augmented Lagrangian items for the introduced constraints,
min Z , S S 1 + λ 1 2 X XZ F 2 + λ 2 t r ( ZL M Z T ) + G , Z S + γ 2 Z S F 2 ,
min Z , S S * + λ 1 2 X XZ F 2 + λ 2 t r ( ZL M Z T ) + G , Z S + γ 2 Z S F 2 ,
where G is the Lagrangian multiplier and γ is the increasing weight parameter enforcing S = Z . The problems in (14) and (15) can be solved by the following sub-problems with respect to Z , S in an alternative manner by updating one when fixing the other.
(1) Fixing S , solve for Z   
When S is fixed, the objective functions in the problems in (14) and (15) will be transformed to the same objective function as follows:
f ( Z ) = λ 1 2 X XZ F 2 + λ 2 t r ( ZL M Z T ) + G , Z S + γ 2 Z S F 2 .
As f ( Z ) is a quadratic function with respect to Z , we set the derivative of f ( Z ) to zero and have
( λ 1 X X T + γ I ) Z + λ 2 ZL M = λ 1 X T X + γ S G .
Then vectorizing the matrix Z , we can transform the above equation into the following linear equation and thus obtain the solution of Z .
I ( λ 1 XX T + γ I ) + λ L M I v e c ( Z ) = v e c ( λ 1 X T X + γ S G ) .
(2) Fixing Z , solve for S   
When Z is fixed, the LLE-SSC and LLE-LRR optimization problems in (14) and (15) degenerate as the following problems:
min S S 1 + G , Z S + γ 2 Z S F 2 ,
min S S * + G , Z S + γ 2 Z S F 2 .
Let P = Z + G γ , the above problems have the following equivalent forms:
min S S 1 + γ 2 Z P F 2 ,
min S S * + γ 2 Z P F 2 .
For the problem in (21), it has a closed-form solution [28] given by
S = s i g n ( P ) m a x { | P | 1 γ , 0 } .
For the problem in (22), it also has a closed-form solution [29] given by
S = U Q [ Σ ] V T ,
where U Σ V T is the singular value decomposition (SVD) of P . Q [ · ] is the soft-thresholding operator defined on the diagonal element Σ i i of Σ as follows:
Q [ Σ i i ] = Σ i i 1 , if Σ i i > 1 ; Σ i i + 1 , if Σ i i < 1 ; 0 , otherwise .
Once the above two steps are solved, the solutions to the proposed LLE-SSC and LLE-LRR models will be obtained. The complete algorithm for the data representation is summarized in Algorithm 1. It is denoted that the convergence of the ADMM method could not be guaranteed as being discussed in [30]. However, our following experiments show that the algorithm always converges. As to the complexity of Algorithm 1, the main computational cost of the algorithm is determined by the step 2, 3, 5 and 6. For the dataset X with n sample vectors in the dimension of D, the step 2 is to get the LLE weights W * by solving (8) and its complexity is O ( D n 2 + ( D + K ) K 2 n ) , where K is the maximal neighbours of the data samples. The calculation of L M in the step 3 has the complexity of O ( n 2 ) according to (10). Step 5 is to solve the data representation matrix Z by the linear equation in (18), which has the complexity of O ( T 0 ( n 2 ) 2 ) , where T 0 is the maximal iterations of (18). The complexity of step 6 is different for LLE-SSC and LLE-LRR methods; the former has the complexity of O ( n 2 ) and the latter has the complexity of O ( r n 2 ) , where r is the estimated rank of the data representation matrix S or Z . Inductively, we obtain the final complexity of Algorithm 1, i.e.,  O ( T T 0 n 4 + ( D + T + 1 ) n 2 + ( D + K ) K 2 n ) for LLE-SSC and O ( T T 0 n 4 + ( D + T r + 1 ) n 2 + ( D + K ) K 2 n ) for LLE-LRR, respectively.
Algorithm 1 Solution to the data representation for LLE-SSC and LLE-LRR.
Input: 
The dataset X , maximal iteration number T, parameters λ 1 , λ 2 , γ .
Output: 
The data representation Z .
1:
Initialization: Z 0 = 0 , S 0 = 0 , Z = S , t = 0 , G = I , ρ = 1.3 .
2:
Obtain W * by solving (8);
3:
Calculate L M by (10);
4:
while t < T do
5:
  Calculate Z t by (18);
6:
  If LLE-SSC, calculate S t by (23);
7:
  If LLE-LRR, calculate S t by (24);
8:
  Update G G + γ ( Z t S t ) ;
9:
  Update γ ρ γ ;
10:
   Update t t + 1 ;
11:
end while
Based on the above LLE-SSC anf LLE-LRR data representation algorithm, we can finally obtain the subspace clustering algorithm as shown in Algorithm 2.
Algorithm 2 Subspace clustering based on LLE-SSC and LLE-LRR.
Input: 
The data for clustering X .
Output: 
The clustering results of X .
1:
Obtain the sparse representation Z of X by Algorithm 1;
2:
Calculate the affinity matrix W = ( | Z | + | Z T | ) / 2 ;
3:
Calculate the Laplace matrix L W by (5);
4:
Implement NCut ( L W ) to get the final clustering result of X .

4. Experiments

To evaluate the performance of the proposed methods, we implement clustering experiments on several public databases. Firstly, to test the proposed clustering methods in an ideal situation, we use a dataset with true subspace structure for clustering. Here, the synthetic dataset in [31] is adopted, which can be regarded as the baseline dataset for the subspace clustering methods. Then we test the proposed clustering methods on several real world datasets, including the Extended Yale B face dataset [32], the handwritten digit dataset of United States Postal Service (USPS) [33], the image dataset of COIL20 [34] and the motion track dataset, Hopkins155 [35], which are considered challenging for clustering as there is too much variety in these datasets.
The proposed methods, LLE-SSC and LLE-LRR, are compared with several related subspace methods, including the classic SSC and LRR methods, which are considered the baseline methods in subspace clustering methods. Additionally, the proposed methods are also compared with several variations of SSC and LRR, including SMR [14], LapLRR [15] and NSLLRR [17], which introduced other constraints for the data representation, like Laplacian graph, to improve the clustering performance. We summarize the metrics of the data representation error, the constraints for the data representation and the methods for constructing the Laplacian matrix in the related methods and our methods in Table 1.
To evaluate the performance of these clustering methods, the Subspace Clustering Error (SCE) defined in [31] is adopted as the measurement with the following form:
S C E = num . misclssified points total num . of points ,
where “num.misclssified points” represents the number of misclassified samples, and “total num.of points” is the number of total samples.
There are two parameters λ 1 , λ 2 which should be set in the proposed LLE-SSC and LLE-LRR methods. However, for the balancing parameters in optimization models, few effective methods are available to obtain the optimal value, as it is generally dependent on the specific datasets. So to get the proper values of these parameters, we adopt a tuning method based on a set of pre-experiments. Concretely, we change the values of these parameters in an estimated range, and for each selected value, we implement some clustering experiments. By comparing the clustering results with different values of these parameters, we can get the favorable values, though they may be not the global optimal values. In the following experiments, we will report the parameters setting for each dataset. The parameters of the compared methods are set as the same as the recommended setting in the original paper or tuned manually based on some pre-experiments similar to the above setting of ours.

4.1. Synthetic Experiment

In the synthetic experiment, we adopt the similar method in [31] to produce the data samples for clustering. Firstly, we randomly select five samples from the pure infrared hyper spectral mineral data and form a matrix A i = [ a 1 , , a 5 ] R 321 × 5 . Then we produce a random vector w i R 5 by using a uniform distribution and obtain a vector by linear combination x i = A i w i . Let the generation of w i be repeated 10 times, we can obtain 10 data samples from the subspace s p a n ( A i ) , denoted by X i = [ x 1 , , x 10 ] R 321 × 10 . If we repeat the process of A i five times, we finally get a dataset derived from the union of five subspaces, denoted by X = [ X 1 , , X 5 ] R 321 × 50 , which will be used as the dataset for clustering. To further evaluate the robustness of the proposed methods, we add Gaussian noise onto X with three different magnitudes (10%, 20%, 30%) and so we produce three noisy datasets for clustering experiments. To get stable results, the experiments on different datasets are repeated 50 times with the reproduced X . The minimum, maximum, median and mean clustering errors are recorded for complete evaluation. In the above experiments, the parameters λ 1 = 20 , λ 2 = 1 for LLE-SSC and λ 1 = 100 , λ 2 = 1000 for LLE-LRR.
The experimental results are shown in Table 2. In the noise free case, almost all clustering methods perform well; SSC, SMR, LapLRR and our model LLE-LRR have completely correct results. However, when the noise amplitude increases, the clustering performances of the compared methods decrease rapidly, while that of our methods drops slowly. It is shown that compared with the other methods, our LLE-SSC method has at least 10% improvements in metric of median and at least 8% improvements in metric of mean when the noise level is 30%, which highlights that our methods have good robustness to noise.
For the subspace clustering methods, the constructed affinity matrix is critical for clustering. The ideal affinity matrix will be a block diagonal form. Generally, the closer to the block diagonal form it is, the better the clustering results will be. Therefore, we provide a visual comparison of affinity matrices produced by different methods in case of adding 20% Gaussian noise, as shown in Figure 1. It is observed that our proposed methods, LLE-SSC and LLE-LRR, have affinity matrices closer to block diagonal shape than other methods. Additionally, we further show the clustering results by visualizing the samples of the same class with the same color, as shown in Figure 2. It is indicted that our methods, LLE-SSC, LLE-LRR and NSLLRR, have the best clustering results.

4.2. Face Clustering on Extended Yale B

To test the proposed LLE-SSC and LLE-LRR methods in more complicated scenarios, we perform clustering experiments on the wildly used face dataset, the Extended Yale B, which has face images of 38 persons captured under different illuminations with each person having 64 front face images. Some samples of the dataset are shown in Figure 3. For clustering application, this dataset is considered challenging as there exists too much varsity of illumination, expression and pose. Here, we randomly select c = [ 2 , 3 , 5 , 8 , 10 ] people and each person chooses 32 face images for clustering. All images in our experiment are resized to 32 × 32 pixels. As a result, we obtain the initial sample vectors x i R 1024 . Then we reduce the dimension of these samples to (9 × c) by the Principal Component Analysis (PCA) method according to the similar data processing procedure in [20]. Based on the dimension reduced datasets, we implement clustering on the selected c classes. To get a stable result, we repeat the clustering experiments 30 times by re-selecting c classes of face images and use the mean clustering error as the final results. In the experiment, the parameters λ 1 = 20 , λ 2 = 1 for LLE-SSC and λ 1 = 3 , λ 2 = 0.5 c for LLE-LRR.
The experimental results are reported in Table 3. It is observed that our methods significantly outperform the other methods in all cases. Both LLE-SSC and LLE-LRR have at least 10% improvements compared to the other methods in terms of the mean error of different number of classes. In addition, our methods have the advantage that the clustering performances decrease little when the number of classes increases. On the contrary, the clustering performances of other methods drop dramatically when the number of classes increases. This demonstrates that the proposed methods have the application potential for clustering large scale datasets.

4.3. Handwritten Digit Clustering on USPS

In this experiment, we test the clustering performance of our methods on the handwritten digit dataset of USPS. In total, this dataset contains 9298 handwritten digit images of “0” through “9”, each of which is in the size of 16 × 16 pixels, with 256 gray levels per pixel. Some sample images are provided in Figure 4. We represent the digital images as 256-dimensional vectors and randomly select c = [ 2 , 3 , 4 , 5 , 6 , 8 , 7 , 8 , 9 , 10 ] digits for clustering. For each selected digit, 48 images are randomly picked up. We repeat these clustering tests 30 times for each c digits and the mean clustering errors are computed. Here the parameters λ 1 = 800 , λ 2 = 10,000 for LLE-SSC and λ 1 = 1 , λ 2 = 10 for LLE-LRR.
Table 4 reports the clustering results. It is observed that the proposed methods, LLE-SSC and LLE-LRR, almost outperform the compared methods in all cases. Compared with the experimental results on the Extended Yale B face dataset, the improvement of our methods in this experiment is not significant, which is analysed and explained as the handwritten digits written by different people, having too much diversity both in the shape and texture as shown in Figure 4.

4.4. Motion Segmentation on Hopkins155

In this experiment, we implement motion segmentation to find the independent moving rigid objects by their trajectories. In theory, when a rigid object moves in a video, the points of its trajectories will be formulated as a low dimensional subspace. Thus one can find the moving objects by using subspace clustering technique on the tracked trajectories. From this point, we test our clustering methods on the motion benchmark dataset, Hopkins155, which contains three categories of 156 video sequences with each sequence having two or three moving objects, including indoor objects covered by checkerboard texture, outdoor objects like car, and articulated objects like people. The feature points of the objects are extracted and tracked in all the frames. The ground-truth segmentation is also provided in the dataset for comparison purpose. Some sample images of the dataset are provided in Figure 5. In our experiment, the parameters λ 1 = 800 , λ 2 = 10,000 for LLE-SSC and λ 1 = 200 , λ 2 = 10,000 for LLE-LRR.
As there only two or three classes in the experiment, we only report the clustering results in maximal, mean, median and the standard deviation values, as shown in Table 5. On the whole, the performances of all clustering methods are quite good as all methods have clustering errors less than 6% in terms of mean. The reason is two fold. The first is that Hopkins155 is a nearly noise-free dataset. The second is that the number of classes is small in this clustering experiment. It is worth noting that our LLE-LRR method has the best result with the mean error of 1.43%.

4.5. Object Image Clustering on COIL20

In this experiment, we test the clustering performance of our methods on the object image dataset of COIL20. The dataset contains 32 × 32 gray scale images of 20 objects. The object images are generally captured at different views. Some sample images of this dataset are provided in Figure 6. Here we randomly select c = [ 2 , 3 , 4 , 5 , 6 , 8 , 7 , 8 , 9 , 10 , 11 ] objects for clustering. For each selected object, 36 images are randomly picked up. For each value of c, we run 30 times of clustering by randomly choosing different images for each class and the mean error is reported as the final result. The parameters λ 1 = 100 , λ 2 = 1000 for LLE-SSC and λ 1 = 5 , λ 2 = 2000 for LLE-LRR.
The clustering results are shown in Table 6. It is shown that our proposed LLE-SSC and LLE-LRR methods still perform excellently in this experiment. Compared with the other methods, LLE-SSC has better clustering results when the classes number is relatively small, while LLE-LRR has better results when the classes number is large. Particularly, LLE-LRR has the best result of 12.97% in terms of mean error.

5. Conclusions

In this paper, we proposed a new data clustering framework based on manifold learning and subspace clustering. Different from the traditional data subspace methods, which try to utilize the data similarity to construct Laplacian graph constraint for data representation. In the proposed framework, LLE is used to learn the manifold structure hidden in the data. The learned non-linear manifold is taken as a regularization for the subspace clustering methods, SSC and LRR, which results in the proposed LLE-SSC and LLE-LRR clustering methods. To evaluate the proposed clustering methods, we implement several clustering experiments on different types of datasets, such as the synthetic data, face images, handwritten digital images, object images and the motion sequences. The experimental results showed that the proposed methods have good clustering performance compared with related methods on different datasets. Particularly, the proposed methods have two advantages. Firstly, from the results of the experiments on the synthetic data with noise, it is shown that the proposed methods are robust to the noisy data. Secondly, from the clustering results on the Extended Yale B, USPS, COIL20 datasets with the number of the clusters increasing, we can observe that our methods show more stability than the compared methods, which means our methods have the potential for clustering a large scale of datasets. These characteristics make them prospective methods in data clustering.
It is shown that using the manifold learning method to reveal the non-linear property of the data and introducing the learned manifold constraint into subspace clustering methods will improve the data clustering performance. However, in this paper, we only use the common used LLE method to verify this idea by integrating it with the classic subspace clustering methods, SSC and LRR. We believe that the developing manifold learning methods are more helpful in dealing with more complicated data in practice. So one of our future works is to exploit a more suitable manifold learning method and integrate it to improve the data clustering performance. The other future work is to extend our methods by parallel computation technique in order to improve their efficiency and make them practical and suitable for clustering large scale data.

Author Contributions

Conceptualization, methodology, writing review and editing, contributed by Y.H.; Investigation, validation, formal analysis, contributed by Y.Y.; Software, data curation, visualization, writing original draft preparation, contributed by F.W.

Funding

This research was funded by National Natural Science Foundation of China under Grant Nos. 61390510, 61672071, 61632006, 61772048, Beijing Natural Science Foundation Nos. 4172003, 4162010, 4152009, 4184082, and Beijing Municipal Science and Technology Project Nos. Z171100000517003, Z171100000517004, Z171100004417023, Z161100001116072.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fahad, A.; Alshatri, N.; Tari, Z.; Alamri, A.; Khalil, I.; Zomaya, A.Y.; Foufou, S.; Bouras, A. A survey of clustering algorithms for big data: Taxonomy and empirical analysis. IEEE Trans. Emerg. Top. Comput. 2014, 2, 267–279. [Google Scholar] [CrossRef]
  2. Saxena, A.; Prasad, M.; Gupta, A.; Bharill, N.; Patel, O.P.; Tiwari, A.; Er, M.J.; Ding, W.; Lin, C.T. A review of clustering techniques and developments. Neurocomputing 2017, 267, 664–681. [Google Scholar] [CrossRef]
  3. Sim, K.; Gopalkrishnan, V.; Zimek, A.; Cong, G. A survey on enhanced subspace clustering. Data Min. Knowl. Discov. 2013, 26, 332–397. [Google Scholar] [CrossRef]
  4. Deng, Z.; Choi, K.S.; Jiang, Y.; Wang, J.; Wang, S. A survey on soft subspace clustering. Inf. Sci. 2016, 348, 84–106. [Google Scholar] [CrossRef] [Green Version]
  5. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [PubMed]
  6. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust Recovery of Subspace Structures by Low-Rank Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Lerman, G.; Zhang, T. Robust recovery of multiple subspaces by geometric lp minimization. Ann. Stat. 2011, 39, 2686–2715. [Google Scholar] [CrossRef]
  8. Elhamifar, E.; Vidal, R. Sparse manifold clustering and embedding. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2011; pp. 55–63. [Google Scholar]
  9. Zheng, M.; Bu, J.; Chen, C.; Wang, C.; Zhang, L.; Qiu, G.; Cai, D. Graph regularized sparse coding for image representation. IEEE Trans. Image Process. 2011, 20, 1327–1336. [Google Scholar] [CrossRef] [PubMed]
  10. Gao, S.; Tsang, I.W.H.; Chia, L.T. Laplacian sparse coding, hypergraph laplacian sparse coding, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 92–104. [Google Scholar] [CrossRef] [PubMed]
  11. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1548–1560. [Google Scholar] [PubMed]
  12. He, X.; Cai, D.; Shao, Y.; Bao, H.; Han, J. Laplacian regularized gaussian mixture model for data clustering. IEEE Trans. Knowl. Data Eng. 2011, 23, 1406–1418. [Google Scholar] [CrossRef]
  13. Zhang, Z.; Zhao, K. Low-rank matrix approximation with manifold regularization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1717–1729. [Google Scholar] [CrossRef] [PubMed]
  14. Hu, H.; Lin, Z.; Feng, J.; Zhou, J. Smooth representation clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3834–3841. [Google Scholar]
  15. Liu, J.; Chen, Y.; Zhang, J.; Xu, Z. Enhancing low-rank subspace clustering by manifold regularization. IEEE Trans. Image Process. 2014, 23, 4022–4030. [Google Scholar] [CrossRef] [PubMed]
  16. Peng, Y.; Lu, B.L.; Wang, S. Enhanced low-rank representation via sparse manifold adaption for semi-supervised learning. Neural Netw. 2015, 65, 1–17. [Google Scholar] [CrossRef] [PubMed]
  17. Yin, M.; Gao, J.; Lin, Z. Laplacian regularized low-rank representation and its applications. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 504–517. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, B.; Hu, Y.; Gao, J.; Sun, Y.; Yin, B. Laplacian LRR on Product Grassmann Manifolds for Human Activity Clustering in Multicamera Video Surveillance. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 554–566. [Google Scholar] [CrossRef] [Green Version]
  19. Yin, M.; Xie, S.; Wu, Z.; Zhang, Y.; Gao, J. Subspace Clustering via Learning an Adaptive Low-Rank Graph. IEEE Trans. Image Process. 2018, 27, 3716–3728. [Google Scholar] [CrossRef] [PubMed]
  20. Feng, J.; Lin, Z.; Xu, H.; Yan, S. Robust subspace segmentation with block-diagonal prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3818–3825. [Google Scholar]
  21. Wu, F.; Hu, Y.; Gao, J.; Sun, Y.; Yin, B. Ordered subspace clustering with block-diagonal priors. IEEE Trans. Cybern. 2016, 46, 3209–3219. [Google Scholar] [CrossRef] [PubMed]
  22. Lu, C.; Feng, J.; Lin, Z.; Mei, T.; Yan, S. Subspace clustering by block diagonal representation. IEEE Trans. Pattern Anal. Mach. Intell. 2018. [Google Scholar] [CrossRef] [PubMed]
  23. Peng, C.; Kang, Z.; Cheng, Q. Subspace clustering via variance regularized ridge regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 21–26. [Google Scholar]
  24. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef] [PubMed]
  25. Tenenbaum, J.B.; De Silva, V.; Langford, J.C. A global geometric framework for nonlinear dimensionality reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef] [PubMed]
  26. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  27. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  28. Bach, F.; Jenatton, R.; Mairal, J.; Obozinski, G. Convex optimization with sparsity-inducing norms. Optim. Mach. Learn. 2011, 5, 19–53. [Google Scholar]
  29. Cai, J.F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  30. Liu, R.; Lin, Z.; Su, Z. Linearized alternating direction method with parallel splitting and adaptive penalty for separable convex programs in machine learning. In Proceedings of the Asian Conference on Machine Learning, Canberra, ACT, Australia, 13–15 November 2013; pp. 116–132. [Google Scholar]
  31. Tierney, S.; Gao, J.; Guo, Y. Subspace clustering for sequential data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1019–1026. [Google Scholar]
  32. Lee, K.C.; Ho, J.; Kriegman, D.J. Acquiring linear subspaces for face recognition under variable lighting. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 684–698. [Google Scholar] [PubMed] [Green Version]
  33. Hull, J.J. A database for handwritten text recognition research. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 550–554. [Google Scholar] [CrossRef]
  34. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (Coil-20); Columbia University: New York, NY, USA, 1996. [Google Scholar]
  35. Tron, R.; Vidal, R. A benchmark for the comparison of 3-d motion segmentation algorithms. In Proceedings of the 2007 CVPR’07 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1–8. [Google Scholar]
Figure 1. The affinity matrices of the clustering methods of (a) SSC; (b) LLE-SSC; (c) LRR; (d) LLE-LRR; (e) NSLLRR; (f) SMR and (g) LapLRR on the synthetic dataset with 20% Gaussian noise.
Figure 1. The affinity matrices of the clustering methods of (a) SSC; (b) LLE-SSC; (c) LRR; (d) LLE-LRR; (e) NSLLRR; (f) SMR and (g) LapLRR on the synthetic dataset with 20% Gaussian noise.
Applsci 08 02175 g001
Figure 2. The clustering results of the clustering methods of (a) SSC; (b) LLE-SSC; (c) LRR; (d) LLE-LRR; (e) NSLLRR; (f) SMR and (g) LapLRR on the synthetic dataset with 20% Gaussian noise.
Figure 2. The clustering results of the clustering methods of (a) SSC; (b) LLE-SSC; (c) LRR; (d) LLE-LRR; (e) NSLLRR; (f) SMR and (g) LapLRR on the synthetic dataset with 20% Gaussian noise.
Applsci 08 02175 g002
Figure 3. Sample faces images from the Extended Yale B databset.
Figure 3. Sample faces images from the Extended Yale B databset.
Applsci 08 02175 g003
Figure 4. Sample digits images from the United States Postal Service (USPS) dataset.
Figure 4. Sample digits images from the United States Postal Service (USPS) dataset.
Applsci 08 02175 g004
Figure 5. Some sample images from Hopkins 155 dataset. The cross marked points with different colors indicate the positions of the tracks of different motions at the current images.
Figure 5. Some sample images from Hopkins 155 dataset. The cross marked points with different colors indicate the positions of the tracks of different motions at the current images.
Applsci 08 02175 g005
Figure 6. Some object images from the COIL20 dataset.
Figure 6. Some object images from the COIL20 dataset.
Applsci 08 02175 g006
Table 1. The metrics of the data representation error, the constraints for the data representation and the methods for constructing the Laplacian matrix in the related methods and our methods.
Table 1. The metrics of the data representation error, the constraints for the data representation and the methods for constructing the Laplacian matrix in the related methods and our methods.
MethodsMetrics of the Data Representation ErrorConstraints for the Data RepresentationMethods for Constructing the Laplacian Matrix
SSC [5] X XZ F 2 Z 1
LRR [6] X XZ 2 , 1 Z *
SMR[14] X XZ F 2 t r ( ZL W Z T ) L W is constructed from the relations
of the samples and their neighbours.
LapLRR [15] X XZ F 2 Z * and t r ( ZL W Z T ) L W is constructed from the similarities
of the samples and their neighbours.
NSLLRR [17] X XZ 1 Z * , Z 1 and t r ( ZL h Z T ) L h is constructed from multiple
relationships between samples.
LLE-SSC X XZ F 2 Z 1 and t r ( ZL M Z T ) L M is learned by manifold learning
method of LLE.
LLE-LRR X XZ F 2 Z * and t r ( ZL M Z T ) L M is learned by manifold learning
method of LLE.
Table 2. Clustering results on the synthetic hyper spectral mineral dataset with various magnitudes of Gaussian noise, lower is better. Numbers in brackets indicate how many times clustering was perfect, i.e., zero error. The best results are in bold font.
Table 2. Clustering results on the synthetic hyper spectral mineral dataset with various magnitudes of Gaussian noise, lower is better. Numbers in brackets indicate how many times clustering was perfect, i.e., zero error. The best results are in bold font.
Gaussian Noise%The Clustering Error
SSCLRRNSLLRRSMRLapLRRLLE-SSCLLE-LRR
0min0.00% (50)0.00% (49)0.00% (44)0.00% (50)0.00% (50)0.00% (46)0.00% (50)
max0.00%14.00%24.00%0.00%0.00%22.00%0.00%
med0.00%0.00%0.00%0.00%0.00%0.00%0.00%
mean0.00%14.28%0.26%0.00%0.00%0.36%0.00%
10min0.00% (48)0.00% (36)0.00% (49)0.00% (47)0.00% (48)0.00% (50)0% (50)
max2.00%30.00%1.00%2.00%1.00%0.00%0.00%
med0.00%0.00%0.00%0.00%0.00%0.00%0.00%
mean0.08%3.52%0.20%0.12%0.04%0.00%0.00%
20min0.00% (10)0.00% (10)0.00% (13)0.00% (15)0.00% (15)0.00% (32)0.00% (27)
max42.00%38.00%28.00%30.00%26.00%24.00%34.00%
med8.00%10.00%4.00%4.00%4.00%0.00%0.00%
mean10.44%12.80%8.08%8.80%6.32%4.20%3.28%
30min4.00%0.00% (2)0.00% (1)20.00%8.00%0.00% (5)0.00% (6)
max56.00%54.00%44.00%60.00%70.00%42.00%38.00%
med36.00%24.00%24.00%42.00%42.00%14.00%15.00%
mean34.23%25.24%23.06%43.84%38.3614.28%16.08%
Table 3. Clustering results on Extended Yale dataset B. The clustering errors for 2, 3, 5, 8, 10 classes and their mean error are shown here. The best results are in bold font.
Table 3. Clustering results on Extended Yale dataset B. The clustering errors for 2, 3, 5, 8, 10 classes and their mean error are shown here. The best results are in bold font.
Number of ClassesThe Clustering Error (%)
SSCLRRNSLLRRSMRLapLRRSSC-LLELRR-LLE
27.746.873.223.494.092.001.23
310.839.815.426.165.132.911.80
523.8118.0117.019.0510.843.312.78
832.4733.8131.9421.2320.923.533.18
1034.7336.1534.3328.3625.523.683.55
mean21.9220.9318.3813.6613.303.092.51
Table 4. Clustering results on USPS dataset. The clustering errors for 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 classes and their mean error are shown here. The best results are in bold font.
Table 4. Clustering results on USPS dataset. The clustering errors for 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 classes and their mean error are shown here. The best results are in bold font.
Number of ClassesThe Clustering Error (%)
SSCLRRNSLLRRSMRLapLRRLLE-SSCLLE-LRR
24.263.807.554.493.164.743.02
311.457.648.759.777.037.238.34
421.5111.4315.5012.3314.558.979.46
527.3518.37514.7416.7814.5012.3213.17
632.1420.0518.8417.9119.0816.4216.74
734.5520.1319.6119.7420.8320.0017.86
836.4522.2421.5822.9123.0724.1721.33
938.1723.7623.6826.0726.3626.4023.84
1039.3228.3026.0329.5827.2928.9424.16
mean27.2417.7216.8817.7617.3016.4715.49
Table 5. Clustering results on Hopkins 155 dataset. The max value (max), mean value (mean), median value (med) and the standard deviation (std) of the errors are reported on the total 155 motion sequences. The best results are in bold font.
Table 5. Clustering results on Hopkins 155 dataset. The max value (max), mean value (mean), median value (med) and the standard deviation (std) of the errors are reported on the total 155 motion sequences. The best results are in bold font.
SSCLRRNSLRRSMRLapLRRLLE-SSCLLE-LRR
max42.3441.1835.0649.6746.4333.8931.12
mean3.114.833.745.565.361.571.43
med0.000.520.000.190.240.200.21
std7.789.357.0210.4210.413.703.51
Table 6. Clustering results on COIL20 dataset. The clustering errors for 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 classes and their mean error are shown here. The best results are in bold font.
Table 6. Clustering results on COIL20 dataset. The clustering errors for 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 classes and their mean error are shown here. The best results are in bold font.
Number of ClassesThe Clustering Error (%)
SSCLRRNSLLRRSMRLapLRRLLE-SSCLLE-LRR
25.326.382.922.132.503.103.33
310.597.759.486.988.555.126.42
414.0512.5510.1216.8911.487.258.87
521.7015.2915.2017.1814.6110.3710.89
625.3117.8416.2016.0316.7816.1312.38
729.0721.3120.6118.5418.7016.7914.06
825.6322.2124.9121.6416.7516.4214.53
927.5425.3224.8723.6919.2223.9020.93
1033.2127.5225.4024.0921.4618.9317.60
1130.2129.4928.9625.2725.3826.3120.72
mean22.2618.5717.8717.2415.5414.4312.97

Share and Cite

MDPI and ACS Style

Yang, Y.; Hu, Y.; Wu, F. Sparse and Low-Rank Subspace Data Clustering with Manifold Regularization Learned by Local Linear Embedding. Appl. Sci. 2018, 8, 2175. https://doi.org/10.3390/app8112175

AMA Style

Yang Y, Hu Y, Wu F. Sparse and Low-Rank Subspace Data Clustering with Manifold Regularization Learned by Local Linear Embedding. Applied Sciences. 2018; 8(11):2175. https://doi.org/10.3390/app8112175

Chicago/Turabian Style

Yang, Ye, Yongli Hu, and Fei Wu. 2018. "Sparse and Low-Rank Subspace Data Clustering with Manifold Regularization Learned by Local Linear Embedding" Applied Sciences 8, no. 11: 2175. https://doi.org/10.3390/app8112175

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop