Tensor Based Multiscale Low Rank Decomposition for Hyperspectral Images Dimensionality Reduction

: Dimensionality reduction is an essential and important issue in hyperspectral image processing. With the advantages of preserving the spatial neighborhood information and the global structure information, tensor analysis and low rank representation have been widely considered in this ﬁeld and yielded satisfactory performance. In available tensor-and low rank-based methods, how to construct appropriate tensor samples and determine the optimal rank of hyperspectral images along each mode are still challenging issues. To address these drawbacks, an unsupervised tensor-based multiscale low rank decomposition (T-MLRD) method for hyperspectral images dimensionality reduction is proposed in this paper. By regarding the raw cube hyperspectral image as the only tensor sample, T-MLRD needs no labeled samples and avoids the processing of constructing tensor samples. In addition, a novel multiscale low rank estimating method is proposed to obtain the optimal rank along each mode of hyperspectral image which avoids the complicated rank computing. Finally, the multiscale low rank feature representation is fused to achieve dimensionality reduction. Experimental results on real hyperspectral datasets demonstrate the superiority of the proposed method over several state-of-the-art approaches.


Introduction
By collecting hundreds of contiguous narrow spectral bands, hyperspectral images contain wealth of spectral information and have been applied in classification [1,2], target detection [3,4], etc. [5,6].On the other hand, the huge spectral bands number may also lead to the curse of dimensionality which not only increases the storage and computational costs but also degrades the processing performance, so dimensionality reduction is an important issue in hyperspectral images processing [7][8][9].Depending on whether labeled training samples are required or not, the dimensionality reduction algorithms can be classified into two major categories: supervised methods and unsupervised methods.In practice, the cost of hyperspectral image samples labeling is extremely high, so the unsupervised methods have attracted much more attention in hyperspectral image processing.
With the advantage of capturing the global structure of the original hyperspectral images, low rank representation has attracted more and more attention [10][11][12][13][14][15], but the basic low rank representation methods need the vectoring processing to convert the 3-order hyperspectral images to vector samples.The vectoring processing makes the vector samples fail at capturing the spatial neighborhood structure information of hyperspectral images.
Tensor analysis is a multilinear algebra tool which needs no vectoring operation.Tensor analysis has been widely considered in hyperspectral image processing and achieved promising performance [16][17][18][19][20][21].In tensor based methods, the spatial and spectral information are preserved simultaneously by representing hyperspectral images in the form of 3-order tensors.By jointly considering the low rank representation and tensor analysis, amount of low rank tensor based methods have been proposed for hyperspectral images processing [5,[22][23][24][25].In [26], a group tensor based low rank model (GTLR) was proposed for hyperspectral images dimensionality reduction.In GTLR, the low rank and non-local properties are jointly considered to preserve the intrinsic structure of hyperspectral images.In [27], a tensor sparse and low-rank graph-based discriminant analysis (TSLGDA) was proposed.TSLGDA jointly considers the sparse and low rank properties to capture both local and global structure of hyperspectral image.A low-rank tensor recovery (LRTR) model was proposed for the denoising task of hyperspectral images in [28], LRTR can preserve the global structure of hyperspectral images and simultaneously remove Gaussian noise and sparse noise by adopting a new tensor singular value decomposition and tensor nuclear norm.Fu. et al. [29] proposed a tensor low-rank representation and sparse coding-based (TLRRSC) subspace clustering method by simultaneously considering feature information and spatial structures.
Though promising performance has been achieved, tensor based low rank methods face two challenges in the field of hyperspectral image processing: (1) The tensor samples are often selected by splitting the original hyperspectral images in a spatial domain with fixed windows, how to determine the windows size and how to determine the number of training samples are serious problems.These two problems greatly depend on the properties of land-covers and are usually solved experimentally, which is time consuming and may degrade the robustness of the corresponding methods.(2) The low rank property has been proved to be important for hyperspectral images, but how to determine the optimal rank of hyperspectral images is difficult.The available methods usually calculate the optimal rank by converting the rank estimating issue to an iteration optimal problem with some constraints.However, this strategy is also time consuming and there are usually trade-off parameters that are difficult to be tuned [21].
Multiscale based methods have also been applied in hyperspectral images processing [30][31][32][33][34][35].Li et al. [36] proposed a multiscale spatial information fusion (MSIF) method for hyperspectral image classification.MSIF captures intrinsic spatial information contained in homogeneous regions of different sizes by multiscale strategy.Fang et al. [37] proposed a multiscale adaptive sparse representation (MASR) model that exploits spatial information at multiple scales via an adaptive sparse strategy.Zhang et al. [38] proposed a multiscale superpixel-based sparse representation (MSSR) algorithm for hyperspectral images classification.After obtaining multiscale superpixels by a modified segmentation strategy, the joint sparse representation classification is used to classify the multiscale superpixels and majority voting is utilized to fuse the labels of different scale superpixels to obtain the final classification results.In [33], Tong et al. proposed a multiscale union regions adaptive sparse representation (MURASR) model for hyperspectral image classification which makes the union region as the overlap of fixed patch and superpixel.MURASR can make full use of the advantages of MASR and MSSR and overcome the weaknesses of them.In general, the multiscale strategy in the available methods refers to different spatial scales of the original hyperspectral image that does not involve the higher level feature representation of hyperspectral image and may degrade the representation ability of multiscale.In addition, most available multiscale based methods are vector based methods that cannot preserve the important spatial neighborhood information.
To overcome the drawbacks of available methods, an unsupervised tensor based multiscale low rank decomposition (T-MLRD) method for hyperspectral images dimensionality reduction is proposed in this paper.With a novel multiscale rank estimating method that can automatically estimate the optimal multiscale rank along each mode of hyperspectral image, the multiscale low rank decomposition is applied on the raw cube hyperspectral image.Finally, the low rank tensor approximation strategy is employed to fuse the multiscale low rank feature representation and achieve dimensionality reduction.The flowchart of T-LMRD is shown in Figure 1.Compared with the existing approaches, the proposed method has three new contributions.(i) We combine the multiscale strategy and the low rank property together that may obtain a higher level and more comprehensive representation of hyperspectral image.In addition, a novel strategy is proposed to calculate the optimal multiscale rank along each mode of cube hyperspectral image that makes the proposed method more efficient.(ii) Different from the vector based methods, the proposed method is a tensor based method that can preserve the spatial neighborhood information.(iii) By regarding the raw cube hyperspectral as the sole sample, the proposed method avoids the complicated operation of selecting tensor samples.In addition, the proposed method needs no label information, which makes the proposed method more practical.

Definition and Notations
For an N-order tensor X ∈ R I 1 ×I 2 ×...×I N , some basic operations are defined as follows.
Definition 1 (n-mode flattening matrix).The n-mode vector is constructed by varying index i n while keeping the remaining indices fixed.Arranging all the vectors in columns, the resulting matrix is called an n-mode flattening matrix and denoted as (1) Definition 2 (n-mode product).The n-mode product refers to a tensor X ∈ R This operation can also be reformulated in matrix form Definition 3 (Tucker decomposition).Tucker decomposition is defined as where "•" is the outer product of the vectors, is the factor matrix along each mode.

Tensor Low Rank Decomposition
With the merits of preserving global intrinsic structure information, low rank representation has been widely considered in hyperspectral image processing.Let X ∈ R I 1 ×I 2 ×I 3 be a hyperspectral image, where I 1 , I 2 and I 3 are the number of row, column and spectral dimensions, respectively.The low rank decomposition aims at finding an approximation tensor X ∈ R I 1 ×I 2 ×I 3 with rank n ( X ) < rank n (X ), n = 1, 2, 3.The objective function of tensor low rank decomposition can be formulated as min X X − X 2 F , s.t.rank n ( X ) < rank n (X ), (5) where rank n (•) means the rank along mode-n.Lathauwer et al. [39] have proved that, using Tucker decomposition, minimizing Equation (5) with respect to X is equivalent to maximizing Equation (6): It is noted that, given the values of low rank r n along each mode, U n , n = 1, 2, 3 consists of the first r n eigenvectors corresponding to covariance matrix R n R T n , where R n is the n-mode flattening matrix of X .Equation ( 6) can be readily solved by alternating least square (ALS) algorithm, and the solution can be reformulated as where P n = U n U T n , n = 1, 2, 3.It is noted that, with the low rank representation, X may lead to a more compact representation of hyperspectral image, and the low rank representations with different rank r n can capture different level features of the original hyperspectral image.
Figure 2 shows the different scale low rank representations on the Indian Pines dataset.It can be seen from Figure 2 that different scale low rank representations show different level features of the original hyperspectral image.The lower the rank is, the more consistent the extracted feature is.The consistent feature representation may eliminate the effects of noise and obtain a more compact feature representation.On the other hand, if the rank is too small, the discriminative feature of different land covers may also be confused.In practice, it is a challenge to identify which scale is the optimal one for the following classification task.In fact, the issue of optimal rank estimating depends on the number and the consistency of different land covers and often is solved experimentally and empirically in practice.In addition, a sole scale may not offer sufficient intrinsic structure information for the following processing and different scale can offer complementary information.

Adaptive Hypepspectral Image Low Rank Estimating
As discussed above, the rank obtained by tensor low rank decomposition is determined by the number of eigenvectors corresponding to covariance matrix R n R T n .Inspired by the principle of Principal Component Analysis (PCA) [40], the eigenvectors associated with larger eigenvalues are regarded to retain more intrinsic information of the original tensor dataset.By employing the analysis of the eigenvalues correspongding to covariance matrix R n R T n , a novel strategy is proposed here to estimate the rank r n reliably.
Let C be the eigenvalues corresponding to covariance matrix R n R T n , a modified array C can be calculated by By sorting the elements in C in decreasing order, the rank r n can be determined by where T is a given threshold value, and max( C) is the maximum value of C. The corresponding values of C on the Indian Pines dataset along each mode are shown in Figure 3. From Figure 3, it can be seen that the elements in the front of C has large values.It is believed that the eigenvectors corresponding to these elements can preserve the most intrinsic information of the original dataset, which is accordance with the theory of Principal Component Analysis.It can also be seen from Equation ( 9) that the resulting rank depends on the property of given dataset, so the obtained rank is adaptive for different datasets.In addition, for a given dataset, max( C) is a constant, thus the resulting rank r n can be controlled by parameter T. The smaller the T is, the larger the resulting r n is and the more detailed features may be kept.Conversely, the larger the T is, the smaller the resulting r n is and the more common features of the original hyperspectral image may be preserved.That is to say, different scales low rank feature can be extracted with different T.
Taking the Indian Pines dataset as an example, the estimation of optimal multiscale rank r n m is illustrated in Figure 3, where r n m denotes the optimal rank along mode n of scale m, n =

Hyperspectral Image Multiscale Representation and Low Rank Fusion
Given a hyperspectral image X , by setting different threshold values T m , m = 1, 2, • • • , M, where M is the number of different scales.It is easy to compute the resulting rank r n by adopting the rank estimating model.Then, the different scale low rank representation X m , m = 1, 2, • • • , M can be obtained by Equation (7).It is noted that different scale low rank representation exhibits different level structure characteristics of hyperspectral images.
It has been proved that a sole scale may not offer sufficient intrinsic structure information for the following processing and different scales may offer complementary and correlated information [37].Thus, we stack the different level representations together in a spectral domain.For hyperspectral image X ∈ R I 1 ×I 2 ×I 3 , suppose that Xm ∈ R I 1 ×I 2 ×I 3 , m = 1, 2, • • • , M are the ith scale low rank representation.By the stacking all Xi in the spectral domain, the final multiscale low rank representation can be represented as Ỹ ∈ R I 1 ×I 2 ×(I 3 ×M) .
For a hyperspectral image, the distinct structures and characteristics will be exhibited more effectively by the multiscale low rank representations.However, different scale representations usually lie on different scales of feature spaces.The strategy of stacking the multiscale low rank representation in the spectral domain faces two disadvantages.
(i) The different scale low rank representations may play different important roles in a specific application, but the stacking strategy treats the different scale representations equally, which ignores the different contributions of different scale representations.(ii) The stacking strategy increases the feature dimensionality which may suffer from the "curse of dimensionality" and the high computational cost.
To fully extract the intrinsic correlative and complementary information offered by the different scale representations, a low rank fusion and dimensionality reduction strategy is proposed.For a hyperspectral image X ∈ R I 1 ×I 2 ×I 3 , Ỹ ∈ R I 1 ×I 2 ×(I 3 ×M) is the multiscale low rank representation of X .The low rank fusion and dimensionality reduction Y can be achieved by Equation (10): where Pn = Ũn ŨT n , Ũn , n = 1, 2, 3 consists of the first r n eigenvectors corresponding to the covariance matrix Rn RT n , Rn is the n-mode flattening matrix of Ỹ.It is noted that, since Ỹ consists of some low rank characters of X , for the spatial dimension, Ũn , n = 1, 2 consists of all the eigenvectors corresponding to the covariance matrix Rn RT n .For the spectral dimension, since the requirement of fusion and dimensionality reduction, P3 = Λ −1/2 × ŨT 3 , Λ is the k × k eigenvector diagonal matrix corresponding to the first k eigenvalues of covariance matrix Ũn ŨT n ; k is the number of reduced dimensionality.Overall, the flowchart of the proposed T-MLRD is shown in Figure 1.As shown in Figure 1, for each mode of the raw hyperspectral image, the optimal rank is estimated according to the specified N-scale threshold values, then the low rank representation associated with the optimal rank of scale N is obtained by multiscale low rank decomposition.To fully extract the complementary information of different scale representation, the tensor low rank decomposition strategy is adopted to fuse the different scale representation.Finally, the dimensionality reduction is achieved under the Tucker decomposition framework.The detailed steps of the proposed method are given in Algorithm 1.

Experimental Results and Analysis
In this section, experiments are carried out on three real hyperspectral datasets to evaluate the performance of the proposed method.The second dataset named Pavia University was acquired by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor over Pavia University, northern Italy.This is an image of 610 × 340 pixels and 103 spectral bands.The spectral range is from 430 to 860 nm with a geometric resolution of 1.3 m.There are nine classes of land covers in total and its pseudo-color and corresponding ground truth maps are shown in Figure 5a,b, respectively.

Experimental Setup
The third dataset was collected by the AVIRIS sensor over Salinas Valley, California.This dataset has a size of 512 × 217 × 224 and a high spatial resolution of 3.7 m pixels.Salinas groundtruth contains 16 classes and its pseudo-color and corresponding ground truth maps are illustrated in Figure 6a,b, respectively.
(ii) Comparison methods: To evaluate the performance of the proposed method, several low rank and tensor based methods are selected as the comparison methods, including low-rank graph discriminant analysis (LGDA) [12], group tensor based low rank decomposition (GTLR) [26], tensor based low-rank representation (TLRR) [21], low rank tensor approximation(LRTA) [17] and tensor-based low rank graph with multi-manifold regularization(T-LGMR) [21].In addition, the original spectral bands and classical PCA are also considered in this letter.For the proposed method, the number of low rank scale is set as 3, and the corresponding threshold parameters are set as [0.01 0.05 0.1].The Nearest Neighborhood (1NN) classifier and Support Vector Machine (SVM) classifier with Gaussian kernel function are chosen as the classifier to valuate the classification performance of the dimensionality reduced dataset.In addition, 10% labeled samples are randomly selected as the training samples and the remaining as the testing samples.Overall accuracy (OA), Average accuracy (AA), Class-specific Accuracy, Kappa coefficients and the classification maps are reported to show the performance of different methods.To reduce the randomness, each experiment is conducted 10 times and the mean value and the corresponding standard deviation are reported.

Classification Results
In this section, classification experiments are carried out on the real hyperspectral datasets.The classification results are shown in Tables 1-3 and the classification maps are shown in Figures 4-6.
For the Indian Pines dataset, it can be seen from Table 1 that the proposed method achieves much higher classification accuracy than comparison methods.For example, the proposed method achieves about 2% and 3% higher accuracy than the second best method in terms of OA and AA.It is noted that, for class 13, the proposed method achieves almost 100% classification accuracy.Figure 4 shows the classification maps of different methods on Indian Pines dataset.It can be observed from Figure 4 that the proposed method can obtain promising performance for the classes with good spatial consistencymarked by white rectangle).In addition, for small samples classes, such as class 7 and class 9, the proposed method can also achieve satisfactory classification results.
For the Pavia University dataset, the advantages of the proposed method over the comparison methods are also evident.It can be seen from Figure 5 that the proposed method achieves good regional uniformity in the area marked by white rectangle which may be easily corrupted by spot noise.Furthermore, for the ribbon area (marked by the ellipse), the proposed method can also achieve promising performance.
For Salinas dataset, the land covers in this scene all have large sample numbers and good region consistency.All of the comparison methods achieve satisfactory classification performance.It is noted that class 8 and class 15 are easy to confuse (marked by the white ellipse in Figure 5) and there are many mis-classified noises in the classification maps obtained by the vector-based methods (Original, PCA and LGDA).For tensor-based methods (GTLR, TLRR, LRTA, T-LGMR and T-MLRD), by utilizing spatial neighborhod information, the region consistency of these two land-covers is obviously superior to that of vector-based methods.In addition, it can be seen from the classification maps that the proposed method achieves the best classification performance by fusing multiscale low rank structure information.

Analysis of Different Reduced Dimensionality
In this section, experiments are carried out to evaluate the classification performance with different dimensionalities.It can be seen from Figure 7 that the proposed method achieves better performance than the comparison methods when the reduced dimensionality is larger than 15.In addition, the performance of the proposed method can be further improved with the increase of dimensionality while the overall accuracy of the comparison methods may remain stable.This further demonstrates that the proposed method can reduce the dimensionality and preserve rich information of the original dataset.

Analysis of Computational Costs
In this section, experiments are carried out on three hyperspectral images to demonstrate the computational efficiency of the proposed method.All the experiments are accomplished using Matlab R2014a on a PC with Intel Core i5-5490 CPU and 8 GB RAM.The mean time (in terms of seconds) of ten repetitions of all comparison methods is listed in Table 4.In can be seen from Table 4 that the running time of vector samples based methods whose solutions do not need to compute the optimal factor matrix for each mode are shorter than tensor sample based methods.For tensor based methods, due to the iterative strategy of obtaining the optimal rank for each mode, TLRR and T-LGMR cost the longest running time.Since the best rank for each mode are specified directly in GTLR, LRTA and T-LMRD, the cost time of these three methods are almost the same.Compared with the excellent performance, the running time of T-LGMR is acceptable.

Analysis of Different Scales
In previous experiments, the number of multiscales is set as 3, and the corresponding threshold parameters are set as [0.01 0.05 0.1], respectively.In this part, the number of low rank scales is expanded and its effect on the overall accuracy is analyzed.The number of low rank scales is set as 7 and the corresponding threshold parameters space is set as [0.01 0.05 0.1 0.2 0.5 0.8 1.0].The classification results of single scale low rank representation are shown in Figure 8 and the classification results of multiscale low rank representation are reported in Figure 8.
It can be seen from Figure 8 that, for a single scale, the correlation of overall accuracy and parameter T is uncertain, which suggests that the optimal single threshold is difficult to be determined.For example, for the Indian Pine dataset, the overall accuracy associated with the first three thresholds is about the same and the second threshold is the optimal one, but, for Pavia University dataset, the first threshold is the optimal one while the overall accuracy associated with the second and third one decreases rapidly.This further reveals that the threshold parameter T is difficult to be determined in real application.Figure 9 shows the classification results with different sizes of multiscale on the three hyperspectral datasets.There are seven pointed multiscale threshold values of the x axis, i.e., MT1 − MT7, the correspongding multiscale threshold values are listed in the legend table of Figure 9.It is seen that the overall accuracy increases dramatically at the beginning and, when the number of multiscales is larger than 3, the curve increases slowly and then remains stable.Thus, the number of multiscale being set as 3 in the experiments is reasonable.

Analysis of Multiscale Threshold Values
In the experiments above, the number of different scale is set as 3 and the corresponding threshold values are set as 0.01, 0.05 and 0.1.To investigate the effect of the threshold values T, we extend the threshold values space of the three scales and discuss the effect of threshold values on the overall classification accuracy.Specifically, [0.005 0.01 0.015], [0.03 0.05 0.07] and [ 0.08 0.1 0.12] are set as the different threshold values for scale 1, scale 2 and scale 3, respectively.The final multiscale threshold values are constructed by selecting one from the three values for each scale, thus the resulting threshold values space is composed of 27 different threshold values combinations, which are shown in Table 5. Classification experiments are carried out with 27 different multiscale threshold values and the corresponding overall classification accuracies are shown in Figure 10.It can be seen from Figure 10 that the overall accuracy varies slightly within the whole threshold values space.For example, the maximum change of overall accuracy on Salinas is only 0.53.This shows the robustness of the proposed method against the threshold values and further demonstrates the fact that different scales' low rank representation can offer complementary structure information.

Classification Results of Different Reduced Dimensionality
It can be seen from the analysis of different reduced dimensionality that, when the reduced dimensionality is larger than 20, the proposed method achieves the best classification performance of all compared methods, while, when the reduced dimensionality k is less than 15, the accuracy performance of T-LMRD is worse than some of the compared methods.This is mainly due to the stacking strategy of all single low rank representations on a spectral domain possibly leading to a high dimensionality of multiscale low rank representation.Taking the Indian Pines dataset, for example, the original dimensionality is 200 while the dimensionality of multiscale low rank representation is 600 when the number of scale is set as 3.If we extract the feature representation with an extremely low dimensionality, such as 5 or 10, from such a high feature space, the intrinsic information of multiscale representation may not be fully preserved by the resulting low dimensionality feature representation, so the classification performance of the T-MLRD may be worse than the compared methods when the dimensionality is extremely low.Considering the outstanding performance with slightly higher dimensionality, the proposed T-MLRD is still a rather good dimensionality reduction method.

Multiscale Threshold Values
From the analysis of different scales, it can be observed that there are obvious advantages of multiscale over the optimal single scale in terms of classification overall accuracy, which demonstrates the superiority of multiscale analysis.Furthermore, Figures 8 and 9 shows that, when the number of scales is larger than 3, the overall classification accuracies of single scale are extremely low, but, if these single scale features are fused into the multiscale feature space, the resulting performance may also be improved.This further reveals the complementarity of the features obtained from different scales.

Conclusions
In this paper, a novel tensor based multiscale low rank decomposition method was proposed for hyperspectral images dimensionality reduction and classification.With a novel multiscale rank estimating method that can automatically estimate the optimal multiscale rank along each mode of hyperspectral images, the multiscale low rank decomposition is employed to obtain a multiscale low rank representation of original hyperspectral images.Then, the low rank tensor approximation strategy is employed to fuse the multiscale low rank feature representation and achieve dimensionality reduction.Furthermore, as an unsupervised dimensionality reduction method, the proposed T-MLRD is able to extract the intrinsic structure information from original cube hyperspectral images directly, which needs no labeled samples and avoids the complicated tensor training samples construction, which makes the T-MLRD more practical.Experimental results on real hyperspectral images demonstrate the advantages of the proposed method over the state-of-the-art approaches.

Figure 1 .
Figure 1.The schematic illustration of T-MLRD.The proposed T-MLRD consists of four steps: (1) estimate the optimal rank for each mode of the raw hyperspectral image according to the specified N-scale threshold values.(2) Construct the single low rank representation associated to the optimal rank of scale N. (3) Fuse the different scale representation tensor low rank decomposition strategy.(4) The dimensionality reduction is achieved under the Tucker decomposition framework.

Figure 2 .
Figure 2. The low rank representation with different rank along each mode on Indian Pines dataset.(a) original pseudo-color map; (b) rank = 30 along each mode; (c) rank = 10 along each mode; (d) rank = 5 along each mode; (e) rank = 2 along each mode.
, 2, 3 denotes the different mode of tensor datasets, m = 1, 2, • • • , M is the different low rank scales and M is the number of all scales.Assuming the number of multiscale is 3 and the corresponding threshold T are 0.01, 0.05 and 0.1, the estimating multiscale low rank values along each mode are obtained as shown in the nine subimages of Figure3.For each mode, the curve of C is shown in blue, while, for each scale, the values of T max( C) are marked in red lines.The optimal rank shown in the top right corner is obtained by the cross point of the two curves.

Figure 3 .
Figure 3.The estimation of optimal multiscale rank along each mode on Indian Pines.

1 :
Estimate the optimal rank r m n of hyperspectral image along mode n of scale m by Equation (9).2: for m=1:M 3: Calculate scale m low rank representation by Equation (7).4: end for 5: Stack all the M low rank representation in spectral dimension to construct the multiscale low rank representation Ỹ. 6: Calculate the reduced dimensionality dataset Y by Equation (10).OUTPUT: Dimensionality reduced of the hyperspectral dataset Y.
(i) Hyperspectral dataset: The first dataset is the Indian Pines dataset that was gathered by AVIRIS sensor in June 1992.This scene contains two-thirds agriculture, and one-third forest or other natural perennial vegetation that are designated into 16 classes.The size of this dataset consists of 145 × 145 × 224 with the wavelength range of 0.4-2.5 µm and the spatial resolution of 20 m.The spectral bands is reduced to 200 by removing the noisy and water absorption bands.Its pseudo-color and corresponding ground truth maps are shown in Figure 4a,b, respectively.

Figure 4 .
Figure 4. Classification maps of different methods on the Indian Pines dataset.

Figure 5 .
Figure 5. Classification maps of different methods on the Pavia University dataset.

Figure 6 .
Figure 6.Classification maps of different methods on the Salinas dataset.

Figure 8 .
Figure 8. Overall accuracy of single scale low rank representation.

Figure 9 .
Figure 9. Overall accuracy of multiscale low rank representation.

Table 4 .
Computational costs (in seconds) on three experimental datasets.

Table 5 .
Multiscale threshold values space.Overall classification accuracy with different multiscale threshold values.