Next Article in Journal
Spectral Method for Liming Recommendation in Oxisol Based on the Prediction of Chemical Characteristics Using Interval Partial Least Squares Regression
Previous Article in Journal
SHAP-Based Interpretable Object Detection Method for Satellite Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-View Structural Feature Extraction for Hyperspectral Image Classification

School of Informatics and Engineering, Suzhou University, Suzhou 234000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(9), 1971; https://doi.org/10.3390/rs14091971
Submission received: 25 February 2022 / Revised: 4 April 2022 / Accepted: 13 April 2022 / Published: 20 April 2022
(This article belongs to the Section AI Remote Sensing)

Abstract

:
The hyperspectral feature extraction technique is one of the most popular topics in the remote sensing community. However, most hyperspectral feature extraction methods are based on region-based local information descriptors while neglecting the correlation and dependencies of different homogeneous regions. To alleviate this issue, this paper proposes a multi-view structural feature extraction method to furnish a complete characterization for spectral–spatial structures of different objects, which mainly is made up of the following key steps. First, the spectral number of the original image is reduced with the minimum noise fraction (MNF) method, and a relative total variation is exploited to extract the local structural feature from the dimension reduced data. Then, with the help of a superpixel segmentation technique, the nonlocal structural features from intra-view and inter-view are constructed by considering the intra- and inter-similarities of superpixels. Finally, the local and nonlocal structural features are merged together to form the final image features for classification. Experiments on several real hyperspectral datasets indicate that the proposed method outperforms other state-of-the-art classification methods in terms of visual performance and objective results, especially when the number of training set is limited.

1. Introduction

A hyperspectral image (HSI) is able to record hundreds of contiguous spectral channels, and thus provides an unique ability to identify different types of materials. Owing to this merit, hyperspectral imaging has been extensively applied in various aspects, such as land cover mapping [1,2], object detection [3,4], and environment monitoring [5,6]. Over the past few years, hyperspectral image classification has been made great progress because of its significance in mineral mapping, urban investigation, and precision agriculture. Nevertheless, object spectrum is usually affected by the imaging equipment and imaging environment, resulting in high spectrum mixture among different land covers.
To alleviate this problem, hyperspectral feature extraction methods have been widely studied to improve the class discrimination among different land covers. Some representative manifold learning tools [7,8,9,10] have been successfully used as feature extractor of HSIs, such as principal component analysis (PCA) [8], minimum noise fraction (MNF) [10], and independent component analysis (ICA) [9]. However, most of these techniques only utilize the spectral information of different objects, and thus fail to achieve satisfactory classification performance.
To fully exploit the spectral and spatial characteristics in HSIs, a mass of spectral–spatial feature extraction techniques have been developed [11,12,13,14,15]. For instance, Marpu et al. developed attribute profiles (APs) to extract discriminative features of HSIs by using morphological operations [12]. Mura et al. studied extended morphological attribute profiles (EMAP) to characterize HSIs by using a series of morphological attribute filters [13]. Kang et al. developed an edge-preserving filtering method for hyperspectral feature extraction to remove low-contrast details [14]. Duan et al. modeled hyperspectral image as a linear combination of structural profile and texture information, in which the structural profile was used as the spatial features [15]. After that, many improved approaches have been also studied for classification of HSIs, such as ensemble learning [16,17], semi-supervised learning [18,19], and active learning [20,21]. For instance, in [17], a random feature ensemble method was proposed by using ICA and edge-preserving filtering to boost the classification accuracy of HSIs. In [18], a rolling guidance filter-based semi-supervised classification method was developed in which an extended label propagation technique was utilized to expand the training set.
In addition, with the development of deep learning models, many deep learning models have been applied to extract the high-order semantic features of HSIs. For example, Chen et al. presented a 3D convolutional neural network (CNN)-based feature extraction technique in which the dropout and L 2 regularization were adopted to prevent overfitting for class imbalance [22]. Liu et al. proposed a Siamese CNN method with a margin ranking loss function to improve the discrimination of different objects [23]. Liu et al. designed a mixed deep feature extraction technique by combining the pixel frequency spectrum features obtained by fast Fourier transform and spatial features. Lately, all kinds of improved versions have been also investigated to increase the classification accuracy [24,25,26,27,28]. For example, Hang et al. designed an attention-guided CNN method to extract spectral–spatial information, in which a spectral attention module and a spatial attention module were considered to capture the spectral and spatial information, respectively [26]. Hong et al. improved the transformer network to characterize the sequence attributes of spectral information, where a cross-layer skip connection was used to fuse spatial features from different layers [27].
To better characterize spectral–spatial information, the multiview technique, which aims to reveal the data characteristics from diverse aspects and provide multiple features for model learning, has been applied in hyperspectral feature extraction [20,29,30]. In more detail, the raw data are first transformed into different views (e.g., attributes, feature subsets), and then, the complementary information of all views is integrated together to achieve a more accurate classification result. For example, in [20], Li et al. proposed a multi-view active learning method for hyperspectral image classification, where the subpixel-level, pixel-level, and superpixel-level information were jointly used to achieve a better identification ability. Xu et al. proposed multi-view attribute components for classification of HSIs, in which an intensity-based query scheme was used to expand the number of training set [29]. In general, these methods can increase the classification performance because of multi-view strategy. However, most of them only utilize the local neighboring information without considering the correlation of pixels in the nonlocal region. Based on the above analysis, it is necessary to develop a novel multi-view method to further boost the classification performance by jointly using the dependencies of pixels in the local and nonlocal regions.
In this work, we propose a novel multi-view structural feature extraction method for the first time, which consists of several key steps. First, the spectral dimension of the original data is reduced to increase the computational efficiency. Then, three multi-view structural features are constructed to characterize varying land covers from different aspects. Finally, different types of features are merged together to increase the discrimination of different land covers, and the fused feature is fed into a spectral classifier to obtain the classification results. Experiments are performed on three benchmark datasets to quantitatively and qualitatively validate the effectiveness of the proposed method. The experimental results verify that the proposed feature extraction method can significantly outperform other state-of-the-art feature extractors. More importantly, our method can obtain promising classification performance over other approaches in the case of limited training samples.
The remainder of this article is divided into four sections. Section 2 provides the proposed method. The experimental results are shown and analyzed in Section 3. Section 4 discusses the influence of different components. Finally, Section 5 concludes this work.

2. Method

Figure 1 shows the flowchart of the multi-view structural feature extraction method, which consists of three key steps. First, the spectral number of the raw data is reduced with the MNF method. Then, the multi-view structural features, i.e., local structural feature, intra-view, and inter-view structural features, are constructed based on the correlation of pixels in the local and nonlocal regions. Finally, the multi-view features are fused together with the kernel PCA (KPCA) method, and the fused feature is fed into the support vector machine (SVM) classifier to obtain the classification map.

2.1. Dimension Reduction

To decrease the computing time and the influence of image noise, the maximum noise fraction (MNF) [31] is first exploited to decrease the spectral dimension of the original data. Specifically, assume I is the raw data, the MNF is to seek a transform matrix W to maximize the signal to noise ratio of transformed data.
R = W T I
where R is the dimension-reduced data, and W is the transform matrix, which can be estimated as
arg max W W T S W W T N W = arg max W W T I W W T N W 1
where I is regarded as a linear combination of the uncorrelated signal S and noise matrix N , and cov ( I ) = I = S + N , in which S and N denote the covariance of S and N . In this work, the first L-dimensional components are preserved for the following feature extraction.

2.2. Multi-View Feature Generation

Since the imaging scene contains different types of land covers with different spatial size, a single structural feature cannot comprehensively characterize the spatial information of diverse objects. To alleviate this problem, a multi-view structural feature extraction method is proposed. Specifically, three different types of structural features are generated, including local structural feature, intra-view, and inter-view structural features.
(1) Local structural feature: The local structural feature aims to remove useless details (e.g., image noise and texture) and preserve the intrinsic spectral–spatial information. Specifically, a relative total variation technique [32] is performed on the dimension-reduced data R to construct local structural feature F 1 , which is expressed as
arg min F 1 i = 1 T ( F 1 i R i ) 2 + α · ( D x ( i ) L x ( i ) + ε + D y ( i ) L y ( i ) + ε ) ,
where T represents the amount of pixels in total. F 1 stands for the desired structural feature. α denotes a smoothing weight. ε is adopted to prevent dividing by zero. D x and D y indicate the variations in two directions. The solution of Equation (3) refers to [32].
D x ( i ) = j R ( i ) g i , j · | ( x S ) j |
D y ( i ) = j R ( i ) g i , j · | ( y S ) j |
where x S and y S stand for the partial derivatives in two directions, which mainly calculates the spatial similarity within a local window R ( i ) , and g i , j is a weight.
g i , j = exp ( x i x j ) 2 + ( y i y j ) 2 2 σ 2 .
where σ denotes the window size.
(2) Intra-view structural feature: The intra-view structural feature is to reduce the spectral difference of pixels belonging to the same land cover and increase the spectral purity in the homogeneous regions. In order to extract the intra-view structural feature, first, an entropy rate superpixel (ERS) segmentation method [33] is adopted to obtain the homogeneous region of the same object. In more detail, the PCA scheme is first conducted on the local structural feature F 1 to obtain the first three components F ^ 1 , and then, the ERS segmentation scheme is utilized to obtain a 2D superpixel resulting map.
S = E R S ( F ^ 1 , T )
where S indicates the segmentation result, and T indicates the number of superpixels, which is determined by
T = L × N ^ N
where L is empirically selected in this work, N ^ represents the amount of nonzero pixels in the detected map obtained by performing Canny filter on the base image F ^ 1 , and N represents the total amount of pixels in the base image. Based on the position indexes of pixels in each superpixel S i , we can obtain the corresponding 3D superpixels Y i , i { 1 , 2 , , T } .
Then, a mean filtering is conducted on each 3D superpixel Y i to calculate the average value. Finally, we assign the pixels in each superpixel to the average value to obtain the intra-view structural feature F 2 .
(3) Inter-view structural feature: The intra-view structural feature is able to reduce the difference of pixels in each superpixel. However, the correlations of pixels for different superpixels are not considered. Thus, we construct an inter-view structural feature to improve the discrimination of different objects. Specifically, a weighted mean operation is performed on the neighboring superpixels Y i , j , { j = 1 , 2 , , J } of the current superpixel Y i to obtain the inter-view structural feature F 3 .
y ˜ i = j = 1 J ω i , j × y i , j
where ω i , j denotes the neighboring superpixels of the ith superpixels. The obtained value y ˜ i is assigned to all pixels in the ith superpixel to produce the inter-view structural feature F 3 .

2.3. Feature Fusion

To make full use of multi-view structural features, the KPCA technique [34] is used to merge three types of features. Specifically, first, three structural features F i , { i = 1 , 2 , 3 } are stacked together F = { F 1 , F 2 , F 3 } . Then, the stacked data F is projected into a high-dimensional space H by using a Gaussian kernel function Φ . Finally, the fused feature F ^ can be calculated:
K α = λ α s . t . α 2 = 1 λ
where K indicates Gram matrix Φ T ( S ) Φ ( S ) . In this paper, K-dimensional features are preserved in the fused feature. Once the fused feature is obtained, the spectral classifier, i.e., SVM, is considered to examine the classification performance. To clearly show the whole procedure of the proposed method, Algorithm 1 presents a pseudocode to summarize the key steps of our method.
Algorithm 1 Multi-view structural feature extraction
Input:
 Input hyperspectral image I;
Output:
 Hyperspectral image feature F ^
 1:
According to (1), reduce the spectral number of the raw data I to obtain the dimension reduced image R
 2:
According to (3), obtain the local structural feature F 1 of the input image.
 3:
According to (6), obtain the 2-D superpixel segmentation map S
 4:
Obtain the corresponding 3-D superpixels Y based on the position indexes of pixels in each superpixels S i
 5:
Construct the intra-view structural feature F 2 by performing the mean filtering on each 3-D superpixel
 6:
According to (8), construct the inter-view structural feature F 3
 7:
According to (9), fuse the multi-view structural features F i , { i = 1 , 2 , 3 } to obtain the final feature F ^
 8:
Return F ^

3. Experiments

3.1. Experimental Setup

(1) Datasets: In the experimental section, three hyperspectral datasets, i.e., Indian Pines, Salinas, and Honghu, are used to examine the classification performance of the proposed feature extraction method. All these datasets are collected from a public hyperspectral database.
The Indian Pines dataset was obtained by the Airborne Visible Infrared Imaging Spectrometer (AVIRIS) sensor over the Indian Pines test scene in northwestern Indiana, which is available online (http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 10 January 2022 )). This image is composed of 220 spectral bands spanning from 0.4 to 2.5 µm. The spatial size is 145 × 145 with a spatial resolution of 20 m. Twenty water absorption channels (No. 104-108, 150-163, and 220) are discarded before experiments. Figure 2a presents the false color composite image. Figure 2b shows the ground truth, which contains 16 different land covers. Figure 2c gives the class name.
The Salinas dataset was collected by the AVIRIS sensor over Salinas Valley, California, USA, which is available online (http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 10 January 2022)). This image consists of 224 spectral channels with spatial size of 512 × 217 pixels. The spatial resolution is of 3.7 m. This scene is an agricultural region, which includes different types of crops, such as vegetables, bare soils, and vineyard fields. Twenty spectral bands (No. 108-112, 154-167, and 224) are discarded before the following experiments. Figure 3a gives the false color composition. Figure 3b displays the ground truth, which consists of 16 different land covers. Figure 3c shows the class name.
The Honghu dataset was captured by the a 17 mm focal length Headwall Nano-Hyperspec imaging sensor over Honghu City, Hubei province, China, which is available online (http://rsidea.whu.edu.cn/e-resource_WHUHi_sharing.htm (accessed on 10 January 2022)). This image consists of 270 spectral channels ranging from 0.4 to 1.0 µm. The spatial size is 940 × 475 pixels with a spatial resolution of 0.043 m. This scene is a complex agricultural area, including various crops and different cultivars of the same crop. Figure 4 gives the false color composition, ground truth, and class name. Table 1 presents the training and test samples of all used datasets for the following experiments.
(2) Evaluation Indexes: To quantitatively calculate the classification accuracies of all considered techniques, four extensively used objective indexes [35,36,37], i.e., class accuracy (CA), overall accuracy (OA), average accuracy (AA), and Kappa coefficient, are used. The definitions of all objective indexes are shown as follows:
(1) CA: CA calculates the percentage of correctly classified pixels of each class in the total number of pixels.
C A = M i i j = 1 C M i j
where M is the confusion matrix obtained by comparing the ground truth with the predicted result, and C is the total number of categories.
(2) OA: OA assesses the proportion of correctly identified samples to all samples.
O A = i = 1 C M i i / N
where N is the total number of labeled samples, M is the confusion matrix, and C is the total number of categories.
(3) AA: AA represents mean of the percentage of the correctly identified samples.
A A = i = 1 C ( M i i M i i j = 1 C M i j j = 1 C M i j ) C
where C is the total number of categories, and M is the confusion matrix.
(4) Kappa: Kappa coefficient denotes the interrater reliability for categorical variables.
K a p p a = N i = 1 C M i i i = 1 C j = 1 C M i j j = 1 C M j i N 2 i = 1 C j = 1 C M i j j = 1 C M j i
where N is the total number of labeled samples, M is the confusion matrix, and C is the total number of categories.

3.2. Classification Results

To examine the effectiveness of the proposed feature extraction method, several state-of-the-art hyperspectral classification methods are selected as competitors, including (1) the spectral classifier, i.e., SVM on the original image (SVM) [38]; (2) the feature extraction methods, i.e., the image fusion and recursive filtering (IFRF) [14], the extended morphological attribute profiles (EMAP) [13], multi-scale total variation (MSTV) [15], the PCA-based edge-preserving features (PCAEPFs) [8]; (3) the spectral–spatial classification methods, i.e., the superpixel-based classification via multiple kernels (SCMK) [39] and the generalized tensor regression approach (GTR) [40]. These methods are adopted because they are either highly cited publications in the remote sensing field or are recently proposed classification methods with state-of-the-art classification performance on several hyperspectral datasets. For all considered approaches, the default parameters follow the corresponding publications for a fair comparison.

3.2.1. Indian Pines Dataset

The first experiment is performed on the Indian Pines dataset, in which 1% labeled samples are randomly selected from the reference image for training (see Table 1). Figure 5 presents the classification maps of all considered methods on the Indian Pines dataset. As shown in this figure, the SVM method yields very noisy visual effects in the classification result, exposing the disadvantages of the spectral classifier without considering the spatial information. By removing image details and preserving the strong edge structures, the IFRF method greatly improves the classification result over the spectral classifier. However, there are still obvious misclassified pixels around the boundaries. For the EMAP method, some homogeneous regions still contain “noise” such as mislabels in the resulting map. For the SCMK method, some obvious misclassified results appear in the edges and corners. The main reason is that the homogeneous region belonging to the same object cannot be accurately segmented. The MSTV method effectively removes the noisy labels. However, it tends to yield an oversmoothed classification map. The PCAEPFs method significantly boosts the classification performance with respect to the IFRF method, since multi-scale feature extraction strategy is adopted. However, some objects with small size fail to be well preserved in the classification map. The GTR method yields spot-like misclassification results since the tensor regression technique cannot fit the spectral curves of different objects well. By contrast, the proposed method obtains a better visual map by integrating multi-view spectral–spatial structural features, in which the edges of the classification map are more consistent with the real scene.
For objective comparison, Table 2 lists the objective results of different approaches including CA, OA, AA, and Kappa. It is easily to observe that the proposed method is superior than all compared approaches in terms of OA, AA, and Kappa. For instance, OA value is increased from 53% to 90% obtained by the proposed method with respect to the SVM method on the original data. Moreover, the proposed feature extraction approach yields the highest classification accuracies for ten classes. This experiment illustrates that the proposed feature extraction method is more effective compared to other approaches.
Furthermore, the influence of the number of training set on all classification methods is discussed. Different numbers of samples varying from 1% to 10% are randomly chosen from the reference data to construct the training set. Figure 6 shows the change tendency of all studied methods with different numbers of training set. It is easily found that the classification performance of all methods tends to be improved when the amount of training set increases. In addition, the proposed method performs promising performance with respect to other methods especially when the number of training samples is limited.

3.2.2. Salinas Dataset

The second experiment is conducted on the Salinas dataset, in which five samples per class are randomly selected from the reference image to constitute the training samples (see Table 1). Figure 7 shows the visual maps of all studied approaches. We can easily observe that the SVM method produces noisy classification performance. The reason is that the spatial priors are not considered in the spectral classifier. The IFRF method removes the noisy labels in some homogeneous regions. However, there are still serious misclassifications, such as Grapes_untrained and Vinyard_treils classes. The EMAP method also yields “pepper and noisy” appearance since this method is a pixel-level feature extraction method. For the SCMK method, some regions are misclassified into other classes due to the inaccurate segmentation. The MSTV method produces an oversmoothed visual resulting map. The main reason is that the feature extraction process removes the spatial information of land covers with low reflectivity. The PCAEPFs method produces noisy labels in the edges and boundaries. The GTR method yields a serious misclassification map since the tensor regression model fails to distinguish the similar spectral curves. Different from other methods, the proposed method provides the best visual classification effect in removing noisy labels and preserving the boundaries of different classes.
Furthermore, Table 3 also verifies the effectiveness of the proposed method. Likewise, the proposed method obtains the highest classification accuracies with regard to OA, AA, and Kappa compared to other studied techniques. In addition, the influence of different numbers of training samples is presented in Figure 8. The number of training set for each class is varying from 5 to 50. It is shown that the increase of the training size is beneficial to the classification performance of all methods. Moreover, our method is always higher than other classification approaches.

3.2.3. Honghu Dataset

The third experiment is performed on a complex crop scene with 22 crop classes, i.e., Honghu dataset, in which the benchmark training and test samples (http://rsidea.whu.edu.cn/e-resource_WHUHi_sharing.htm (accessed on on Jaunary 2022)) are adopted. The classification results of all methods are shown in Figure 9. Similarly, the SVM method produces a very noisy visual map since only the spectral information is used. The IFRF method greatly improves this problem by using the image filtering on the raw data, obtaining a better classification result over the SVM method. The EMAP method also obtains a noisy classification map. The reason is that this feature extraction method is in a pixel-wise pattern. For the SCMK method, the classification map has a small amount of misclassification labels for some classes. The MSTV method obtains a better classification map due to multi-scale technique. However, the edges of different classes still have misclassification appearance. The PCAEPFs method yields a similar classification effect to the MSTV method. For the GTR method, the classification result suffers from serious misclassification for this complex scene when the amount of training set is limited. By contrast, the proposed method yields the best visual map with respect to other studied approaches, since multi-view structural features are effectively merged to yield a complete characterization of different objects.
Furthermore, the objective results obtained by all considered approaches are listed in Table 4. It is obvious from Table 4 that our method still provides the highest classification accuracies concerning OA, AA, and Kappa coefficient. In addition, Figure 10 gives the OA, AA, and Kappa coefficient of all studied approaches as functions of the amount of training samples from 25 to 300. It should be mentioned that the training samples follow the benchmark dataset (http://rsidea.whu.edu.cn/e-resource_WHUHi_sharing.htm (accessed on 10 January 2022)). It is found that the classification accuracies of all methods tend to improve when the training size increases, and our method still produces the highest OA, AA, and Kappa coefficient.

4. Discussion

4.1. The Influence of Different Parameters

In this part, the influence of different free parameters, i.e., the number of dimension reduction L, the number of fused feature K, the smoothing weight α , and the window size σ , on the classification accuracy of our method is analyzed. An experiment is performed on the three datasets with training set listed in Table 1. When L and K are discussed, α and σ are fixed as 0.005 and 3, respectively. Similarly, when α and σ are analyzed, L and K are set as 30 and 20, respectively. Figure 11 presents the classification accuracy OA of the proposed method with different parameter settings. It can be observed that the proposed method can yield satisfactory classification accuracy for all used datasets when L and K are set to be 30 and 20, respectively. Moreover, when L and K are relatively small, the classification performance tends to be decreased, since the limited number of features cannot well represent the spectral–spatial information in HSIs. Figure 11d–f shows the influence of different α and σ . It is shown that when α and σ increase, the classification performance of the proposed method decreases. The reason is that the structural feature extraction technique smooths out the spatial structures of HSIs. When α and σ are set to be 0.005 and 3, respectively, the proposed method obtains the highest classification accuracy. Based on this observation, L, K, α , and σ are set as 30, 20, 0.005, and 3, respectively.

4.2. The Influence of Three Different Views

In this subsection, the influence of three different views on the classification accuracy is investigated. An experiment is conducted on the Indian Pines dataset. The classification performance obtained by the proposed framework with different views is shown in Table 5. It can be seen that the inter-view feature performs the best classification performance among three different views. Furthermore, the combination of two different views outperforms individual view in terms of classification accuracies. Overall, when three types of features are combined, the proposed method provides the highest classification results. The reason is that three different views have complementary information, which can be jointly utilized to improve the classification performance.

4.3. Effect of Different Hyperspectral Feature Methods

To demonstrate the advantage of the proposed feature extraction method, several widely used feature extraction methods for HSIs are selected as competitors, including extended morphological profiles (EMP) [41], extended morphological attribute profiles (EMAP) [13], Gabor filtering (Gabor) [42], image fusion and recursive filtering (IFRF) [14], intrinsic image decomposition (IID) [43], PCA-based edge-preserving filters (PCAEPFs) [8], invariant attribute profiles (IAPs) [44], low rank representation (LRR) [45], multi-scale total variation (MSTV) [15], and random patches network (RPNet) [46]. An experiment is performed on the Indian Pines dataset with 1% of training samples listed in Table 1. The classification accuracy of all considered approaches is shown in Figure 12. The EMP and EMAP methods only yield around 60% classification performance when the number of training samples is scarce. The classification performance obtained by the edge-preserving filtering-based feature extraction methods such as IFRF and PCAEPFs also tends to decrease when the number of training set is scarce. The RPNet-based deep feature extraction method also fails to achieve satisfactory performance. By contrast, it is found that the proposed feature method obtains the highest classification performance among all feature extraction techniques for three indexes, which further illustrates that the proposed method can better characterize the spectral–spatial information compared to other methods by fusing local and nonlocal multi-view structural features.

4.4. Computing Time

The computing efficiency of all considered techniques for all datasets is provided in Table 6. All experiments are tested a laptop with 8 GB RAM and 2.6 GHz with Matlab 2018. We can observe from Table 6 that when the spatial and spectral dimensions of HSIs increase, the computing time of all methods tends to increases. Furthermore, the computing time of our method is quite competitive among all considered approaches (taking the Indian Pines dataset as an example, the running time of our method is around 5.36 s). The GTR method is the fastest as it is a regression model.

5. Conclusions

In this work, a multi-view structural feature extraction method is developed for hyperspectral image classification, which consists of three key steps. First, the spectral number of the raw data is decreased. Then, the local structural feature, intra-view structural feature, and inter-view structural feature are constructed to characterize spectral–spatial information of diverse ground objects. Finally, the KPCA technique is exploited to merge multi-view structural features, and the fused feature is incorporated with the spectral classifier to obtain the classification map. Our experimental results on three datasets reveal that the proposed feature extraction method can consistently outperform other state-of-the-art classification methods even when the number of training set is limited. Furthermore, with regard to 10 other representative feature extraction methods, our method still produces the highest classification performance.

Author Contributions

N.L. performed the experiments and wrote the draft. P.D. provided some comments and carefully revised the presentation. H.X. modified the presentation of this work. L.C. checked the grammar of this work. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Key Laboratory of Mine Water Resource Utilization of Anhui Higher Education Institutes, Suzhou University, under grant KMWRU202107, in part by the Key Natural Science Project of the Anhui Provincial Education Department under grant KJ2021ZD0137, in part by the Key Natural Science Project of the Anhui Provincial Education Department under grant KJ2020A0733, in part by the Top talent project of colleges and universities in Anhui Province under grant gxbjZD43, and in part by the Collaborative Innovation Center—cloud computing industry under grant 4199106.

Data Availability Statement

The Indian Pines and Salinas datasets are freely available from this site (http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 10 January 2022)). The Honghu dataset is freely available from this site (http://rsidea.whu.edu.cn/e-resource_WHUHi_sharing.htm (accessed on 10 January 2022)).

Acknowledgments

We would like to thank Y. Zhong from Wuhan University for sharing the Honghu datset.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MNFMinimum noise fraction
HSIHyperspectral image
PCAPrincipal component analysis
ICAIndependent component analysis
APsAttribute profiles
EMAPExtended morphological attribute profiles
CNNConvolutional neural network
ERSEntropy rate superpixel
KPCAKernel PCA
SVMSupport vector machine
AVIRISAirborne Visible Infrared Imaging Spectrometer
CAClass accuracy
OAOverall accuracy
AAAverage accuracy
IFRFImage fusion and recursive filtering
SCMKSuperpixel-based classification via multiple kernels
MSTVMulti-scale total variation
PCAEPFsPCA-based edge-preserving features
GTRGeneralized tensor regression
EMPExtended morphological profiles
GaborGabor filtering
IIDIntrinsic image decomposition
IAPsInvariant attribute profiles
LRRLow rank representation
RPNetRandom patches network

References

  1. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution From Shallow to Deep: Overview and Toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  2. Duan, P.; Lai, J.; Ghamisi, P.; Kang, X.; Jackisch, R.; Kang, J.; Gloaguen, R. Component Decomposition-Based Hyperspectral Resolution Enhancement for Mineral Mapping. Remote Sens. 2020, 12, 2903. [Google Scholar] [CrossRef]
  3. Liang, J.; Zhou, J.; Tong, L.; Bai, X.; Wang, B. Material Based Salient Object Detection from Hyperspectral Images. Pattern Recognit. 2018, 76, 476–490. [Google Scholar] [CrossRef] [Green Version]
  4. Li, S.; Zhang, K.; Duan, P.; Kang, X. Hyperspectral Anomaly Detection With Kernel Isolation Forest. IEEE Trans. Geosci. Remote Sens. 2020, 58, 319–329. [Google Scholar] [CrossRef]
  5. Duan, P.; Lai, J.; Kang, J.; Kang, X.; Ghamisi, P.; Li, S. Texture-Aware Total Variation-Based Removal of Sun Glint in Hyperspectral Images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 359–372. [Google Scholar] [CrossRef]
  6. Stuart, M.B.; McGonigle, A.J.S.; Willmott, J.R. Hyperspectral Imaging in Environmental Monitoring: A Review of Recent Developments and Technological Advances in Compact Field Deployable Systems. Sensors 2019, 19, 3071. [Google Scholar] [CrossRef] [Green Version]
  7. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  8. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-Based Edge-Preserving Features for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  9. Wang, J.; Chang, C.I. Independent Component Analysis-Based Dimensionality Reduction with Applications in Hyperspectral Image Analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef]
  10. Gao, L.; Zhao, B.; Jia, X.; Liao, W.; Zhang, B. Optimized Kernel Minimum Noise Fraction Transformation for Hyperspectral Image Classification. Remote Sens. 2017, 9, 548. [Google Scholar] [CrossRef] [Green Version]
  11. Zhang, Y.; Kang, X.; Li, S.; Duan, P.; Benediktsson, J.A. Feature Extraction from Hyperspectral Images using Learned Edge Structures. Remote Sens. Lett. 2019, 10, 244–253. [Google Scholar] [CrossRef]
  12. Marpu, P.R.; Pedergnana, M.; Dalla Mura, M.; Benediktsson, J.A.; Bruzzone, L. Automatic Generation of Standard Deviation Attribute Profiles for Spectral–Spatial Classification of Remote Sensing Data. IEEE Geosci. Remote Sens. Lett. 2013, 10, 293–297. [Google Scholar] [CrossRef]
  13. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of Hyperspectral Images by Using Extended Morphological Attribute Profiles and Independent Component Analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  14. Kang, X.; Li, S.; Benediktsson, J.A. Feature Extraction of Hyperspectral Images With Image Fusion and Recursive Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  15. Duan, P.; Kang, X.; Li, S.; Ghamisi, P. Noise-Robust Hyperspectral Image Classification via Multi-Scale Total Variation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1948–1962. [Google Scholar] [CrossRef]
  16. Duan, P.; Ghamisi, P.; Kang, X.; Rasti, B.; Li, S.; Gloaguen, R. Fusion of Dual Spatial Information for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7726–7738. [Google Scholar] [CrossRef]
  17. Xia, J.; Bombrun, L.; Adalı, T.; Berthoumieu, Y.; Germain, C. Spectral–Spatial Classification of Hyperspectral Images Using ICA and Edge-Preserving Filter via an Ensemble Strategy. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4971–4982. [Google Scholar] [CrossRef] [Green Version]
  18. Cui, B.; Xie, X.; Hao, S.; Cui, J.; Lu, Y. Semi-Supervised Classification of Hyperspectral Images Based on Extended Label Propagation and Rolling Guidance Filtering. Remote Sens. 2018, 10, 515. [Google Scholar] [CrossRef] [Green Version]
  19. Sellars, P.; Aviles-Rivero, A.I.; Schönlieb, C.B. Superpixel Contracted Graph-Based Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4180–4193. [Google Scholar] [CrossRef] [Green Version]
  20. Li, Y.; Lu, T.; Li, S. Subpixel-Pixel-Superpixel-Based Multiview Active Learning for Hyperspectral Images Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4976–4988. [Google Scholar] [CrossRef]
  21. Li, Q.; Zheng, B.; Yang, Y. Spectral-Spatial Active Learning With Structure Density for Hyperspectral Classification. IEEE Access 2021, 9, 61793–61806. [Google Scholar] [CrossRef]
  22. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  23. Liu, B.; Yu, X.; Zhang, P.; Yu, A.; Fu, Q.; Wei, X. Supervised Deep Feature Extraction for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1909–1921. [Google Scholar] [CrossRef]
  24. Kang, X.; Zhuo, B.; Duan, P. Dual-Path Network-Based Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 447–451. [Google Scholar] [CrossRef]
  25. Xie, Z.; Hu, J.; Kang, X.; Duan, P.; Li, S. Multi-Layer Global Spectral-Spatial Attention Network for Wetland Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 58, 3232–3245. [Google Scholar]
  26. Hang, R.; Li, Z.; Liu, Q.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral Image Classification With Attention-Aided CNNs. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2281–2293. [Google Scholar] [CrossRef]
  27. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  28. Duan, P.; Xie, Z.; Kang, X.; Li, S. Self-Supervised Learning-Based Oil Spill Detection of Hyperspectral Images. Sci. China Technol. Sci. 2022, 65, 793–801. [Google Scholar] [CrossRef]
  29. Xu, X.; Li, J.; Li, S. Multiview Intensity-Based Active Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 669–680. [Google Scholar] [CrossRef]
  30. Zhou, X.; Prasad, S.; Crawford, M.M. Wavelet-Domain Multiview Active Learning for Spatial-Spectral Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4047–4059. [Google Scholar] [CrossRef]
  31. Green, A.; Berman, M.; Switzer, P.; Craig, M. A Transformation for Ordering Multispectral Data in terms of Image Quality with Implications for Noise Removal. IEEE Trans. Geosci. Remote Sens. 1988, 26, 65–74. [Google Scholar] [CrossRef] [Green Version]
  32. Xu, L.; Yan, Q.; Xia, Y.; Jia, J. Structure Extraction from Texture via Relative Total Variation. ACM Trans. Graph. 2012, 31, 1–10. [Google Scholar] [CrossRef]
  33. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy-Rate Clustering: Cluster Analysis via Maximizing a Submodular Function Subject to a Matroid Constraint. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 99–112. [Google Scholar] [CrossRef] [PubMed]
  34. Schölkopf, B.; Smola, A.; Müller, K.R. Kernel Principal Component Analysis. In Artificial Neural Networks—ICANN’97, Proceedings of the 7th International Conference, Lausanne, Switzerland, 8–10 October 1997 Proceeedings; Gerstner, W., Germond, A., Hasler, M., Nicoud, J.D., Eds.; Springer: Berlin/Heidelberg, Germany, 1997; pp. 583–588. [Google Scholar]
  35. Duan, P.; Kang, X.; Li, S.; Ghamisi, P.; Benediktsson, J.A. Fusion of Multiple Edge-Preserving Operations for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10336–10349. [Google Scholar] [CrossRef]
  36. Duan, P.; Kang, X.; Li, S.; Ghamisi, P. Multichannel Pulse-Coupled Neural Network-Based Hyperspectral Image Visualization. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2444–2456. [Google Scholar] [CrossRef]
  37. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef] [Green Version]
  38. Melgani, F.; Bruzzone, L. Classification of Hyperspectral Remote Sensing Images with Support Vector Machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  39. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of Hyperspectral Images by Exploiting Spectral–Spatial Information of Superpixel via Multiple Kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  40. Liu, J.; Wu, Z.; Xiao, L.; Sun, J.; Yan, H. Generalized Tensor Regression for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1244–1258. [Google Scholar] [CrossRef]
  41. Benediktsson, J.; Palmason, J.; Sveinsson, J. Classification of Hyperspectral Data from Urban Areas Based on Extended Morphological Profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  42. He, L.; Li, J.; Plaza, A.; Li, Y. Discriminative Low-Rank Gabor Filtering for Spectral–Spatial Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1381–1395. [Google Scholar] [CrossRef]
  43. Kang, X.; Li, S.; Fang, L.; Benediktsson, J.A. Intrinsic Image Decomposition for Feature Extraction of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2241–2253. [Google Scholar] [CrossRef]
  44. Hong, D.; Wu, X.; Ghamisi, P.; Chanussot, J.; Yokoya, N.; Zhu, X.X. Invariant Attribute Profiles: A Spatial-Frequency Joint Feature Extractor for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3791–3808. [Google Scholar] [CrossRef] [Green Version]
  45. Rasti, B.; Ulfarsson, M.O.; Sveinsson, J.R. Hyperspectral Feature Extraction Using Total Variation Component Analysis. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6976–6985. [Google Scholar] [CrossRef]
  46. Xu, Y.; Du, B.; Zhang, F.; Zhang, L. Hyperspectral Image Classification via a Random Patches Network. ISPRS J. Photogramm. Remote Sens. 2018, 142, 344–357. [Google Scholar] [CrossRef]
Figure 1. The flow chart of the proposed multi-view structural feature extraction method.
Figure 1. The flow chart of the proposed multi-view structural feature extraction method.
Remotesensing 14 01971 g001
Figure 2. Indian Pines dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Figure 2. Indian Pines dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Remotesensing 14 01971 g002
Figure 3. Salinas dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Figure 3. Salinas dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Remotesensing 14 01971 g003
Figure 4. Honghu dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Figure 4. Honghu dataset. (a) False color composite. (b) Ground truth. (c) Label name.
Remotesensing 14 01971 g004
Figure 5. Classification results of all considered approaches on Indian Pines dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 55.30%. (d) IFRF [14], OA = 70.09%. (e) EMAP [13], OA = 67.70%. (f) SCMK [39], OA = 71.20%. (g) MSTV [15], OA = 88.25%. (h) PCAEPFs [8], OA = 85.58%. (i) GTR [40], OA = 63.25%. (j) Our method, OA = 90.32%.
Figure 5. Classification results of all considered approaches on Indian Pines dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 55.30%. (d) IFRF [14], OA = 70.09%. (e) EMAP [13], OA = 67.70%. (f) SCMK [39], OA = 71.20%. (g) MSTV [15], OA = 88.25%. (h) PCAEPFs [8], OA = 85.58%. (i) GTR [40], OA = 63.25%. (j) Our method, OA = 90.32%.
Remotesensing 14 01971 g005
Figure 6. Classification performance of different approaches on the Indian Pines with different numbers of training samples. (a) OA. (b) AA. (c) Kappa. The widths of the line areas are the standard deviation of accuracies produced in ten experiments.
Figure 6. Classification performance of different approaches on the Indian Pines with different numbers of training samples. (a) OA. (b) AA. (c) Kappa. The widths of the line areas are the standard deviation of accuracies produced in ten experiments.
Remotesensing 14 01971 g006
Figure 7. Classification results of all considered approaches on Salinas dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 80.08%. (d) IFRF [14], OA = 90.67%. (e) EMAP [13], OA = 85.54%. (f) SCMK [39], OA = 88.70%. (g) MSTV [15], OA = 94.46%. (h) PCAEPFs [8], OA = 95.12%. (i) GTR [40], OA = 85.59%. (j) Our method, OA = 98.13%.
Figure 7. Classification results of all considered approaches on Salinas dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 80.08%. (d) IFRF [14], OA = 90.67%. (e) EMAP [13], OA = 85.54%. (f) SCMK [39], OA = 88.70%. (g) MSTV [15], OA = 94.46%. (h) PCAEPFs [8], OA = 95.12%. (i) GTR [40], OA = 85.59%. (j) Our method, OA = 98.13%.
Remotesensing 14 01971 g007
Figure 8. Classification performance of different approaches on the Salinas with different numbers of training samples. (a) OA. (b) AA. (c) Kappa. The widths of the line areas are the standard deviation of accuracies produced in ten experiments.
Figure 8. Classification performance of different approaches on the Salinas with different numbers of training samples. (a) OA. (b) AA. (c) Kappa. The widths of the line areas are the standard deviation of accuracies produced in ten experiments.
Remotesensing 14 01971 g008
Figure 9. Classification results of all considered approaches on Honghu dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 64.43%. (d) IFRF [14], OA = 84.23%. (e) EMAP [13], OA = 76.11%. (f) SCMK [39], OA = 86.41%. (g) MSTV [15], 90.74%. (h) PCAEPFs [8], OA = 87.37%. (i) GTR [40], OA = 51.87%. (j) Our method, OA = 94.01%.
Figure 9. Classification results of all considered approaches on Honghu dataset. (a) False color image. (b) Ground truth. (c) SVM [33], OA = 64.43%. (d) IFRF [14], OA = 84.23%. (e) EMAP [13], OA = 76.11%. (f) SCMK [39], OA = 86.41%. (g) MSTV [15], 90.74%. (h) PCAEPFs [8], OA = 87.37%. (i) GTR [40], OA = 51.87%. (j) Our method, OA = 94.01%.
Remotesensing 14 01971 g009
Figure 10. Classification performance of different approaches on the Honghu with different numbers of training samples. (a) OA. (b) AA. (c) Kappa.
Figure 10. Classification performance of different approaches on the Honghu with different numbers of training samples. (a) OA. (b) AA. (c) Kappa.
Remotesensing 14 01971 g010
Figure 11. The influence of different parameters in the proposed method. The first row is the influence of the number of dimension reduction L and the number of fused feature K. The second row is the influence of the smoothing parameter α and the window size σ . (a) Indian Pines dataset; (b) Salinas dataset; (c) Honghu dataset; (d) Indian Pines dataset; (e) Salinas dataset; (f) Honghu dataset.
Figure 11. The influence of different parameters in the proposed method. The first row is the influence of the number of dimension reduction L and the number of fused feature K. The second row is the influence of the smoothing parameter α and the window size σ . (a) Indian Pines dataset; (b) Salinas dataset; (c) Honghu dataset; (d) Indian Pines dataset; (e) Salinas dataset; (f) Honghu dataset.
Remotesensing 14 01971 g011
Figure 12. Classification accuracies of different hyperspectral feature extraction techniques on Indian Pines dataset.
Figure 12. Classification accuracies of different hyperspectral feature extraction techniques on Indian Pines dataset.
Remotesensing 14 01971 g012
Table 1. The number of training and test set. The colors represent different land covers in the classification map.
Table 1. The number of training and test set. The colors represent different land covers in the classification map.
No.Indian Pines DatasetSalinas DatasetHonghu Dataset
NameTrainTestNameTrainTestNameTrainTest
1Alfalfa640Weeds_152004Red roof2514,016
2Corn_N71421Weeds_253721Road253487
3Corn_M6824Fallow51971Bare soil2521,796
4Corn6231Fallow_P51389Cotton25163,260
5Grass_M6477Fallow_S52673Cotton firewood256193
6Grass_T6724Stubble53954Rape2544,532
7Grass_P622Celery53574Chinese cabbage2524,078
8Hay_W7471Grapes511,266Packchoi254029
9Oats614Soil56198Cabbage2510,794
10Soybean_N7965Corn53273Tuber mustard2512,369
11Soybean_M82447Lettuce_451063Brassica parachinensis2510,990
12Soybean_C6587Lettuce_551922Brassica chinensis258929
13Wheat6199Lettuce_65911Small Brassica chinensis2522,482
14Woods61259Lettuce_751065Lactuca sativa257331
15Buildings6380Vinyard_U57263Celtuce25977
16Stone786Vinyard_T51802Film covered257237
17Total10210,147Total8054,049Romaine lettuce252985
18 Carrot253192
19 White radish258687
20 Garlic sprout253461
21 Broad bean251303
22 Tree254015
Total550386,143
Table 2. Classification accuracies of all methods on Indian Pines dataset. The bold denotes the best classification accuracy.
Table 2. Classification accuracies of all methods on Indian Pines dataset. The bold denotes the best classification accuracy.
ClassSVMIFRFEMAPSCMKMSTVPCAEPFsGTROur Method
131.53 (9.74)63.03 (27.93)94.21 (9.19)98.00 (1.05)95.92 (10.41)98.02 (4.65)96.75 (2.06)100.0 (0.00)
247.00 (6.00)70.64 (15.98)59.74 (8.25)60.27 (11.34)86.93 (5.35)76.38 (7.21)55.64 (8.15)86.37 (5.87)
334.03 (14.66)48.62 (12.35)53.23 (13.54)57.23 (13.07)71.01 (8.26)73.79 (13.56)47.42 (8.58)83.07 (16.27)
426.70 (6.66)52.45 (11.34)36.54 (6.17)94.42 (4.94)67.17 (9.51)66.83 (7.53)72.47 (11.74)86.05 (10.51)
563.24 (9.75)75.81 (10.59)67.96 (10.27)78.97 (12.76)97.49 (3.45)93.86 (6.29)83.06 (9.97)98.30 (3.94)
679.83 (8.04)91.26 (4.32)90.31 (3.84)89.78 (11.19)98.70 (2.05)94.28 (3.29)88.07 (5.81)99.96 (0.13)
731.35 (14.53)57.65 (21.85)61.05 (20.27)97.27 (2.35)98.70 (2.10)73.05 (31.40)99.09 (2.87)80.73 (25.62)
895.57 (2.12)99.89 (0.27)100.0 (0.00)100.0 (0.00)100.0 (0.00)100.0 (0.00)88.77 (7.52)100.0 (0.00)
917.18 (8.59)28.91 (13.60)40.28 (10.36)100.0 (0.00)96.67 (10.54)72.87 (21.18)98.57 (4.52)98.08 (4.27)
1041.51 (5.35)66.42 (8.13)52.5 (10.53)67.95 (13.90)81.22 (9.40)78.55 (10.42)54.30 (8.60)90.70 (8.08)
1160.65 (8.31)73.91 (6.22)75.15 (9.85)59.69 (9.88)92.28 (4.04)90.14 (4.52)41.52 (11.20)89.40 (6.90)
1229.10 (5.43)56.65 (9.17)48.07 (5.31)65.54 (9.57)78.81 (12.25)77.19 (12.22)67.00 (8.52)83.81 (15.76)
1379.14 (3.12)74.79 (12.60)85.46 (9.48)100.0 (0.00)100.0 (0.00)97.45 (4.78)99.25 (1.09)100.0 (0.00)
1488.92 (5.36)93.03 (4.09)91.79 (4.98)81.92 (9.64)99.54 (0.65)99.66 (0.45)85.77 (9.36)99.61 (0.24)
1533.04 (7.36)59.07 (13.37)65.61 (14.81)76.47 (11.34)89.69 (10.56)93.72 (5.04)65.53 (11.80)93.78 (6.94)
1676.05 (19.59)93.16 (5.08)90.53 (7.63)97.79 (0.37)93.82 (3.82)98.37 (1.07)97.67 (4.17)98.80 (0.02)
OA53.30 (3.23)70.09 (4.51)67.70 (5.14)71.20 (3.52)88.25 (2.45)85.58 (3.35)63.25 (3.46)90.32 (3.58)
AA52.18 (2.78)69.08 (4.43)69.53 (3.88)82.83 (1.89)90.49 (1.93)86.51 (2.53)77.55 (1.61)93.04 (2.75)
Kappa47.57 (3.44)66.35 (4.72)63.67 (5.52)67.50 (3.90)86.63 (2.77)83.65 (3.75)58.79 (3.59)88.96 (4.02)
Table 3. Classification performance of all methods on Salinas dataset. The bold denotes the best classification accuracy.
Table 3. Classification performance of all methods on Salinas dataset. The bold denotes the best classification accuracy.
ClassesSVMIFRFEMAPSCMKMSTVPCAEPFsGTROur Method
199.13 (0.97)99.95 (0.16)99.98 (0.05)97.53 (5.50)100.0 (0.00)100.0 (0.00)97.33 (2.25)100.0 (0.00)
298.74 (1.00)98.66 (1.31)99.75 (0.14)98.90 (3.48)99.97 (0.05)99.95 (0.08)99.84 (0.32)100.0 (0.00)
379.50 (8.77)98.34 (1.93)94.55 (2.25)96.40 (7.65)97.87 (2.83)97.72 (2.38)85.58 (8.80)99.47 (0.10)
496.00 (2.41)91.59 (3.95)95.93 (1.39)90.01 (8.71)96.90 (1.79)93.19 (4.02)99.83 (0.07)97.03 (0.28)
594.01 (7.06)97.08 (1.77)96.66 (6.27)97.93 (1.56)96.86 (1.49)98.98 (3.21)90.12 (6.82)99.96 (0.01)
699.79 (0.53)100.0 (0.00)99.41 (0.73)99.75 (0.00)98.34 (2.92)99.98 (0.04)98.89 (2.86)99.97 (0.01)
795.27 (3.61)97.43 (2.27)96.27 (3.16)99.85 (0.06)96.23 (5.52)99.92 (0.06)99.18 (0.65)99.83 (0.01)
864.94 (4.82)91.99 (3.74)80.00 (7.80)69.96 (8.23)92.73 (5.95)93.66 (6.53)69.49 (11.96)98.22 (2.96)
998.66 (0.87)98.95 (0.67)98.99 (0.21)99.91 (0.19)98.61 (1.03)99.58 (0.46)98.84 (0.95)99.33 (0.67)
1077.21 (6.38)97.51 (1.62)87.98 (4.27)82.44 (14.88)95.02 (7.28)99.31 (0.97)80.51 (10.31)98.98 (0.19)
1182.43 (10.35)93.20 (3.51)79.41 (10.37)94.93 (5.93)99.65 (0.78)95.04 (4.63)96.43 (2.38)97.42 (7.25)
1289.11 (7.96)94.98 (5.23)88.64 (6.36)91.05 (12.43)99.11 (1.13)94.71 (4.32)98.12 (5.45)98.62 (2.48)
1379.04 (12.81)83.39 (11.25)92.32 (3.45)93.94 (10.28)95.32 (6.68)91.81 (11.56)98.79 (0.71)81.85 (6.29)
1485.09 (12.77)84.15 (17.99)97.23 (1.08)88.92 (3.95)89.54 (10.17)92.54 (10.36)91.01 (5.19)94.81 (6.19)
1546.93 (5.71)70.34 (13.57)56.84 (7.75)82.66 (11.52)84.53 (7.06)84.52 (8.47)65.16 (14.05)95.59 (4.72)
1693.27 (5.57)99.12 (0.76)96.43 (2.26)93.83 (10.51)97.68 (4.67)99.96 (0.08)86.38 (7.44)100.0 (0.00)
OA80.08 (1.96)90.67 (3.33)85.54 (2.57)88.70 (2.84)94.46 (1.47)95.12 (1.37)85.59 (3.54)98.13 (0.59)
AA86.19 (1.07)93.54 (1.84)91.27 (1.07)92.38 (2.16)96.14 (1.18)96.30 (0.94)90.96 (1.67)97.57 (0.75)
Kappa77.87 (2.15)89.66 (3.65)83.97 (2.81)87.47 (3.15)93.84 (1.64)94.56 (1.53)83.97 (3.94)97.92 (0.67)
Table 4. Classification performance of all methods on Honghu dataset. The bold denotes the best classification accuracy.
Table 4. Classification performance of all methods on Honghu dataset. The bold denotes the best classification accuracy.
ClassesSVMIFRFEMAPSCMKMSTVPCAEPFsGTROur Method
188.1798.2291.0486.9197.7389.8475.4189.69
254.7367.7964.5486.2373.3178.2857.3657.95
388.5196.1692.6293.4493.2796.6461.1397.83
496.1299.3198.3087.6499.5099.5644.6699.72
517.8557.8423.4998.6663.2452.0755.5379.29
685.1890.8389.9789.7493.7893.2669.4199.40
774.0089.3881.9577.2489.6288.4450.9395.81
86.0916.0510.7390.0243.8731.3942.7497.99
991.6099.2189.6794.6190.2386.7990.7597.82
1049.2973.4573.6276.5991.8883.7232.9999.44
1128.1758.8241.5077.4369.8864.2821.3679.63
1243.3361.2052.5693.6464.1861.5443.7796.26
1350.5877.5470.8670.7379.5281.9930.3878.44
1443.3570.7360.0186.7790.1478.7761.2980.40
153.9731.7323.4697.2457.0551.6982.8091.99
1680.9393.5187.5794.5399.0198.6682.3498.37
1754.3972.8262.1793.1085.5392.6673.8097.09
1821.8332.8539.2795.0872.9067.5374.9168.77
1948.8864.4551.5264.0683.4158.4156.2374.37
2038.0054.1959.3499.8679.5965.6046.0688.71
2111.0424.3238.85100.0086.1143.4671.1491.70
2221.5047.3425.7399.5582.9076.2173.2389.30
OA64.4384.2376.1186.4190.7487.3751.8794.01
AA49.8867.1760.3988.1881.2174.5859.0188.64
Kappa57.6880.2770.8285.8688.3784.2345.7992.47
Table 5. Classification performance of three different views. The bold denotes the best classification performance.
Table 5. Classification performance of three different views. The bold denotes the best classification performance.
Local-ViewIntra-ViewInter-ViewOAAAKappaTime (s)
87.2285.0885.513.68
86.2286.7884.383.87
87.3590.6685.633.95
88.3289.4186.714.17
88.9691.2187.434.31
88.2592.0786.584.48
90.3293.0488.965.36
Table 6. The computing time of all considered approaches for all datasets. The bold denotes the best computing efficiency.
Table 6. The computing time of all considered approaches for all datasets. The bold denotes the best computing efficiency.
DatasetsSVMIFRFEMAPSCMKMSTVPCAEPFsGTROur Method
Indian Pines5.532.323.254.494.352.672.045.36
Salinas21.082.684.563.2312.3412.982.4919.35
Honghu128.2616.5569.9415.7545.6720.799.37102.23
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liang, N.; Duan, P.; Xu, H.; Cui, L. Multi-View Structural Feature Extraction for Hyperspectral Image Classification. Remote Sens. 2022, 14, 1971. https://doi.org/10.3390/rs14091971

AMA Style

Liang N, Duan P, Xu H, Cui L. Multi-View Structural Feature Extraction for Hyperspectral Image Classification. Remote Sensing. 2022; 14(9):1971. https://doi.org/10.3390/rs14091971

Chicago/Turabian Style

Liang, Nannan, Puhong Duan, Haifeng Xu, and Lin Cui. 2022. "Multi-View Structural Feature Extraction for Hyperspectral Image Classification" Remote Sensing 14, no. 9: 1971. https://doi.org/10.3390/rs14091971

APA Style

Liang, N., Duan, P., Xu, H., & Cui, L. (2022). Multi-View Structural Feature Extraction for Hyperspectral Image Classification. Remote Sensing, 14(9), 1971. https://doi.org/10.3390/rs14091971

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop