Next Article in Journal
Neotectonics of the Western Suleiman Fold Belt, Pakistan: Evidence for Bookshelf Faulting
Previous Article in Journal
Point-Line Visual Stereo SLAM Using EDlines and PL-BoW
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Classification Based on Sparse Superpixel Graph

1
State Key Laboratory of Resources and Environmental Information System, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Sciences, Beijing 100101, China
2
College of Resources and Environment, University of Chinese Academy of Sciences, Beijing 100049, China
3
Collaborative Innovation Center of South China Sea Studies, Nanjing University, Nanjing 210093, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(18), 3592; https://doi.org/10.3390/rs13183592
Submission received: 3 August 2021 / Revised: 2 September 2021 / Accepted: 7 September 2021 / Published: 9 September 2021
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Hyperspectral image (HSI) classification is one of the major problems in the field of remote sensing. Particularly, graph-based HSI classification is a promising topic and has received increasing attention in recent years. However, graphs with pixels as nodes generate large size graphs, thus increasing the computational burden. Moreover, satisfactory classification results are often not obtained without considering spatial information in constructing graph. To address these issues, this study proposes an efficient and effective semi-supervised spectral-spatial HSI classification method based on sparse superpixel graph (SSG). In the constructed sparse superpixels graph, each vertex represents a superpixel instead of a pixel, which greatly reduces the size of graph. Meanwhile, both spectral information and spatial structure are considered by using superpixel, local spatial connection and global spectral connection. To verify the effectiveness of the proposed method, three real hyperspectral images, Indian Pines, Pavia University and Salinas, are chosen to test the performance of our proposal. Experimental results show that the proposed method has good classification completion on the three benchmarks. Compared with several competitive superpixel-based HSI classification approaches, the method has the advantages of high classification accuracy (>97.85%) and rapid implementation (<10 s). This clearly favors the application of the proposed method in practice.

1. Introduction

Although continuous development of hyperspectral sensors makes it easy to collect a large amount of hyperspectral data, labeling the acquired data is expensive [1]. As a result of this fact, semi-supervised hyperspectral image (HSI) classification methods have been extensively investigated in the past decades [2,3,4,5]. In turn, the successful application of HSI classification in urban mapping [6], environment monitoring [7], precision agriculture [8] and other fields promotes the improvement of classification methods. Currently, it still remains an issue worthy of further study to develop new semi-supervised methods to classify an HSI fast and accurately, due to its characteristics of big data [9,10].
Unlike big data recorded in other fields, one of the remarkable features of hyperspectral data is that it clearly contains spatial structure information in addition to spectral information. This means that classification results may be improved by integrating spatial and spectral information based on classification techniques or classifier. Satisfactory classification results of plenty of existing spectral-spatial HSI classification methods have demonstrated the feasibility of this integration [11,12,13,14,15,16]. Using fixed-size window technique, the methods of Markov random field [17], guided filter [18], discontinuity preserving relaxation [19,20], and recursive filtering [21] successfully adopted spatial information to smooth the noisy pixels contained in the HSI. The use of these smoothing techniques in classification results in an effective reduction in the number of misclassified pixels. In multiple kernels-based classification methods, the spatial and spectral information was effectively fused by using composite kernels to enhance the classification of hyperspectral images [22,23]. The spatial structure information was also used to correct some misclassified pixels in the classification map by majority voting rule [24,25]. Multi-scale covariance maps were introduced to solve the overfitting problem of convolutional neural network-based methods in HSI classification [26]. Additionally, the existing works show that in the classification process, the combination of spatial information with dimensionality reduction [27], sparse representation [28,29], low rank representation [30], convolutional neural network [31] and various techniques do contribute to improving the classification accuracy.
Among a mass of spectral-spatial HSI classification methods, the superpixel-based approach is a representative method and has been extensively investigated in recent years because superpixels are homogeneous and adaptive in size and shape [32,33,34,35]. Compared with fixed-size window, the adaptive size and shape are more favorable to explore the spectral and spatial relationships in HSI classification. Based on superpixels and multi-view active learning, an HSI classification method was proposed, aiming at improving the classification performance and reducing the manual labeling effort [36]. Under the assumption of superpixel homogeneity, a series of superpixel-level HSI classification methods have been introduced in recent years [34,37,38,39,40]. Xie et al. [37] developed a semi-supervised spectral-spatial classification approach at the superpixel-level by the improved simple linear iterative clustering and k-nearest neighbor classifier. Using k-nearest neighbor rules twice, Tu et al. [39] defined the distance between superpixels to explore the optimal representations of superpixels and then suggested a superpixel-level HSI classifier. One of the advantages of superpixel-level classifier is to quickly implement the classification process through reducing an HSI from the aspect of pixels. However, the adaptive size and shape feature of superpixels makes it arduous to measure the similarity between superpixels. Therefore, although there are many superpixel-based classification methods, few works are dedicated to superpixel-level classification algorithms.
A great deal of work has shown that graph is a powerful tool in the fields of machine learning, computer vision and remote sensing, due to its surprisingly flexible representation ability [41,42,43,44,45]. By means of graph theory, label propagation rule or complex network theory, HSI classification is successfully transformed into a graph partition in graph-based HSI classification methods. Camps-Valls et al. early suggested a semi-supervised graph-based classifier specifically designed for processing hyperspectral images [46]. Based on superpixels and a spectral–spatial sparse graph, a semi-supervised adaptive label propagation algorithm was depicted to address the pollution of label information by random noise and boundary noise [47]. Using the spectral and spatial information extracted from superpixels, Sellars et al. [40] constructed a weighted superpixel graph, where each vertex represents a superpixel. Then a graph-based semi-supervised HSI classification framework was proposed. Recently, graph convolutional network (GCN) schemes and its various improvements have been proposed to classify HSI [48,49,50,51]. GCN adopts graph to carry out the convolution on arbitrarily structured non-Euclidean data to improve the classification performances of convolutional neural network with fixed size and weight convolution kernel. In these graph-based HSI classification methods, various techniques for constructing and partitioning graphs were adopted to achieve accurate classification of HSI. In general, the performance of graph-based classification methods relies on the construction and division of graphs. It is known that superpixel graphs usually have smaller sizes than those with pixels as vertices. This means that for the same label propagation on graph algorithm, it will take less time to complete the classification task. Therefore, it is interesting to further develop novel superpixel graphs based HSI classification methods.
The applications of discrete potential theory on graphs have been previously investigated in various fields, such as image segmentation [52], text data classification [53], community structure detection [54] and so on. Motivated by random walks and based on discrete potential theory, Grady proposed a novel algorithm for image segmentation [52]. An effective semi-supervised algorithm was developed to detect the community structure in complex networks through discrete potential theory [54]. The potential theory on directed graph was discussed to discover factors of the network formation [55]. The difference between different applications makes it meaningful to further extend this theory to more fields, especially to remote sensing.
Inspired by the works on superpixel-level classifier, label propagation on graph and discrete potential theory, in this work, a novel superpixel graph-based scheme is proposed to classify HSI efficiently. In this proposal, a novel and computationally simple distance is first defined to measure the similarity of two superpixels. A sparse superpixel graph (with superpixels as vertices) is then constructed by local spatial connection and global spectral connection. Based on the constructed superpixel graph and discrete potential theory, the HSI classification is converted into an optimization problem that can be solved by a system of linear algebraic equations. We here adopt the conjugate gradient descent method (CGDM) to find the approximate solution because of its rapid convergence and strong robustness. Finally, according to the greatest potential of each vertex computed by the CGDM, superpixels can be properly labeled. A range of experimental and comparative results on three benchmarks extensively validate our scheme.
The main technical contributions of the proposal include:
  • An efficient HSI classification scheme is suggested based on sparse superpixel graph.
  • A computationally simple but effective distance between superpixels is newly defined.
  • A sparse superpixel graph is constructed by using spectral-spatial connection strategy.
  • The use of CGDM in the proposal speeds up the process of label propagation on graph.
The rest of this paper is organized as follows: Section 2 details the process of the proposed method, including superpixel segmentation, the new definition of distance between superpixels, the construction of sparse superpixel graph and label propagation on graph. Experimental and comparative results are reported in Section 3. Section 4 analyzes the influence of the parameters used in our method. Finally, discussions and conclusions are given in Section 5 and Section 6.

2. Methods

In this section, we describe the proposed method in detail. The classification framework includes dimension reduction, superpixel segmentation, construction of sparse superpixel graph and label propagation on graph, as shown in Figure 1.

2.1. Superpixel Segmentation

Superpixel segmentation is often understood as the problem of region localization of an image. Generally, the number of pixels contained in each superpixel is automatically generated in terms of local image structure by using superpixel segmentation algorithms. Thus, different superpixels may have different shapes. These adaptive features of superpixels are obviously propitious to good homogeneity of the generated superpixels.
Among many segmentation algorithms, the graph-based entropy rate superpixel segmentation (ERS) [56] is commonly used in remote sensing because of its excellent performance in good effectiveness and high efficiency. However, the high-dimension property of hyperspectral data does not allow to divide an HSI into superpixels by using classical ERS straightly. Therefore, we need to execute principal component analysis approach in advance and take the first principal component to generate the base image used for segmentation.
The ERS method is an optimization algorithm based on edge clustering. For user-defined number of superpixel p and a set of edges M, the ERS approach splits the base image into superpixels through optimizing the following objective function with restrict conditions according to set M [56]:
{ m a x M ( R ( M ) + ω B ( M ) ) s u b j e c t M E N M p , |
where R(M) and B(M) are entropy rate term and balancing term, respectively. NM indicates the number of the obtained superpixels; ω is a balancing parameter, and E is an edge set.
In Equation (1), the introduction of entropy rate term is to include spatial neighbor pixels with similar spectral information as much as possible in a superpixel. However, this may cause significant differences in superpixel size. The use of the balance term and balancing parameter effectively avoids this phenomenon. Interested readers can consult with Liu’s work for more details of ERS method [56].

2.2. Distance between Superpixels

It is one of the key issues to properly define the distance between superpixels in the superpixel-level HSI methods. A desired distance should not only better measure the intimate relationship between two superpixels but also should be computationally simple. However, this is a nontrivial task because superpixels have different sizes. In [34,37,38,39,40], the distance/similarity of a pair of superpixels was defined by using various techniques. Although the similarity of two superpixels can be better measured by these designed distances, it is time-consuming to calculate them.
To address this problem, a simple but effective distance between superpixels is defined in this work. In the suggested distance, we carefully consider the statistical features of a superpixel, that is, mean, median and mode. Specifically, for a generated superpixels S i containing n i pixels, we first construct the mean vector α 1 i , the median vector α 2 i and the mode vector α 3 i by orderly computing the mean, the median and the mode of each band of these n i pixels. The superpixel S i is then approximately represented by a newly defined sample with the same dimension as the original HSI,
α i = w 1 α 1 i + w 2 α 2 i + ( 1 w 1 w 2 ) α 3 i
where w 1 and w 2 are the weights that control the importance of each component.
Finally, a novel distance between superpixels S i and S j can be defined as
d ( S i , S j ) = d ( α i , α j ) = h = 1 B ( α i , h α j , h ) 2 .
where B is the number of bands.
In Equation (3), distance between the vectors α i and α j may choose the desired distance, for example, the commonly used Euclidean distance. Furthermore, the adoption of the median vector and the mode vector in Equation (2) effectively reduces the impact of noise pixels on distance computation. The advantage of the distance defined here is simple to calculate because it is independent of the shape and size of the superpixels. The validity of the above definition has been widely verified in the experiments carried out in this work.

2.3. The Construction of Sparse Superpixel Graph

A majority of existing work on HSI classification demonstrates that the graph-based methods have a promising future, partially due to amazing data presentation ability of graphs. In the field of remote sensing, a vertex of the graphs may represent a pixel or a superpixel. We term a graph as a superpixel graph if the vertices in a graph represent superpixels. It is clear that superpixel graph has a smaller size than the one with pixels as vertices. This means that it is possible to take less time to partition a graph. Considering this advantage of superpixel graph, we use a superpixel graph to represent an HSI for classification.
It is known that there is abundant spatial information in a given HSI, besides rich spectral information. Notably, the effective fusion of spectral and spatial information is helpful to improve the classification results. Therefore, we adopt the method of global spectral connection and local spatial connection to construct desired superpixel graph by using KNN rule twice.
During the spectral connection, we firstly calculate the distances of the superpixel S i to the rest according to Equation (3), and then arrange them in an ascending order. For user-specified k 1 , we finally connect the superpixel S i to its first k 1 superpixels. In this stage, we only consider the spectral information of superpixels and search the nearest neighbors of the superpixel in a global range. This connection approach attempts to deal with the problem that the superpixels belonging to the same class are disconnected or far from each other in space.
In the process of local connection, for each superpixel, one connects it to k 2 nearest neighbors in all its spatial adjacent superpixels in terms of Equation (3). This link strategy aims to maintain local consistency by density connections.
The constructed graph is an efficient combination of spectral-based graph and spatial-based graph. Please note that a superpixel may be linked twice in global and local connections. It has no effect on the constructed superpixel graph because we only record the link regardless of the number of links.
The superpixel graph constructed here is an unweighted and undirected sparse graph. This provides the basis to predict the classes of label-free vertices using a variety of classical or popular label propagation techniques. In next subsection, we would like to use the discrete potential theory and conjugate gradient descent algorithm to predict the class of each unlabeled vertex in the constructed graph.
Let G = (V, E) be the constructed superpixel graph, where V = { v 1 , v 2 , , v p } is the set of vertices and E V × V is the collection of edges. Commonly, we use an adjacent matrix A to describe the graph G, where the element a i , j of A is equal to 1 if there is an edge between vertices v i and v j and 0 otherwise.
The Laplace matrix of graph G is usually defined as
L = D A ,
where the diagonal matrix D is the degree matrix of A. The matrix L is obviously symmetry and positive semi-definite.

2.4. Label Propagation on Graph

A graph can be regarded as an electrostatic field if one or more vertices are assigned to an electric potential 1 and some vertices to 0. Particularly, the electrostatic field is termed as m-label generation if we assign an unit potential to the vertices with label m and zero to other labeled vertices with class label rather than m. This electrostatic field is simply denoted by m-EF. The potentials of the reminding unlabeled vertices can be calculated by solving a system of linear algebraic equations derived from combinatorial Dirichlet problem with the boundary conditions [52]. In this sense, the process of potential transmission in the electrostatic field may be considered as the procedure of label propagation on the graph.
In a continuous situation, for a field u and a region Ω , the Dirichlet integral is defined as [52,54]
D ( u ) = 1 2 Ω | u | 2 d Ω .
The Dirichlet problem is to find a harmonic function satisfying Laplace equation and its boundary condition and minimize the Dirichlet integral.
The graph based discrete combinatorial formulation of Equation (5) can be expressed as [54]
Min : D ( X ) = 1 2 X T LX ,
where p-dimensional vector X denotes the potentials of all vertices. Now, the aim is to search for a vector X to minimize Equation (6).
It is easy to divide the set of vertices V into two groups, V L (labeled vertices) and V U (unlabeled vertices) such that V L V U = and V L V U = V . Without loss of generality, we put the labeled vertices in front, then label-free vertices. In light of matrix theory, Equation (6) may also be decomposed into the following form
D ( X U ) = 1 2 ( X L T X U T ) ( L L H H T L U ) ( X L X U ) = 1 2 ( X L T L L X L + 2 X U T H T X L + X U T L U X U ) ,
where the vectors X L and X U denote the potentials of labeled vertices and unlabeled vertices, respectively. Differentiating D X U with respect to X U and setting it to zero yield an algebraic system [54]
L U X U = H T X L .
In Equation (8), the matrix L U is sparse, symmetric and positive definite. The vector X L is composed of 1 or 0 which can be easily determined by an m-EF (regardless of the order of the labeled vertices). Among many methods to solve a system of linear algebraic equations, the conjugate gradient descent method (CGDM) has been widely used in various engineering problems because of its advantages of rapid convergence, less space requirement and strong robustness [54,57]. In this work, we would like to adopt the CGDM to solve the above equation. Therefore, the potentials of unlabeled vertices can be obtained by the solution of Equation (8).
Assume that the given HSI consists of C different classes and at least one pixel in each class is labeled. The superpixel will be assigned a class label if one or more pixels in this superpixel are labeled. Thus, based on this assumption, C such systems are deduced by using the proposed method. Solving these C systems, for each unlabeled vertex v i , one can obtain its C potentials, that is, x i 1 , x i 2 , , x i C . Under the restriction condition m = 1 C x i m = 1 (for any v i ), the potential x i m may be understood as the probability that vertex v i belongs to the m-th class. Accordingly, we use the following rule to label the unlabeled vertices
C l a s s ( v i ) = a r g m   m a x { x i m } .
In Figure 2, we take a graph with nine vertices as an example to explain our method. It is clear that this graph contains two apparent groups. Marking the vertex v 1 and v 6 as labeled vertices, one can derive two algebraic systems like Equation (8). Using CGDM to solve them, the electrostatic field shown in Figure 2b is generated. Similarly, we can get another electromagnetic field illustrated in Figure 2c. As a result, one can easily classify these vertices into two classes according to Equation (9).
The above labeling rule also means that we only consider the approximate solution of Equation (8). In other words, the iteration process of the CGDM will be terminated early by specifying a relatively large threshold, for example 10−2. Additionally, the realization of finding the solutions to C systems can also be carried out in parallel. These considerations are conducive to the fast implementation of our method.
The proposed algorithm can be summarized as follows.
Algorithm 1. SSG
procedure ( p , k 1 , k 2 , C)
    Initialize A
    Assign parameters to variables
    Call for PCA to generate the base image
    Execute ERS to segment the base image into p superpixels
    for i = 1 to p
       for j = 1 to p
           Compute d ( S i , S j )
       if S j belongs to the first k 1 nearest neighbors of S i then
            a i , j 1
        q i the number of spatial neighbors of S i
       for h = 1 to q i
           Compute d ( S i , S h )
       if S h belongs to the first k 2 nearest neighbors of S i then
            a i , h 1
    Compute D
    L D − A
    for m = 1 to C
       Derive Equation (8) from the m-EF
       Execute CGDM to solve Equation (8)
    Use Equation (9) to assign the class label to each unlabeled vertex
end procedure
The main computation time of the Algorithm 1 is spent on computing PCA, ERS, global spectral connection and CGDM. Suppose the given HSI has n pixels, B bands and C classes and is split into p superpixels. It will take about O( B 2 ), O( n l o g n + p 2 ), O( p 2 B + p 2 l o g p ) and O( p 2 C ) to implement above four steps, respectively. Therefore, the complexity of the proposed method is approximately O( n l o g n + p 2 B + p 2 C ). Compared with pixel-level classification methods with at least O( n 2 B ) complexity ( n p ), this means that our superpixel-level classification scheme is computationally efficient.

3. Experiments

3.1. Description of Three Datsets

The Indian Pines scene, recorded by the AVIRIS sensor in June 1992, contains 16 different types of crops planted in Indian Pines test site in Northwestern Indiana. The image has a spatial resolution of 20 m per pixel and 200 bands of size 145 × 145 (after 20 water absorption bands removing).
Different from Indian Pines image, Pavia University scene is an urban scene around Pavia University, Pavia, Italy. This image acquired by ROSIS satellite sensor in 2001 has 103 bands (after 12 most noisy channels were abandoned) and is of size 610 × 340, with a spatial resolution of 1.3m. It contains nine different urban ground objects.
The Salinas scene was also recorded by the AVIRIS sensor over Salinas Valley, CA, USA, in October 1998. The image has 224 bands, the size of 512 × 217 and a spatial resolution of 3.7m per pixel. Sixteen classes of interest are taken into account for this image. The detailed class information for these three scenes is listed in Table 1.

3.2. Experimental Setup

To validate the performance of our proposal, we have tested the proposed classification method SSG on three common hyperspectral benchmarks, that is, Indian Pines, Pavia University and Salinas scenes. In our experiments, weighted parameters w 1 and w 2 are 0.5 and 0.4, respectively; k 1 is equal to 2 for three images; the connection parameters k 2 is 5 for Pavia University and Salinas images and 6 for Indian Pines image. In our experiments, each class is randomly labeled, and the number of labeled samples and label-free samples are listed in Table 1. The training set consists of all labeled samples. The remaining label-free pixels make up the test set. In order to objectively show the performance of the proposed method, all experiments are reported using an average and standard deviation of ten independent tests. Like in many exiting work, three evaluation criteria, Overall Accuracy (OA), Average Accuracy (AA) and Kappa Coefficient (κ) are employed to evaluate the whole performance of classification methods.
Additionally, we compared the proposed method with several competitors to prove the superiority of our proposal. These comparative algorithms include graph convolutional network (GCN) [48], edge-preserving filters (EPF) [58], image fusion and recursive filtering (IFRF) [21], superpixel-based classification via multiple kernels (SCMK) [22], spectral–spatial HSI classification method at superpixel level (SSC-SL) [37], multiscale dynamic graph convolutional network (MDGCN) [50] and superpixel pooling convolutional neural network (SPCNN) [10]. We re-implemented these seven comparison methods (the authors have shared their codes) which use the same parameter settings as the original. The methods of GCN, MDGCN and SPCNN are based on deep learning; SCMK, SSC-SL, MDGCN and SPCNN are superpixel level HSI classification methods; EPF and IFRF are pixel-wise spectral–spatial HSI classification methods.

3.3. Classification Results

The classification results of seven competitors and our method on Indian Pines image are tabulated in Table 2. The classification accuracy obtained by the five superpixel-level classification methods is better than that of the other three pixel-wise classification algorithms. This may be due to the fact that the spatial spectral information of this image can be more fully explored by using adaptive size and shape of superpixels in the classification. Compared with IFRF, the accuracy of EPF method is relatively low (82.59%), probably because the edge-preserving filtering technique is only used in the classification maps to correct misclassified pixels. The classification accuracy of our method is improved by at least 2% (97.85% vs. 95.83%) contrast to other four superpixel-level competitors, SCMK, SSC-SL, MDGCN and SPCNN. Furthermore, the proposed SSG method outperforms other competitive algorithms in terms of the three evaluation indicators. The visualization of classification results of eight different methods are shown in Figure 3.
For Indian Pines image, there are 145 × 145 (21,025) pixels. Among them, the number of pixels to be classified is 10,249. The classification accuracy 97.83% means that there are about 223 misclassification pixels. Assuming that this image is divided into 1000 superpixels, each one contains approximately 20 pixels. Therefore, at least 21 superpixels are misclassified because of many misclassified small volume superpixels produced after removing the background pixels. This is why the visuals are unsatisfactory.
Table 3 reports the statistical results of eight algorithms on Pavia University scene. With the exception of the GCN method, the classification results of all spatial-spectral methods are satisfactory (greater than 95%) because of the high resolution of the image and the good spatial separation between classes. Both MDGCN, SPCNN and our method have a classification accuracy of more than 98.5%. There is no significant difference in the classification results of these three methods. This shows that the proposed method is comparable to that based on deep learning. Compared with EPF, IFRF, SCMK and SSC-SL methods, the classification accuracy raised by at least 2.6%. Of the eight approaches, the GCN method has the lowest classification accuracy. The reason should be that, on the one hand, this method utilizes only spectral information for pixel-wise classification. On the other hand, the neural network classifier adopted in GCN requires a large number of training samples to achieve satisfactory result. The classification maps of these methods are presented in Figure 4.
Table 4 lists the classification results of various algorithms for Salinas image. The classification accuracy of seven methods is greater than 90% in the case of 1% pixels randomly labeled per class. The classification accuracy of EPF and IFRF methods is 91.55% and 92.06%, respectively, which is lower than that of the other five superpixel based methods. It may be due to the use of dimensionality reduction technique in these two methods, resulting in the loss of some information. There was no significant difference in classification results (from 95. 88% to 96.83%) among of SCMK, SSC-SL and SPCNN methods since they all used spatial adaptation technique. Our method is slightly better than the MDGCN method, which reaches 98.97%. In particular, for the two classes easily misclassified, i.e., the 8-th class and the 15-th class, the average accuracy of 99.52% and 99.26% indicates the superior performance of our method locally. As can be seen from Figure 5, our method shows good classification performance in both the global and local regions.

4. Effect of the Number of Superpixels and Different Number of Training Samples

The classification results of our proposed scheme depend on the choice of superpixel number and labeled sample number. As with many existing superpixel-based methods, the optimal parameter value was obtained as an experimental result, not as a pre-specified one. Therefore, the effect of different superpixel number and different number of training samples on the performance of the proposed SSG approach will be analyzed in this section. The OA values recorded in our experiments are the average of ten independent runs.
In our experiments, the test superpixel number was from 700 to 1800. The number of training samples and test samples is the same as that listed in Table 1. In Figure 6, with the increase of the number of superpixels, the change of OA values shows a similar trend on three scenes, that is, rising first and then decreasing. The difference between them is that the superpixel number corresponding to the optimal OA value and the magnitude of the change are different. The reason may be that a small superpixel number usually means that the generated superpixels have large sizes. This increases the risk that the objects belonging to different classes are located in the same superpixel. In addition, although the superpixel with small size has better homogeneity, it weakens the role of spatial information in classification. The experimental results show that when the number of superpixels is equal to 1000 for Indian Pines and Pavia University images and 1500 for Salinas image, our method provides satisfactory classification results on three scenes.
The performance of semi-supervised classification methods generally relies on the number of training samples. Although excellent classification results can be obtained using a large number of training samples, the marking of samples is expensive. Thus, using a small number of labeled samples to obtain satisfactory classification results can often verify the performance of the classifier. The number of superpixels is identical to that determined in the previous part. Figure 7 displays the impact of different numbers of training sample on several classification methods. The number of randomly labeled pixels per class varies from 5 to 30 for these three scenes. It is easy to see that the classification accuracy of all methods gradually improves as the increase of training pixels. Due to lack of spatial information and the use of neural network in classification, the classification results of GCN were unsatisfactory. Compared with other six spectral-spatial classification algorithms, the proposed scheme can always acquire better performances for different numbers of training samples for different images.
Additionally, we further compared the running time of these classification methods. All experiments were implemented using MATLAB 2018a on the computer with TM i7-6700 CPU, 3.40GHz, 16GB memory and NVIDIA GEFORCE RTX 1660 GPU. The running time here includes all the computational times for PCA, ERS, graph construction, label propagation and classification. Table 5 reports the running time (average of ten runs) of these methods on Indian Pines (IP), Pavia University (PU) and Salinas (SA) images with 30 labeled pixels per class. Three deep learning-based classification framework, GCN, MDGCN and SPCNN, are time-consuming since there are massive parameters to be optimized. One can observe from Table 5 that the computation time of our method is less than 10 seconds on three images and much less than that of other competitors. This satisfactory result benefits from the computational efficiency of the ERS algorithm, the abstract representation of superpixels, the computational simplicity of the graph construction and the rapid implementation of solving algebraic equations by CGDM approach. The results in Table 2, Table 3, Table 4 and Table 5 confirm that the proposed method can accurately and quickly realize the classification of hyperspectral data.

5. Discussions

Based on Table 2, Table 3 and Table 4 and earlier studies [10,20,22,27,32,34,35,36,37,38,39,40], although superpixel-based HSI classification methods can achieve better classification results than pixel-based ones, superpixel-based HSI classification methods often suffer from two problems.
One is how to determine the optimal number of superpixels. As shown in Figure 5, different superpixel segmentation scales will result in different classification results. To date, the determination of the optimal number of superpixels still depends on the experimental results [10,20,22,27,32,34,35,36,37,38,39,40]. Although the multiscale-based approach can partially address this problem [18,33], the problem of how to choose segmentation scales still remains. There may be two ways to address this problem: (1) define an indicator to determine superpixel segmentation scale for a given hyperspectral image; (2) eliminate or alleviate the effect of segmentation scale on the classification results by merging superpixels.
The other is how to properly measure the similarity between two superpixels and pixels. Generally, it is difficult to properly define the distance between the two superpixels due to its characteristic of adaptive shape and size. In [32,37,38,39,40], the authors try to use various techniques to address this problem, such as affine hull model [32,38], covariance matrix [40], KNN-based [37,39]. The computation complexity of the distance between two superpixels is at least O  ( n i × n j × B ) in these studies. In our work, the complexity of calculating the distance between the two superpixels is approximately O ( n i + n j ) × B . As expected, calculating the defined distance is more efficient than existing methods.
In Equation (2), we only consider three commonly used statistical features of superpixels to construct a new sample to approximately represent a superpixel. Perhaps, more factors should be considered in such a representation, for example, the entropy, texture features or variance of superpixels. In addition, it is also important to choose an appropriate metric when calculating the distance between superpixels defined in Equation (3). For example, if the Euclidean distance (ED) in Equation (3) is replaced with spectral angular mapping (SAM) [59], different classification results and running times are obtained, as shown in Table 6. In general, using different metrics may yield different classification results for the same classification algorithm. Thus, it is necessary to choose the appropriate metrics for different data types.
It is also important to select the appropriate metrics when calculating the distance between the superpixels defined in Formula (3).
Additionally, the sparse superpixel graph constructed here is an unweighted graph. Thus, some popular graph partitioning algorithms can be used directly. Theoretically, weighted directed graphs can better represent the close relationship between vertices. However, the intimacy between the two vertices is often asymmetric. In this case, the proposed method cannot work well, and a redesign of new label propagation method is required. Therefore, it is still worth to further explore how to represent hyperspectral data with a weighted directed superpixel graph and study its division.

6. Conclusions

Based on the constructed superpixel graph and discrete potential theory, in this work, we present a superpixel-level semi-supervised HSI classification method. The merits of the proposal are the following: (i) Unlike the existing definition of distance between superpixels, the only use of a vector to approximately represent a superpixel makes it easy and fast to calculate the defined distance between a pair of superpixels. (ii) The classification results of three scenes demonstrate that the strategy of global spectral connection and local spatial connection can better preserve the spectral and spatial relations of the HSI in the constructed superpixel graph. (iii) In the constructed superpixel graph, taking each superpixel as a vertex instead of a pixel actually reduces the hyperspectral data spatially, so that our method can be realized quickly. (iv) The label propagation procedure based on discrete potential theory, the CGDM and approximate solution speeds up the implementation of the scheme again. Experimental and comparative results in this paper, respectively, confirm the validity of the proposed classification scheme and outperform other state-of-the-art methods in terms of classification accuracy and running time.

Author Contributions

Conceptualization, Y.Z. and F.Y.; methodology, Y.Z.; software, Y.Z.; formal analysis, Y.Z. and F.Y.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, F.Y.; visualization, Y.Z.; funding acquisition, F.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grants 41890854). The APC was also funded by the National Natural Science Foundation of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The three hyperspectral images used in our experiments are freely available from http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.

Acknowledgments

We would like to thank the authors of GCN, EPF, IFRF, SCMK, SSC-SL, MDGCN and SPCNN for sharing the source code. In addition, constructive suggestions and valuable comments on the manuscript from the reviewers are appreciated. Special thanks are delivered to Professor Miloš Ilić for his kind help.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, X.; Chen, Y.; Ghamisi, P. Heterogeneous transfer learning for hyperspectral image classification based on convolutional neural network. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3246–3263. [Google Scholar] [CrossRef]
  2. Jiao, L.; Feng, J.; Liu, F.; Sun, T.; Zhang, X. Semisupervised affinity propagation based on normalized trivariable mutual information for hyperspectral band selection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 8, 2760–2773. [Google Scholar] [CrossRef]
  3. Romaszewski, M.; Glomb, P.; Cholewa, M. Semi-supervised hyperspectral classification from a small number of training samples using a co-training approach. ISPRS J. Photo. Remote Sens. 2016, 121, 60–76. [Google Scholar] [CrossRef]
  4. Sellami, A.; Farah, M.; Riadh, I.; Solaiman, F. Hyperspectral imagery classification based on semi-supervised 3-D deep neural network and adaptive band selection. Expert Syst. Appl. 2019, 129, 246–259. [Google Scholar] [CrossRef]
  5. Fang, B.; Li, Y.; Zhang, H.; Chan, J. Collaborative learning of lightweight convolutional neural network and deep clustering for hyperspectral image semi-supervised classification with limited training samples. ISPRS J. Photo. Remote Sens. 2020, 161, 164–178. [Google Scholar] [CrossRef]
  6. Benediktsson, J.; Palmason, J.; Sveinsson, J. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  7. Clark, M.; Roberts, D. Species-level differences in hyperspectral metrics among tropical rainforest trees as determined by a tree-based classifier. Remote Sens. 2012, 4, 1820–1855. [Google Scholar] [CrossRef] [Green Version]
  8. Kanning, M.; Siegmann, B.; Jarmer, T. Regionalization of uncovered agricultural soils based on organic carbon and soil texture estimations. Remote Sens. 2016, 8, 927. [Google Scholar] [CrossRef] [Green Version]
  9. Xie, X.; Li, B. A unified framework of multiple kernels learning for hyperspectral remote sensing big data. J. Inf. Hiding Multim. Signal Process. 2016, 7, 296–303. [Google Scholar]
  10. Xie, F.; Gao, Q.; Jin, C.; Zhao, F. Hyperspectral Image Classification Based on Superpixel Pooling Convolutional Neural Network with Transfer Learning. Remote Sens. 2021, 13, 930. [Google Scholar] [CrossRef]
  11. Barca, E.; Castrignanò, A.; Ruggieri, S.; Rinaldi, M. A new supervised classifier exploiting spectral-spatial information in the Bayesian framework. Int. J. Appl. Earth Obs. Geo. 2019, 86, 101990. [Google Scholar] [CrossRef]
  12. Mukherjee, S.; Prasad, S. A spatial-spectral semisupervised deep learning framework using siamese networks and angular loss. Comput. Vis. Image Underst. 2020, 194, 102943. [Google Scholar] [CrossRef]
  13. Kizel, F.; Benediktsson, J. Spatially Enhanced Spectral Unmixing Through Data Fusion of Spectral and Visible Images from Different Sensors. Remote Sens. 2020, 12, 1255. [Google Scholar] [CrossRef] [Green Version]
  14. Ghamisi, P.; Mura, M.; Benediktsson, J. A Survey on Spectral-Spatial Classification Techniques Based on Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  15. Ghamisi, P.; Benediktsson, J.; Cavallaro, G.; Plaza, A. Automatic Framework for Spectral-Spatial Classification Based on Supervised Feature Extraction and Morphological Attribute Profiles. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2147–2160. [Google Scholar] [CrossRef]
  16. Camps-Valls, G.; Shervashidze, N.; Borgwardt, K. Spatio-Spectral Remote Sensing Image Classification with Graph Kernels. IEEE Geosci. Remote Sens. Lett. 2010, 7, 741–745. [Google Scholar] [CrossRef]
  17. Li, J.; Bioucas-Dias, J.; Plaza, A. Spectral-spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  18. Dundar, T.; Ince, T. Sparse Representation-Based Hyperspectral Image Classification Using Multiscale Superpixels and Guided Filter. IEEE Geosci. Remote Sens. Lett. 2019, 16, 246–250. [Google Scholar] [CrossRef]
  19. Li, J.; Khodadadzadeh, M.; Plaza, A.; Jia, X.; Bioucas-Dias, J. A discontinuity preserving relaxation scheme for spectral–spatial hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. 2017, 9, 625–639. [Google Scholar] [CrossRef]
  20. Xie, F.; Lei, C.; Yang, J.; Jin, C. An effective classification scheme for hyperspectral image based on superpixel and discontinuity preserving relaxation. Remote Sens. 2019, 11, 1149. [Google Scholar] [CrossRef] [Green Version]
  21. Kang, X.; Li, S.; Benediktsson, J. Feature Extraction of Hyperspectral Images with Image Fusion and Recursive Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 3742–3752. [Google Scholar] [CrossRef]
  22. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of Hyperspectral Images by Exploiting Spectral-Spatial Information of Superpixel via Multiple Kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  23. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite Kernels for Hyperspectral Image Classification. IEEE Trans. Geos. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
  24. Xie, F.; Hu, D.; Li, F.; Yang, J.; Liu, D. Semi-Supervised Classification for Hyperspectral Images Based on Multiple Classifiers and Relaxation Strategy. ISPRS Int. J. Geo-Info. 2018, 7, 284. [Google Scholar] [CrossRef] [Green Version]
  25. Tarabalka, Y.; Benediktsson, J.; Chanussot, J.; Tilton, J. Multiple Spectral–Spatial Classification Approach for Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4122. [Google Scholar] [CrossRef]
  26. He, N.; Paoletti, M.E.; Haut, J.M.; Fang, L.; Li, S.; Plaza, A.; Plaza, J. Feature Extraction with Multiscale Covariance Maps for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 755–769. [Google Scholar] [CrossRef]
  27. Xu, H.; Zhang, H.; He, W.; Zhang, L. Superpixel-based spatial-spectral dimension reduction for hyperspectral imagery classification. Neurocomputing 2019, 360, 138–150. [Google Scholar] [CrossRef]
  28. Yang, S.; Hou, J.; Jia, Y.; Mei, S.; Du, Q. Hyperspectral Image Classification via Sparse Representation with Incremental Dictionaries. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1598–1602. [Google Scholar] [CrossRef]
  29. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral Image Classification Using Dictionary-Based Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  30. Zhang, X.; Li, C.; Zhang, J.; Chen, Q.; Feng, J.; Jiao, L.; Zhou, H. Hyperspectral Unmixing via Low-rank Representation with Space Consistency Constraint and Spectral Library Pruning. Remote Sens. 2018, 10, 339. [Google Scholar] [CrossRef] [Green Version]
  31. Kwan, C.; Gribben, D.; Ayhan, B.; Li, J.; Bernabe, S.; Plaza, A. An Accurate Vegetation and Non-Vegetation Differentiation Approach Based on Land Cover Classification. Remote Sens. 2020, 12, 3880. [Google Scholar] [CrossRef]
  32. Cao, X.; Lu, H.; Ren, M.; Jiao, L. Non-overlapping classification of hyperspectral imagery with superpixel segmentation. Appl. Soft Comput. J. 2019, 83, 105630. [Google Scholar] [CrossRef]
  33. Zhan, T.; Lu, Z.; Wan, M.; Yang, G. Multiscale Superpixel Kernel-Based Low-Rank Representation for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1642–1646. [Google Scholar] [CrossRef]
  34. Jia, S.; Deng, X.; Xu, M.; Zhou, J.; Jia, X. Superpixel-Level Weighted Label Propagation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5077–5091. [Google Scholar] [CrossRef]
  35. Fang, L.; Zhuo, H.; Li, S. Super-resolution of hyperspectral image via superpixel-based sparse representation. Neurocomputing 2018, 273, 171–177. [Google Scholar] [CrossRef]
  36. Li, Y.; Lu, T.; Li, S. Subpixel-Pixel-Superpixel-Based Multiview Active Learning for Hyperspectral Images Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4976–4988. [Google Scholar] [CrossRef]
  37. Xie, F.; Lei, C.; Jin, C.; An, N. A novel spectral-spatial classification method for hyperspectral image at superpixel Level. Appl. Sci. 2020, 10, 463. [Google Scholar] [CrossRef] [Green Version]
  38. Lu, T.; Li, S.; Fang, L.; Bruzzone, L.; Benediktsson, J.A. Set-to-set distance-based spectral-spatial classification of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1–13. [Google Scholar] [CrossRef]
  39. Tu, B.; Wang, J.; Kang, X.; Zhang, G.; Ou, X.; Guo, L. KNN-based representation of superpixels for hyperspectral image classification. IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens. 2018, 11, 4032–4047. [Google Scholar] [CrossRef]
  40. Sellars, P.; Avilés-Rivero, A.; Chonlieb, C. Superpixel Contracted Graph-Based Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4180–4193. [Google Scholar] [CrossRef] [Green Version]
  41. Chen, P.; Jiao, L.; Liu, F.; Zhao, Z.; Zhao, J. Adaptive sparse graph learning based dimensionality reduction for classification. Appl. Soft Comput. J. 2019, 82, 105459. [Google Scholar] [CrossRef]
  42. Xue, Z.; Du, P.; Li, J.; Su, H. Sparse Graph Regularization for Hyperspectral Remote Sensing Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2351–2366. [Google Scholar] [CrossRef]
  43. Shao, Y.; Sang, N.; Gao, C.; Ma, L. Spatial and class structure regularized sparse representation graph for semi-supervised hyperspectral image classification. Pattern Recognit. 2018, 81, 81–94. [Google Scholar] [CrossRef]
  44. Ahmadi, S.; Mehrshad, N.; Razavi, S. Semisupervised graph-based hyperspectral images classification using low-rank representation graph with considering the local structure of data. J. Elect. Imag. 2018, 27, 063002. [Google Scholar] [CrossRef]
  45. Chen, M.; Wang, Q.; Li, X. Discriminant Analysis with Graph Learning for Hyperspectral Image Classification. Remote Sens. 2018, 10, 836. [Google Scholar] [CrossRef] [Green Version]
  46. Camps-Valls, G.; Marsheva, T.; Zhou, D. Semi-Supervised Graph-Based Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3044–3054. [Google Scholar] [CrossRef]
  47. Leng, Q.; Yang, H.; Jiang, J. Label Noise Cleansing with sparse Graph for Hyperspectral image Classification. Remote Sens. 2019, 11, 1116. [Google Scholar] [CrossRef] [Green Version]
  48. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  49. Ding, Y.; Guo, Y.; Chong, Y.; Pan, S.; Feng, J. Global Consistent Graph Convolutional Network of Hyperspectral Image Classification. IEEE Trans. Instrum. Meas. 2021, 70, 1–16. [Google Scholar]
  50. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale Dynamic Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3162–3177. [Google Scholar] [CrossRef] [Green Version]
  51. Cai, Y.; Zhang, Z.; Liu, X.; Cai, Z. Efficient Graph Convolutional Self-Representation for Band Selection of Hyperspectral Image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4869–4880. [Google Scholar] [CrossRef]
  52. Grady, L. Random walks for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1768–1783. [Google Scholar] [CrossRef] [Green Version]
  53. Wang, F.; Zhang, C. Semisupervised learning based on generalized point charge models. IEEE Trans. Neural Netw. 2008, 19, 1307–1311. [Google Scholar] [CrossRef]
  54. Liu, D.; Liu, X.; Wang, W.; Bai, H. Semi-supervised community detection based on discrete potential theory. Physica. A 2014, 416, 173–182. [Google Scholar] [CrossRef]
  55. Zhang, Q.; Lü, L.; Wang, W.; Zhou, T. Potential theory for directed networks. PLoS ONE 2013, 8, e55437. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Liu, E.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the CVPR2011, Providence, RI, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  57. Golub, G.; Van Loan, C. Matrix Computations; JHU Press: Baltimore, MD, USA, 2012. [Google Scholar]
  58. Kang, X.; Li, S.; Benediktsson, J. Spectral-spatial hyperspectral image classification with edge-preserving filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
  59. Tumelien, E.; Visockien, J.; Malien, V. The Influence of Seasonality on the Multi-Spectral Image Segmentation for Identification of Abandoned Land. Sustainability 2021, 13, 6941. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed method. PCA: principal component analysis; ERS: entropy rate superpixel segmentation; CGDM: conjugate gradient descent method.
Figure 1. The flowchart of the proposed method. PCA: principal component analysis; ERS: entropy rate superpixel segmentation; CGDM: conjugate gradient descent method.
Remotesensing 13 03592 g001
Figure 2. A graph and two generated electrostatic fields. (a) A graph with nine vertices. (b,c) The electrostatic fields generated by known potentials, respectively. Red digits represent the potential assigned to the labeled vertices, and the other numbers are the potentials calculated by the CGDM.
Figure 2. A graph and two generated electrostatic fields. (a) A graph with nine vertices. (b,c) The electrostatic fields generated by known potentials, respectively. Red digits represent the potential assigned to the labeled vertices, and the other numbers are the potentials calculated by the CGDM.
Remotesensing 13 03592 g002
Figure 3. Indian Pines image. (a) The false-color composite image. (b) Ground-truth. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Figure 3. Indian Pines image. (a) The false-color composite image. (b) Ground-truth. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Remotesensing 13 03592 g003
Figure 4. Pavia University image. (a) The false-color composite image. (b) Ground-truth. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Figure 4. Pavia University image. (a) The false-color composite image. (b) Ground-truth. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Remotesensing 13 03592 g004
Figure 5. Salinas image. (a) The false-color composite image. (b) Reference image. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Figure 5. Salinas image. (a) The false-color composite image. (b) Reference image. The classification maps of (c) GCN, (d) EPF, (e) IFRF, (f) SCMK, (g) SSL-SL, (h) MDGCN, (i) SPCNN and (j) SSG.
Remotesensing 13 03592 g005aRemotesensing 13 03592 g005b
Figure 6. Variation of classification results with different superpixel numbers on three scenes.
Figure 6. Variation of classification results with different superpixel numbers on three scenes.
Remotesensing 13 03592 g006
Figure 7. Change of classification accuracy on three images with the increase of the number of labeled pixels. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Figure 7. Change of classification accuracy on three images with the increase of the number of labeled pixels. (a) Indian Pines. (b) Pavia University. (c) Salinas.
Remotesensing 13 03592 g007
Table 1. Number of training samples and test samples for each class on three images.
Table 1. Number of training samples and test samples for each class on three images.
Indian PinesPavia UniversitySalinas
ClassNameTrainTestNameTrainTestNameTrainTest
1Alfalfa343Asphalt3426489Weeds_1201989
2Corn_no till721356Meadows93317,716Weeds_2373689
3Corn_min till42788Gravel1051994Fallow201956
4Corn12225Trees1532911Fallow_P141380
5Grass/Pasture24459Metal sheets681277Fallow_S272651
6Grass/Trees37693Bare soil2524847Stubble403919
7Grass/Pasture mowed226Bitumen671263Celery363543
8Hay_windrowed24454Bricks1843498Grapes11311,158
9Oats119Shadows48899Soil626141
10Soybean_no till49923 Corn333245
11Soybean_min till1232332 Lettuce_4wk111057
12Soybean_clean30563 Lettuce_5wk201907
13Wheat10195 Lettuce_6wk9907
14Woods641201 Lettuce_7wk111059
15Building_G_T_D20366 Vinyard_U737192
16Stone-steel_T588 Vinyard_T181789
5189731 215240,894 54453,582
Table 2. Classification results (%) of eight methods for Indian Pines image.
Table 2. Classification results (%) of eight methods for Indian Pines image.
GCNEPFIFRFSCMKSSC-SLMDGCNSPCNNSSG
193.65 ± 0.3526.28 ± 6.4394.41 ± 1.8695.25 ± 076.51 ± 3.6790.23 ± 1.4796.38 ± 0.9297.39 ± 1.95
251.36 ± 1.8368.69 ± 5.6486.14 ± 3.6890.07 ± 2.5494.65 ± 2.0193.37 ± 1.2291.96 ± 0.4197.46 ± 0.55
350.30 ± 3.0258.74 ± 2.6284.67 ± 4.6792.89 ± 392.4 ± 1.7192.66 ± 4.7095.80 ± 1.8297.87 ± 1.49
425.41 ± 6.9840.89 ± 18.7173.16 ± 6.6784.51 ± 3.4471.25 ± 9.7594.80 ± 5.6889.61 ± 0.1898.39 ± 2.16
50 ± 091.77 ± 2.5191.18 ± 4.5293.57 ± 4.2092.07 ± 3.6793.34 ± 3.4489.32 ± 2.1695.57 ± 1.87
697.52 ± 1.4299.73 ± 2.7798.44 ± 1.1298.68 ± 1.1599.34 ± 0.7997.89 ± 0.7099.04 ± 0.2999.81 ± 0.06
798.3 ± 2.3493.46 ± 8.0794.62 ± 10.7796.17 ± 1.3597.31 ± 4.0790.77 ± 7.3095.1 ± 2.2896.43 ± 0.24
897.46 ± 0.3799.96 ± 0.13100 ± 099.62 ± 0.1594.58 ± 4.4799.47 ± 0.5699.39 ± 0.02100 ± 0
9100 ± 00 ± 0100 ± 046.62 ± 5.2182.63 ± 5.2257.37 ± 30.9395.06 ± 1.14100 ± 0
1073.58 ± 1.7368.63 ± 4.9387.56 ± 3.3991.36 ± 2.5693.65 ± 1.5493.26 ± 1.4692.48 ± 2.1794.01 ± 0.74
1178.38 ± 5.8293.64 ± 2.9198.03 ± 0.5995.19 ± 1.996.84 ± 0.8597.50 ± 0.7881.02 ± 2.4198.60 ± 0.81
1256.37 ± 2.2754.49 ± 13.2774.46 ± 9.4391.3 ± 2.285.81 ± 4.0892.43 ± 3.1797.83 ± 0.0295.95 ± 1.76
1399.01 ± 0.0999.43 ± 0.1599.07 ± 0.697.42 ± 1.2398.09 ± 2.8599.23 ± 1.2499.43 ± 0.0299.12 ± 0.43
1499.6 ± 0.0899.54 ± 0.2598.75 ± 0.9299.3 ± 0.5698.96 ± 0.6399.78 ± 0.1396.36 ± 0.1799.84 ± 0.08
1576.67 ± 2.4057.9 ± 12.9094.97 ± 3.1192.94 ± 3.7695.33 ± 3.7996.39 ± 5.71100 ± 098.86 ± 0.28
1689.24 ± 2.5498.98 ± 1.0798.86 ± 1.1786.69 ± 6.5171.14 ± 3.8895.80 ± 2.9898.15 ± 0.6095.70 ± 1.52
OA67.86 ± 0.5982.59 ± 1.4291.94 ± 1.3894.21 ± 0.5494.29 ± 0.6195.83 ± 0.3295.26 ± 0.5397.85 ± 0.07
AA70.79 ± 3.2672.01 ± 3.1792.15 ± 3.2890.76 ± 3.0590.05 ± 5.692.77 ± 2.3394.89 ± 1.1697.75 ± 1.80
κ63 ± 0.0280 ± 0.1691 ± 0.4393 ± 0.0691 ± 0.0195 ± 0.0194 ± 0.0198 ± 0.08
Table 3. Statistical results (%) of seven competitive algorithms and the proposed method for Pavia University image.
Table 3. Statistical results (%) of seven competitive algorithms and the proposed method for Pavia University image.
GCNEPFIFRFSCMKSSC-SLMDGCNSPCNNSSG
162.85 ± 12.3799.75 ± 0.4397.54 ± 1.0695.94 ± 0.8299.32 ± 0.1299.24 ± 0.3399.03 ± 0.0699.76 ± 0.08
258.07 ± 2.2399.94 ± 0.2299.85 ± 0.1495.65 ± 0.2999.92 ± 0.0599.73 ± 0.0799.26 ± 0.0399.95 ± 0.04
346.56 ± 9.2872.32 ± 5.3384.08 ± 5.1893.60 ± 4.8778.28 ± 0.2597.25 ± 0.9999.19 ± 0.0499.37 ± 0.48
453.45 ± 8.4595.77 ± 0.6993.04 ± 1.2495.74 ± 0.8396.46 ± 0.2196.22 ± 1.0298.94 ± 0.1688.11 ± 0.65
588.72 ± 2.5199.92 ± 0.2399.83 ± 0.1394.33 ± 0.1599.92 ± 0.2998.56 ± 1.06100 ± 099.73 ± 0.13
659.57 ± 6.2277.25 ± 0.0799.75 ± 0.0396.16 ± 1.2990.24 ± 0.0699.99 ± 0.0399.27 ± 0.0299.90 ± 0.07
788.81 ± 2.2790.50 ± 0.3997.66 ± 1.3292.23 ± 5.8291.53 ± 0.2697.92 ± 1.9899.32 ± 0.04100 ± 0
876.06 ± 2.2698.91 ± 0.8586.41 ± 2.1295.2 ± 1.0698.91 ± 0.1596.14 ± 1.3298.45 ± 0.0498.65 ± 1.51
981.34 ± 4.5296.89 ± 3.3742.67 ± 2.6897.59 ± 0.697.33 ± 0.2893.58 ± 2.2798.26 ± 0.6899.62 ± 0.55
OA75.78 ± 2.6495.14 ± 0.3295.73 ± 0.3296.49 ± 0.1596.47 ± 0.0398.77 ± 0.1398.53 ± 0.0399.12 ± 0.12
AA77.50 ± 3.5792.36 ± 1.2888.98 ± 4.3095.29 ± 2.8694.66 ± 0.1997.63 ± 0.3899.06 ± 0.2898.34 ± 3.64
κ75 ± 0.0493 ± 0.4394 ± 0.2496 ± 0.1996 ± 0.0198 ± 0.0198 ± 0.0199 ± 0.15
Table 4. Classification results (%) of eight spatial-spectral approaches on Salinas.
Table 4. Classification results (%) of eight spatial-spectral approaches on Salinas.
GCNEPFIFRFSCMKSSC-SLMDGCNSPCNNSSG
1100 ± 099.59 ± 1.1099.80 ± 0.02100 ± 0100 ± 0100 ± 0100 ± 0100 ± 0
299.78 ± 0.0599.86 ± 0.1699.21 ± 0.2699.78 ± 097.84 ± 0.1799.85 ± 0.1699.16 ± 0.02100 ± 0
3100 ± 088.18 ± 3.03100 ± 097.11 ± 0.7698.38 ± 5.1399.36 ± 0.93100 ± 0100 ± 0
492.38 ± 0.2599.56 ± 0.1796.27 ± 2.2695.78 ± 3.3396.07 ± 6.4496.12 ± 4.7999.67 ± 0.0495.56 ± 2.73
598.91 ± 0.1699.26 ± 0.3497.79. ± 0.4394.95 ± 0.2196.9 ± 0.296.94 ± 0.8899.21 ± 0.0797.21 ± 1.44
699.67 ± 0.2199.98 ± 0.0199.28 ± 0.0199.69 ± 0.1895.88 ± 0.0599.93 ± 0.0199.02 ± 0.12100 ± 0
799.40 ± 0.0499.75 ± 0.0399.57 ± 0.2195.27 ± 0.3199.85 ± 0.199.48 ± 0.7099.07 ± 0.0499.80 ± 0.13
868.78 ± 0.6392.49 ± 0.0374.34 ± 9.2493.01 ± 2.795.53 ± 1.0199.02 ± 0.8298.45 ± 0.8099.52 ± 0.33
999.90 ± 0.0199.5 ± 0.6299.98 ± 096.67 ± 1.8596.77 ± 0.2999.99 ± 0.0399.81 ± 0.0199.93 ± 0.05
1091.91 ± 0.4390.71 ± 1.6598.94 ± 0.9796.38 ± 1.2496.84 ± 1.7898.89 ± 0.8298.61 ± 0.0798.75 ± 0.40
1199.19 ± 0.0397.68 ± 1.3293.87 ± 3.9394.19 ± 3.3998.19 ± 0.8196.40 ± 1.81100 ± 096.27 ± 2.11
1287.66 ± 0.91100 ± 098.26 ± 2.5697.34 ± 3.23100 ± 093.32 ± 3.9193.42 ± 1.0393.95 ± 1.84
1398.04 ± 0.2897.79 ± 0.4286.24 ± 6.8597.12 ± 096.83 ± 3.1187.48 ± 13.8298.44 ± 0.1198.07 ± 0.19
1495.21 ± 1.3093.54 ± 6.2193.45 ± 1.4695.46 ± 3.6892.96 ± 6.4191.08 ± 3.3793.22 ± 0.0795.44 ± 2.38
1569.01 ± 2.2759.94 ± 2.4789.62 ± 9.5893.5 ± 3.3395.04 ± 1.7398.02 ± 1.7693.69 ± 2.4699.26 ± 0.33
1697.22 ± 0.3597.51 ± 2.9697.51 ± 0.4794.64 ± 4.6694.05 ± 0.3499.11 ± 2.1296.71 ± 0.2597.71 ± 1.97
OA88.67 ± 0.9691.55 ± 1.8792.06 ± 2.0796.01 ± 0.2796.83 ± 0.3698.41 ± 0.3595.88 ± 0.1498.97 ± 0.08
AA88.63 ± 1.0394.51 ± 2.792.59 ± 3.7196.12 ± 1.8096.01 ± 1.7297.19 ± 0.8198.02 ± 1.3698.22 ± 1.92
κ87 ± 0.0191 ± 0.3291 ± 0.5896 ± 0.0395 ± 0.0498 ± 0.0195 ± 0.1299 ± 0.10
Table 5. Running time (seconds) of eight classification methods on three images.
Table 5. Running time (seconds) of eight classification methods on three images.
GCNEPFIFRFSCMKSSC-SLMDGCNSPCNNOURS
IP58.72 ± 0.3613.041 ± 0.364.732 ± 0.116.23 ± 0.0915.09 ± 0.5644.64 ± 0.5365.83 ± 2.921.481 ± 0.01
PU783.06 ± 2.2117.564 ± 1.2212.864 ± 0.5612.97 ± 0.2121.33 ± 1.02154.16 ± 3.86205.86 ± 3.514.267 ± 0.26
SA73.59 ± 1.6427.693 ± 0.2114.122 ± 0.1414.33 ± 0.2123.77 ± 0.6924.53 ± 0.9151.54 ± 3.736.057 ± 0.38
Table 6. Classification results and running time (seconds) on three images.
Table 6. Classification results and running time (seconds) on three images.
Indian PinesPavia UniversitySalinas
SAMEDSAMEDSAMED
OA94.8 ± 2.0197.85 ± 0.0797.97 ± 0.2799.12 ± 0.1292.12 ± 4.2198.97 ± 0.08
AA93.02 ± 9.5897.75 ± 1.8096.36 ± 4.8298.34 ± 3.6488.67 ± 8.9898.22 ± 1.92
κ94.1 ± 2.2898 ± 0.0897.54 ± 0.3999 ± 0.1591.26 ± 4.6599 ± 0.10
Times1.82 ± 0.501.481 ± 0.014.75 ± 0.074.267 ± 0.266.375 ± 0.086.057 ± 0.38
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Yan, F. Hyperspectral Image Classification Based on Sparse Superpixel Graph. Remote Sens. 2021, 13, 3592. https://doi.org/10.3390/rs13183592

AMA Style

Zhao Y, Yan F. Hyperspectral Image Classification Based on Sparse Superpixel Graph. Remote Sensing. 2021; 13(18):3592. https://doi.org/10.3390/rs13183592

Chicago/Turabian Style

Zhao, Yifei, and Fengqin Yan. 2021. "Hyperspectral Image Classification Based on Sparse Superpixel Graph" Remote Sensing 13, no. 18: 3592. https://doi.org/10.3390/rs13183592

APA Style

Zhao, Y., & Yan, F. (2021). Hyperspectral Image Classification Based on Sparse Superpixel Graph. Remote Sensing, 13(18), 3592. https://doi.org/10.3390/rs13183592

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop