Next Article in Journal
Low Blind Zone Atmospheric Lidar Based on Fiber Bundle Receiving
Previous Article in Journal
Spatiotemporal Analysis of Ecosystem Status in China’s National Key Ecological Function Zones
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Decompressed Spectral-Spatial Multiscale Semantic Feature Network for Hyperspectral Image Classification

1
National Key Laboratory of Optical Field Manipulation Science and Technology, Chinese Academy of Sciences, Chengdu 610209, China
2
Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China
3
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
4
China Automotive Engineering Research Institute Co., Ltd., Chongqing 410022, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(18), 4642; https://doi.org/10.3390/rs15184642
Submission received: 21 August 2023 / Revised: 10 September 2023 / Accepted: 18 September 2023 / Published: 21 September 2023

Abstract

:
Convolutional neural networks (CNNs) have shown outstanding feature extraction capability and become a hot topic in the field of hyperspectral image (HSI) classification. However, most of the prior works usually focus on designing deeper or wider network architectures to extract spatial and spectral features, which give rise to difficulty for optimization and more parameters along with higher computation. Moreover, how to learn spatial and spectral information more effectively is still being researched. To tackle the aforementioned problems, a decompressed spectral-spatial multiscale semantic feature network (DSMSFNet) for HSI classification is proposed. This model is composed of a decompressed spectral-spatial feature extraction module (DSFEM) and a multiscale semantic feature extraction module (MSFEM). The former is devised to extract more discriminative and representative global decompressed spectral-spatial features in a lightweight extraction manner, while the latter is constructed to expand the range of available receptive fields and generate clean multiscale semantic features at a granular level to further enhance the classification performance. Compared with progressive classification approaches, abundant experimental results on three benchmark datasets prove the superiority of our developed DSMSFNet model.

1. Introduction

Hyperspectral image (HSI), generally captured by hyperspectral remote sensing sensors or imaging spectrometers, combines subdivisional spectroscopy with 2D imaging technology [1]. HSI utilizes subdivisional spectroscopy to decompose the total radiation of each pixel into the radiation spectrum of different bands and exploits 2D imaging technology to collect the spatial information of the surface, generating a 3D data cube containing rich spectrum, space, and radiation [2,3,4]. Owning to the abundant spatial and spectral information, HSI plays an important role in many practical applications, such as urban planning, land-cover investigation, precision agriculture, military detection, and environmental monitoring [5,6,7,8,9]. In the last few decades, HSI classification has been one of the most popular research fields [10] and attracted multitudinous scholars to devote great efforts to improving classification accuracy [11,12,13,14,15].
In the early period, plentiful HSI classification approaches based on machine learning (ML) utilizing spectral information were presented. Representative methods included support vector machine (SVM) [16], random forest (RF) [17], neural network [18], multinomial logistic regression (MLR) [19], and so on. Although these methods could obtain good classification results, the pure spectra contained much redundant information and noise, which were adverse to the classification performance. Therefore, researchers gave more attention to dimension reduction by using decomposition functions, for example, factor analysis (FA) [20], linear discriminant analysis (LDA) [21], principal component analysis (PCA) [22], and independent component analysis (ICA) [23]. The spatial features included the dependence between the central pixel and its neighborhood pixels, which were introduced into HSI classification to greatly improve the classification performance. An algorithm combining the band selection (BS) with a Markov random field (MRF) was proposed [24]. Jiang et al. designed a random label propagation algorithm (RLPA), which was based on spatial-spectral constraint knowledge, to cleanse the labelled noise [25]. The traditional classification approaches, whether spectral features-based or spatial-spectral features-based, all depended on handcraft features that lacked sufficient representation ability and poor generalization performance.
Of late, deep learning (DL) has developed rapidly and received a substantial amount of attention. Due to excellent performance, DL has been broadly applied to HSI classification. Chen et al. used an autoencoder to classify hyperspectral pixels, which was the first time DL was introduced into the field of HSI classification [26]. Hu et al. introduced 1D CNN into HSI classification for the first time, and the classification accuracy obtained by this method exceeded the traditional ML methods [27]. A 1D CNN model was designed by Li et al., which extracted pixel pairs from the raw HSI as the input data and obtained the spectral relationship between pixels [28]. Although these approaches could obtain good classification results, the input data of them needed to be flattened into a1Dvector, which seriously ignored the rich spatial information. Therefore, many methods integrated the spatial context with spectral information to improve classification accuracy. For example, Yang et al. presented a double-channel CNN to capture the spatial features and spectral features separately [29]. Li et al. developed a two-stream CNN with deep feature fusion to enhance spatial-spectral feature representative power [30]. An end-to-end residual spectral-spatial attention network without additional feature engineering was presented [31]. To reduce the computational complexity and obtain better classification results, Zhang et al. combined 2D CNN with spectral partition [32]. Considering the 3D characteristics of HSI data, plentiful classification algorithms based on 3DCNN were constructed. For example, Wang et al. [33] and Zhong et al. [34] devised spatial and spectral blocks under the construction of 3DCNN to obtain spectral-spatial information. To directly learn joint spatial-spectral features from the original HSI, Li et al. designed a 3D CNN algorithm [35]. To deal with the loss of vast initial information problems, Lin et al. constructed an attention-aware pseudo-3D CNN [36]. The 3D CNN framework was built to capture spatial-spectral joint features [37,38].
With the development of DL, many ancillary strategies have emerged, such as multiscale feature extraction, dense connection, multilayer feature fusion, and residual learning. For example, Gao et al. developed a network framework based on 3D convolution, which not only adopted multiscale blocks to extract spatial-spectral information at different scales but also introduced the dense connection to boost feature propagation and reuse [39]. Xue et al. utilized the shortcut connection structure to learn discriminative spatial-spectral features [40]. Safari et al. designed a multiscale DL learning approach to effectively capture spectral-spatial joint information over different scales [41]. To increase the network depth, Song et al. embedded residual learning into the built-deep feature fusion network [42]. To make full use of HSI’s data, Paoletti et al. presented a deep and dense 3DCNN framework [43]. To achieve complementary spatial-spectral information from different levels, Li et al. combined multilevel fusion with multiattention mechanisms [44].
Although the above-described CNN-based methods obtained promising classification accuracy, two challenging problems remain. The first challenge is that most of the prior works usually focus on wider or deeper network structures to invigorate the discrimination ability of CNNs to capture spatial-spectral features. However, wider or deeper network structures give rise to difficulty for optimization and more parameters along with higher computation, which critically influence the classification performance. The second challenge is how to learn spatial and spectral information more effectively. To resolve the problems mentioned above, this article proposes a decompressed spectral-spatial multiscale semantic feature network (DSMSFNet) for HSI classification. The designed DSMSFNet includes a decompressed spectral-spatial feature extraction module (DSFEM) and a multiscale semantic feature extraction module (MSFEM). The former is utilized to obtain more representative and discriminative global decompressed spectral-spatial features in a lightweight extraction manner, while the latter is constructed to capture clean multiscale semantic features at a granular level and further boost the classification performance. In conclusion, the contributions of this work are twofold as follows:
(1)
To decrease the training parameters and computational complexity, we devise a compressed-weight convolutional (CConv) layer, which takes the place of the traditional 2D convolutional layer, to extract spatial and spectral information through cheap operations.
(2)
To conduct an efficient and lightweight spectral-spatial feature extraction, we construct a compressed residual block (CRB), embedding the CConv layer into a residual block, to alleviate the overfitting and achieve spectral-spatial feature reuse effectively.
(3)
To obtain more representative and discriminative global decompressed spectral-spatial features, we build a decompressed spectral-spatial feature extraction module (DSFEM) in a lightweight extraction manner. For one thing, DSFEM is composed of multiple decompressed dense blocks (DDBs), which provide abundant local decompressed spectral-spatial features. For another thing, the dense connection is introduced into DSFEM to integrate features from shallow and deep layers, thereby acquiring robust complementary information.
(4)
To further enhance the classification performance, we raise a multiscale semantic feature extraction module (MSFEM). The MSFEM can not only expand the range of available receptive fields but also generate clean multiscale semantic features for classification tasks at a granular level.
The remainder of this article is organized as follows: Section 2 provides an elaborate description of the developed HSI classification network model. Section 3 reports the experimental results and discussion. Section 4 summarizes the conclusion of this article and provides an outlook for future research.

2. Method

This article utilizes the Indian Pines dataset as an example to graphically describe the architecture of our constructed DSMSFNet, as exhibited in Figure 1. The developed method includes two main submodules: DSFEM to obtain more representative and discriminative global decompressed spectral-spatial features in a lightweight extraction manner and MSFEM to expand the range of available receptive fields and capture clean multiscale semantic features at a granular level for classification.

2.1. Decompressed Spectral-Spatial Feature Extraction Module

In this article, we devise a decompressed spectral-spatial feature extraction module (DSFEM) to obtain more representative and discriminative global decompressed spectral-spatial features. A DSFEM is composed of five decompressed dense blocks (DDBs) whose nuclear component is a compressed residual block (CRB). For the CRB, the compressed-weight convolutional (CConv) layer is the primary component, which can downgrade the number of training parameters and computational burden. Figure 2 gives the diagram of a DSFEM.

2.1.1. Compressed-Weight Convolution Layer

Recently, pointwise and depthwise convolution operations have attracted considerable attention and have been introduced into many computer vision tasks, such as MobileNet [45], Xception [46], and HSI classification [47]. Pointwise convolution is the standard convolution with a 1 × 1 filter. Only one convolution kernel of depthwise convolution is responsible for one channel. Compared with standard convolution, depthwise convolution can reduce the number of network training parameters and improve the training speed. Figure 3a shows a standard convolution with 3 × 3 filters. Gao et al. [47] presented a mixed depthwise convolution (MDSConv) to replace a 3 × 3 ordinary depthwise separable convolution, as shown in Figure 3b. First, each feature map is divided equally into two parts: one is convoluted by the 1 × 1 depthwise convolution, and the other is convoluted by the 3 × 3 depthwise convolution. Then, the output feature maps of pointwise convolution and depthwise convolution are fused by a concatenation operation. Finally, the connected feature maps are sent into a pointwise convolution. Figure 3c is the architecture of our constructed CConv layer. Different from MDSConv, a 1 × 1 point convolution is first performed to reduce the number of input channels. Then, we employ the 3 × 3 depthwise convolution to take the place of the 3 × 3 standard convolution. Finally, the feature maps from depthwise convolution are concatenated with the previous pointwise convolution feature maps.
Let S i n and S o u t represent the number of input feature maps and output feature maps, respectively. The 3 × 3 standard convolution requires 3 × 3 × S i n × S o u t parameters. For MDSConv, the number of parameters in this layer can be calculated as 3 × 3 × ( S i n / 2 ) + 1 × 1 × ( S i n / 2 ) + 1 × 1 × S i n × S o u t . For CConv, the parameter number of this layer is 1 × 1 × S i n × ( S o u t / 2 ) + 3 × 3 × ( S o u t / 2 ) . For example, we assume that both the number of input feature maps S i n and output feature maps S o u t are 32, and then the parameter numbers of standard convolution, MDSConv, and CConv are 9216, 1184, and 656, respectively. As we can see, our designed CConv requires approximately 14 × fewer parameters than the 3 × 3 standard convolution, and 2 × fewer parameters than MDSConv, demonstrating that the proposed CConv exhibits ascendant capability for reducing the number of network training parameters.

2.1.2. Compressed Residual Block

To capture discriminative spectral-spatial information in a lightweight feature extraction manner, we devise a compressed residual block (CRB), which embeds the CConv layer into a residual block. Figure 4 provides the structure of CRB. CRB is composed of two 3 × 3 CConv layers to extract spatial and spectral information through cheap operations, two BN layers to accelerate the convergence of network, and two ReLU activation functions. In addition, we introduce skip transmission [48] into CRB to increase the network depth and achieve a good generalization performance. Each CRB can be expressed as follows:
Z ( X ) = σ ( F ( X ) + H ( X ) )
where X and Z ( X ) are the input and output of CRB, respectively. σ denotes the ReLU activation function. F ( ) represents a series of operations, including the CConv layer, BN layer, and ReLU. H ( X ) is the identity operator, which is achieved by utilizing the skip transmission.

2.1.3. Decompressed Dense Block

As shown in Figure 2, each DDB comprises a 1 × 1 convolution layer, a BN layer, a ReLU activation function, and a CRB. The 1 × 1 convolution layer is used to reduce the channel numbers and boost the calculation efficiency in this block. CRB is adopted to learn more careful and representative spectral-spatial features. DDB can capture local spectral-spatial information while making the network converge faster. In addition, to obtain more comprehensive global decompressed spectral-spatial features and avoid gradient disappearance, the dense connection [49] is employed between each DDB.

2.2. Multiscale Semantic Feature Extraction Module

During the training process, as the increase in network depth, spectral-spatial information will gradually disappear. The local decompressed spectral-spatial features generated by each DDB are conducive to HSI classification. Therefore, adequately utilizing these features can provide significant classification accuracy. However, only exploiting an ordinary concatenation operation to aggregate these locally decompressed spectral-spatial features may generate noise and redundant information, thereby downgrading the classification accuracies. In this section, we build a multiscale semantic feature extraction module (MSFEM), which can not only expand the range of available receptive fields but also extract clean multiscale high-level semantic features at a granular level to further boost the classification performance. The whole structure of the MSFEM is exhibited in Figure 5.
According to Figure 5, after aggerating all local decompressed spectral-spatial features generated by each DDB, the obtained global decompressed spectral-spatial features first are equally divided into four subsets, represented by x 1 , x 2 , x 3 , and x 4 . Except for x 1 , the other subsets have a corresponding 3 × 3 CConv layer following a BN layer and a ReLU function. Second, to enhance feature reuse of the previous layer, the output features of previous layers and current subset input features are added by elementwise summation. Then, these features are fed into the corresponding 3 × 3 CConv layer to generate new subset features. Furthermore, we utilize the concatenation operation among four new subset features to obtain multiscale high-level semantic features for HSI classification. Finally, to achieve more comprehensive features and avoid information loss, we introduce a “squeeze-and-excitation” block [50] and skip transmission [48] into our devised MSFEM. Mathematically, MSFEM can be described as follows:
z 1 = x 1
z 2 = P ( z 1 + x 2 )
z 3 = P ( z 1 + z 2 + x 3 )
z 4 = P ( z 1 + z 2 + z 3 + x 4 )
z = A t t ( [ z 1 , z 2 , z 3 , z 4 ] ) + X
where X and z are the input and output of MSFEM, respectively. P denotes the 3 × 3 CConv operation. z 1 , z 2 , z 3 , and z 4 represent the output of each subset, respectively. A t t refers to the “squeeze-and-excitation” block. [ ] stands for the concatenation operation.

2.3. The Overall Framework of the Proposed DSMSFNet

The main procedure of our presented DSMSFNet is exhibited in Figure 1. HSI contains hundreds of highly correlated spectral bands, which causes the Hughes effect and thereby impairs the classification performance. Therefore, the first PCA was performed on the original HSI to effectively reduce the spectral dimension and restrain the noisy bands. Second, we selected w × w sized neighborhoods around the target pixels to construct 3-D image cubes, which can fully exploit the property of HSI containing spatial and spectral information. Then, the 3-D image cubes of the size 19 × 19 × 25 were sent to the DSFEM. The DSFEM is composed of five DDBs and each DDB can extract local decompressed spectral-spatial features of the size 19 × 19 × 48 in a lightweight extraction manner. Additionally, we applied the dense connection to the designed DSFEM, which facilitated the local decompressed spectral-spatial features flow and achieved more discriminative global decompressed spectral-spatial features of the size 19 × 19 × 265 . Next, the obtained global decompressed spectral-spatial features were fed to a reduced dimension block, which consists of three 5 × 5 2-D convolutional layers, three BN layers, and three PReLU activation functions, acquiring reduced dimension features of the size 7 × 7 × 24 . Furthermore, we transmitted the reduced dimension features to MSFEM to generate clean multiscale semantic features of the size 7 × 7 × 24 at a granular level, and, thereby, boosted the HSI classification performance. Finally, we utilized a 2D global average pooling layer, two fully connected layers and a softmax function to obtain the output classification maps. In addition, L2 regularization was also introduced into the developed DSMSFNet to improve the classification ability.

3. Experimental Results and Discussion

3.1. Datasets Description

We evaluate the classification performance of our presented approach on three commonly used hyperspectral datasets, including the Botswana (BOW) dataset, the Indian Pines (IP) dataset, and the Houston 2013 dataset.
The BOW dataset [51] was provided by the NASA EO-1 Hyperion sensor over the Okavango Delta. This scene contains 14 diverse land-cover categories and 1476 × 256 hyperspectral pixels. After eliminating uncalibrated and noisy bands, 145 bands with a spatial resolution of 30 m per pixel ranging from 0.4 to 2.5 um remain.
The IP dataset [52] was gathered by the airborne visible/infrared imaging spectrometer (AVIRIS) in northwestern Indiana. This scene is made up of 145 × 145 hyperspectral pixels and 16 different land-cover categories. After removing noisy bands, 200 spectral bands remained with a spatial resolution of 20 m per pixel ranging from 0.4 to 2.5 um.
The Houston 2013 dataset [53] was collected by the 2013 IEEE GRSS Data Fusion Competition. This scene involves 349 × 1905 hyperspectral pixels and 15 different land-cover categories. It has 144 spectral bands with a spatial resolution of 2.5 m per pixel and the wavelength range is from 0.38 to 1.05 um.
Table 1, Table 2 and Table 3 list landcover classes, colors, the numbers of randomly selected training samples, and testing samples.

3.2. Experimental Setup

TensorFlow 2.3.0 was utilized as the DL framework and all experiments were conducted on a PC with an Intel(R) Core(TM) i7-9700F CPU and NVIDIA GeForce RTX 2060 SUPER GPU.
Different hyperspectral datasets contain different annotated sample numbers and exist in the imbalanced data dilemma, so different training sample portions were used for three benchmark datasets. For the BOW, IP, and Houston 2013 datasets, we randomly chose 10%, 20%, and 10% annotated samples for training and the remaining 90%, 80%, and 90% annotated samples for testing, respectively. Adam was used to optimize the parameters and the learning rate was set to 0.001. Additionally, the epoch and batch size are 400 and 16, respectively.
We adopted the overall accuracy (OA), average accuracy (AA), and Kappa coefficient (Kappa) as criteria metrics to measure the classification performance of the proposed DSMSFNet.

3.3. Comparison Methods

The developed DSMSFNet was comprehensively compared with eleven advanced classification algorithms. These comparison approaches can be classified into two categories: one includes SVM, RF, KNN, and GuassianNB, which are based on traditional ML; the other includes HybridSN [54], MSRN_A [55], 3D_2D_CNN [56], RSSAN [31], MSRN_B [47], DMCN [57], and MSDAN [58], which are based on DL. Concretely, HybridSN and 3D_2D_CNN utilize 2D and 3D convolutional layers to extract spectral-spatial features. MSRN_A integrates a 2-D CNN stage involving a spatial attention module and a multiscale spatial feature extraction block with a 3-D CNN stage composed of a spatial-spectral attention module and a multiscale spectral feature extraction block to obtain salient spatial and spectral features. RSSAN utilizes a spectral-spatial attention learning module to find vital spatial and spectral parts and a spectral-spatial feature learning module to acquire important spectral-spatial information. MSRN_B proposes a multiscale residual network with mixed depthwise convolutions to generate discriminative spectral-spatial information. DMCN is composed of a grouped residual 2-D CNN, a dense 3-D CNN, and coordinate attention for acquiring spectral-spatial fusion features. MSDAN uses three multiscale dense connectivity blocks with three different scales to capture multiscale spectral-spatial features and embeds spectral-spatial-channel attention to boost feature representations. To ensure the reliability and authenticity of experimental results, for the BOW, IP, and Houston 2013 datasets, all experiments randomly chose 10%, 20%, and 10% annotated samples as the training set and the remaining 90%, 80%, and 90% annotated samples as the testing set, respectively. Table 4, Table 5 and Table 6 report land-cover class accuracy and three criteria metrics of all experiments on three public datasets. Additionally, to better display the superiority of our presented model for minority category classification, we define a noun named the minority category comparison (MCC), which is the training sample numbers of a category that is smaller than the average training sample numbers.
Table 4 reports each land-cover category and OA, AA, and Kappa accuracies of all comparison approaches for the BOW dataset. It is clear that our developed DSMSFNet obtains superb OA, AA, and Kappa, which demonstrates the stability and superiority of our constructed DSMSFNet. The GuassianNB has the lowest OA, AA, and Kappa, which are 20.96%, 18.66%, and 22.68% lower than those of the DSMSFNet, respectively. This is because the GuassianNB only extracts spectral features and ignores the abundant information in the spatial domain, which results in unsatisfactory classification accuracies. The MSRN_A achieves competitive OA, AA, and Kappa, which are 0.78%, 0.7%, and 0.86% lower than those of the DSMSFNet, respectively. This is because the MSRN_A devises multiscale spectral and spatial feature extraction blocks to acquire spectral-spatial information from diverse scales while utilizing spatial and spatial-spectral attention modules to emphasize the important information and suppress useless ones. In addition, the BOW dataset has five MCCs, including the 1st, 3rd, 7th, 11th, and 13th. Our proposed DSMSFNet acquires decent accuracy on the 3rd, 7th, 11th, and 13th CCMs, which indicates that the designed DSFEM and MSFEM can effectively capture global decompressed and clean semantic spectral-spatial features of the minority class. Similarly, the MSRN_B obtains optimal accuracy on the 3rd, 7th, 11th, and 13th CCMs. But the OA, AA, and Kappa of MSRN_B are 8.29%, 7.61%, and 8.97% lower than those of the DSMSFNet, respectively. This is because although both the MSRN_B and our developed DSMSFNet utilize mixed depthwise convolutions, the DSMSFNet needs fewer training parameters and provides more discriminative features for classification.
Table 5 presents each land-cover category and OA, AA, and Kappa accuracies of all comparison approaches for the IP dataset. We can obviously see that in terms of OA, AA, and Kappa, our proposed DSMSFNet outperforms eleven other comparison approaches. Concretely, the DSMDFNet classifies the most MCCs with high precision and obtains optimal accuracy on the 5th, 7th, 8th, 9th, 12th, and 13th ones, which verifies the effectiveness of the DSMSFNet in identifying the minority categories. Some similar observations can be found on the IP dataset. Among comparison methods, the GuassianNB achieves the most terrible OA, AA, and Kappa, which are 48.79%, 46.61%, and 55.50% lower than those of the DSMSFNet. This is because ML-based classification approaches need to count on prior knowledge, resulting in poor generalization ability. The RSSAN obtains suboptimal OA and AA accuracies, which are 0.55% and 1.47% lower than those of the DSMSFNet. But it does not show good classification performance on the CCMs. In addition, MSRN_A and 3D_2D_CNN have fine classification accuracy on the 1st, 7th, 8th, 9th, and 16th CCMs, which illustrates that they can effectively identify the minority categories. However, the three evaluation indices of MSRN_A and 3D_2D_CNN are not splendid. DSMSFNet achieves 99.62% OA, 99.26% AA, and 99.57% Kappa, which are 1.12%, 2.70%, and 3.16% higher than those of MSRN_A and 2.77%, 4.92%, and 3.16% higher than those of 3D_2D_CNN. These phenomena can adequately show the superiority and robustness of our proposed DSMSFNet.
Table 6 displays each land-cover category and OA, AA, and Kappa accuracies of all comparison approaches for the Houston 2013 dataset. Compared with eight DL-based classification approaches, SVM, RF, KNN, and GaussianNB obtain low criteria metrics. Specifically, our developed DSMSFNet again achieves splendid OA, AA, and Kappa and obtains excellent accuracy on the 3rd, 6th, 13th, and 14th CCMs. Uniformly, the GuassianNB produces unsatisfactory classification performance, and the three criteria metrics of it are 39.06%, 36.73%, and 42.14% lower than those of the DSMSFNet. This is because the eight DL-based approaches can utilize hierarchical feature extraction structures to capture automatically high-level features, thereby acquiring good classification results. Additionally, although the MSRN_A obtains encouraging classification accuracies, the three criteria metrics are 0.52%, 0.76%, and 0.56% lower than those of the DSMSFNet, and it only achieves worthy classification results on the 3rd, 6th, and 14th CCMs. The stabilized classification performance on the three benchmark datasets manifests that our presented DSMSFNet can fully excavate global decompressed spectral-spatial features and clean multiscale semantic features, which are conducive to enhancing the classification performance.
Moreover, the visual classification maps of all comparison approaches for the three benchmark datasets are depicted in Figure 6, Figure 7 and Figure 8, which are highly consistent with Table 4, Table 5 and Table 6. In Figure 6, Figure 7 and Figure 8, we can observe that the visual maps of four ML-based classification approaches are poor and contain a lot of noise; especially for the GuassianNB, the visual classification map has the most misclassified pixels and is the coarsest. By comparison, the visual classification results of the HybridSN, MSRN_A, 3D_2D_CNN, RSSAN, MSRN_B, DMCN, MSDAN, and DSMSFNet are relatively smooth and exhibit less noise. Among them, the visual classification map of the presented DSMSFNet is the smoothest and generates the least noise. To furthermore demonstrate the robust generalization performance of the DSMSFNet, the training sample proportions of eight DL-based classification algorithms are also changed, namely, 1%, 3%, 5%, 7%, and 10%. Figure 9 exhibits the matching classification results. As the number of training samples increases, the classification performance gap between diverse approaches gradually narrows. Our designed approach still acquires terrific classification results and demonstrates robust generalization performance.

3.4. Discussion

3.4.1. Influence of Different Spatial Sizes

The feature distributions and intrinsic structures of the three benchmark datasets are different, and the most suitable spatial sizes for them are different. Therefore, to investigate the impacts of different spatial sizes, we varied them in the grid of { 15 × 15 , 17 × 17 , 19 × 19 , 21 × 21 , 23 × 23 , 25 × 25 , 27 × 27 , 29 × 29 }; the results of the classification are given in Figure 10. In Figure 10, for the BOW and IP datasets, it is clear that the optimal spatial size is 19 × 19 . For the Houston 2013 dataset, when the spatial size is 23 × 23 , the accuracies of three criteria metrics are unexceptionable. This is because a small spatial size results in too little spatial context, while a large spatial size leads to too many pixels of different classes and noises, which indicate that the middle spatial sizes are appropriate for three benchmark datasets. Hence, for the BOW and IP datasets, the proper spatial size was set to 19 × 19 . For the Houston 2013 dataset, the proper spatial size was set to 23 × 23 .

3.4.2. Influence of Diverse Training Percentage

To analyze the classification performance of our constructed DSMSFNet under different training sample proportions on the three benchmark datasets, the labelled samples were chosen randomly in the grid of { 1 % , 3 % , 5 % , 7 % 10 % , 20 % , 30 % } as the training set and the corresponding remaining labelled samples as the testing set; the results of the classification are provided in Figure 11. In Figure 11, we can obviously see that for the three benchmark datasets, the accuracies of three criteria metrics gradually increase as the number of training samples increases. For the BOW and Houston 2013 datasets, when the training sample number exceeds 10%, the accuracies of OA, AA, and Kappa increase slowly and become gradually stable. For the IP dataset, when the training sample number exceeds 20%, the accuracies of OA, AA, and Kappa increase slowly and become gradually stable. This is because the IP dataset involves many continuous areas of the species; it needs more training samples to obtain excellent classification accuracies. While the other two datasets have many smaller areas of the species, they need fewer training samples to acquire decent classification results. Therefore, for the three public datasets, we set, respectively, the optimal training sample proportion to 10%, 20%, and 10%.

3.4.3. Influence of Different Numbers of Principal Components

The hyperspectral datasets consist of hundreds of continuous spectral bands with high correlation, which is adverse to the HSI classification. Therefore, before the DSMSFNet extracts decompressed global spectral-spatial features, we conducted PCA on the original hyperspectral datasets to reduce the training parameters of the model and abundant redundant information by lessening the dimension of the original hyperspectral datasets. To explore the relationship between the classification results and the number of principal components, we also conducted some experiments and varied the number of principal components in the grid of {5, 10, 15, 20, 25, 30, 35, 40}. The classification results are provided in Figure 12. In Figure 12, it can be easily found that for the three benchmark datasets, the three criteria metrics gradually increase as the number of principal components increases. For the BOW dataset, when the principal component number is 35, the three criteria metrics are optimal. When the principal component number is over 35, the three criteria metrics begin to decrease. For the IP and Houston 2013 datasets, when the principal component number is 25, the three criteria metrics are optimal. When the principal component number is over 25, the criteria metrics begin to descend. These phenomena partly explain that the principal component number is greater; the developed DSMSFNet can generate more discriminative spectral-spatial information from these principal components. But if the principal component number is too large, the three criteria metrics are not good, which is due to the interference of noise bands and excessive redundant information. Therefore, for the BOW, IP, and Houston 2013 datasets, we set the optimal principal component number to 35, 25, and 25, respectively.

3.4.4. Influence of Diverse Compressed Ratio in the MSFEM

We applied the “squeeze-and excitation” block to our constructed MSFEM to enhance the useful information and suppress other ones. The number of neurons in the first fully connected layer is determined by the compressed ratio r . We discuss the classification accuracies of our developed method under diverse compressed ratios r , which were set to {1, 2, 3, 4, 5, 6}. The classification results are provided in Figure 13. In Figure 13, we can clearly observe that for the three benchmark datasets, the optimal compressed ratio r is 3, 4, and 6, respectively.

3.4.5. Influence of Different L2 Regularization Parameters

To avoid the overfitting problem, L2 regularization was also introduced into the proposed model. We analyzed the classification results under different L2 regularization parameters, including 0.0005, 0.002, 0.01, 0.02, 0.03, 0.1, and 1. The classification results are provided in Figure 14. In Figure 14, for the three benchmark datasets, as the L2 parameter increases, the three criteria metrics also gradually increase. For the BOW and Houston 2013 datasets, when the L2 parameter is 0.002, the three criteria metrics are superior. When the L2 parameter is over 0.002, the accuracies of OA, AA, and Kappa begin to decline. For the IP dataset, when the L2 parameter is 0.02, the three criteria metrics are optimal. When the L2 parameter is over 0.02, the accuracies of OA, AA, and Kappa begin to decrease. Therefore, for the three common datasets, we set the optimal L2 parameter to 0.002, 0.02, and 0.002, respectively.

3.4.6. Influence of Diverse Convolutional Kernel Numbers of DDBs

The DDB is composed of a 1 × 1 convolutional layer, a BN layer, a ReLU activation function, and a CRB. Among them, the 1 × 1 convolutional layer is utilized to reduce the dimension of local decompressed spectral-spatial features. The dimension of output local decompressed spectral-spatial features of each DDB directly influences the classification performance and computational complexity of the presented method. Therefore, we discuss the classification accuracies of our developed method under diverse convolutional kernel numbers of DDBs, which were set to {12, 24, 32, 48, 64}. Figure 15 provides the classification results. In Figure 15, for the BOW dataset, we can observe that when the convolutional kernel number is 64, the accuracies of three criteria metrics are first-rank. For the IP and Houston 2013 datasets, when the convolutional kernel number is 48, the accuracies of three criteria metrics are at their best and the classification performance of our developed DSMSFNet is excellent. Therefore, for the three benchmark datasets, we set the optimal convolutional kernel number to 64, 48, and 48, respectively.

3.4.7. Influence of Various Numbers of DDBs in the DSFEM

The DSFEM includes multiple DDBs and the number of DDBs directly affects the classification performance of our presented approach. If the number of DDBs is too low, the decompressed spectral-spatial features are inadequately extracted. If the number of DDBs is too high, the complexity and training parameters of the model increase. These are adverse to the HSI classification. Therefore, we conducted some experiments to investigate the impacts of various numbers of DDBs in the DSFEM in the grid of { 2 , 3 , 4 , 5 , 6 } . Figure 16 provides the results of the classification of the three public datasets. According to Figure 16, it is easy to see that for the IP datasets, when the number of DDBs is five, the accuracies of the three criteria metrics are obviously better than the other conditions. For the other two datasets, we can visibly see that the three evaluation indices gradually rise with the increase of DDB numbers. As the number of the DDBs is five, the three criteria metrics are excellent. As the number of the DDBs exceeds five, the accuracies of the three criteria metrics begin to decrease. Therefore, for the three benchmark datasets, we set the optimal number of DDBs to five, five, and five, respectively.

3.5. Ablation Study

3.5.1. The Effect of Constructed CConv Layer

To verify the validity of our constructed CConv layer, the ablation experiments were performed on the three benchmark datasets under six different conditions, named case1, case2, case3, case4, case5, and case6, where “×” represents each corresponding DDB that uses a conventional 3 × 3 convolutional layer instead of CConv layers, and “√” refers to each corresponding DDB that uses CConv layers. Table 7 shows the corresponding results on the three public datasets, including the training parameter number and testing time.
As shown in Table 7, it can be obviously seen that case 1 requires more parameters for training and a longer time for testing on the three benchmark datasets. This is because all cases utilize the CConv layer, except case 1. This means case 1 cannot reduce network parameters and computational complexity. Compared with the other cases, case 6 requires fewer parameters for training and a shorter time for testing on the three benchmark datasets. This is because the CConv layer is the primary component in the CRB. Each DDB can obtain representative local spectral-spatial information, where the main part of our designed DDB is CRB. DSFEM can capture more discriminative global spectral-spatial features and it comprises five DDBs through dense connection. This confirms that the CConv layer is effective and can effectively decrease the computational cost of our proposed model while enhancing the classification performance.

3.5.2. The Effect of the Designed DSMSFNet Model

To further discuss and demonstrate the importance of the DSFEM and MSFEM of our presented DSMSFNet on the three benchmark datasets, we conducted comparative experiments under three conditions, namely, network 1 (only using the DSFEM), network 2 (only using the MSFEM), and network 3 (using the DSFEM and MSFE, i.e., our proposed method). Figure 17 exhibits the corresponding classification results of the three benchmark datasets.
In Figure 17, for the three benchmark datasets, it is obvious that the three criteria metrics of network 2 are the lowest. For example, for the Houston 2013 dataset, the three criteria metrics are 6.66%, 7.26%, and 7.2% lower than those of network 3, respectively. For the IP dataset, the three criteria metrics are 32.06%, 59.25%, and 36.63% lower than those of network 3, respectively. For the BOW dataset, the three criteria metrics are 1.26%, 1.8%, and 1.37% lower than those of network 3, respectively. These experimental results certify that our designed DSFEM is conducive to the constructed DSMSFNet to not only capture adequately local decompressed spectral-spatial features of each DDB but also generate global decompressed spectral-spatial features by introducing dense connections, which further increase the diversity of spectral-spatial information. Compared with network 2, the three criteria metrics of network 1 are perceptibly raised. For example, for the Houston 2013 dataset, the three criteria metrics are 0.06%, 0.02%, and 0.06% lower than those of network 3, respectively. For the IP dataset, the three criteria metrics are 0.74%, 7.73%, and 0.85% lower than those of network 3, respectively. For the BOW dataset, the three criteria metrics are 0.2%, 0.28%, and 0.22% lower than those of network 3, respectively. These results indicate that the MSFEM can expand the range of available receptive fields while capturing multiscale high-level semantic features at a granular level to further boost the classification performance. Therefore, up to a point, the DSFEM and the MSFEM in our proposed model considerably boost the classification performance.

4. Conclusions

This article presents a decompressed spectral-spatial multiscale semantic feature network (DSMSFNet) for HSI classification. First, we design a CConv layer instead of conventional 3 × 3 convolutional layer to decrease the training parameter number and computational complexity of our developed model. Second, taking the CConv layer as the dominating component, a CRB is constructed to capture spectral-spatial features in a lightweight feature extraction manner. Third, we devise a DDB, which is composed of a 1 × 1 convolution layer, a BN layer, a ReLU activation function and a CRB, to generate local decompressed spectral-spatial features. In addition, to avoid gradient disappearance and obtain more comprehensive global decompressed spectral-spatial features, we also introduce the dense connection into our presented DSMSFNet. Furthermore, considering dispelling the redundant information and noise, we raise an MSFEM, which can not only expand the range of available receptive fields but also extract high-level clean multiscale semantic features at a granular level to further boost the classification performance. Finally, a 2D global average pooling layer, two fully connected layers, and a softmax function are utilized to acquire the output results. Additionally, L2 regularization is also introduced into the developed DSMSFNet to improve the classification ability. The experimental results on the three public datasets prove the superiority and effectiveness of our constructed model, while exhibiting competitive performance compared with the advanced classification algorithm.
Nevertheless, there is still room for improvement in our proposed framework. In future work, it is suggested to adaptively design the architecture of a DL model and not rely on specialized knowledge. Another important future avenue of exploration is integrating the semi-supervised or unsupervised training strategies into the presented model while maintaining high-quality classification results.

Author Contributions

Conceptualization, D.L.; validation, J.Z.; formal analysis, D.L.; investigation, D.L. and J.Z.; original draft preparation, D.L.; review and editing, D.L., Q.L., M.L. and J.Z.; funding acquisition, J.Z. and M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant number 62101529.

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  2. Han, X.; Yu, J.; Xue, J.-H.; Sun, W. Hyperspectral and multispectral image fusion using optimized twin dictionaries. IEEE Trans. Image Process. 2020, 29, 4709–4720. [Google Scholar] [CrossRef]
  3. Gu, Y.; Chanussot, J.; Jia, X.; Benediktsson, J.A. Multiple Kernel learning for hyperspectral image classification: A review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
  4. Li, S.; Zhu, X.; Liu, Y.; Bao, J. Adaptive spatial–spectral feature learning for hyperspectral image classification. IEEE Access 2019, 7, 61534–61547. [Google Scholar] [CrossRef]
  5. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef]
  6. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph convolutional networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5966–5978. [Google Scholar] [CrossRef]
  7. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More diverse means better: Multimodal deep learning meets remote sensing imagery classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4340–4354. [Google Scholar] [CrossRef]
  8. Wu, Z.; Zhu, W.; Chanussot, J.; Xu, Y.; Osher, S. Hyperspectral anomaly detection via global and local joint modeling of background. IEEE Trans. Signal Process. 2019, 67, 3858–3869. [Google Scholar] [CrossRef]
  9. Vaglio Laurin, G.; Chan, J.C.; Chen, Q.; Lindsell, J.A.; Coomes, D.A.; Guerriero, L.; Frate, F.D.; Miglietta, F.; Valentini, R. Biodiversity Mapping in a Tropical West African Forest with Airborne Hyperspectral Data. PLoS ONE 2014, 9, e97910. [Google Scholar] [CrossRef]
  10. Zhang, X.; Wang, Y.; Zhang, N.; Xu, D.; Luo, H.; Chen, B.; Ben, G. SSDANet: Spectral-spatial three-dimensional convolutional neural network for hyperspectral image classification. IEEE Access 2020, 8, 127167–127180. [Google Scholar] [CrossRef]
  11. Lin, C.; Wang, T.; Dong, S.; Zhang, Q.; Yang, Z.; Gao, F. Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification. Electronics 2022, 11, 3992. [Google Scholar] [CrossRef]
  12. Savelonas, M.A.; Veinidis, C.V.; Bartsokas, T.K. Computer Vision and Pattern Recognition for the Analysis of 2D/3D Remote Sensing Data in Geoscience: A Survey. Remote Sens. 2022, 14, 6017. [Google Scholar] [CrossRef]
  13. Gong, H.; Li, Q.; Li, C.; Dai, H.; He, Z.; Wang, W.; Li, H.; Han, F.; Tuniyazi, A.; Mu, T. Multiscale information fusion for hyperspectral image classification based on hybrid 2D-3D CNN. Remote Sens. 2021, 13, 2268. [Google Scholar] [CrossRef]
  14. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  15. Li, Q.; Wang, Q.; Li, X. Exploring the relationship between 2D/3D convolution for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8693–8703. [Google Scholar] [CrossRef]
  16. Sun, Z.; Wang, C.; Wang, H.; Li, J. Learn multiple-kernel SVMs domain adaptation in hyperspectral data. IEEE Geosci. Remote Sens. 2013, 10, 1224–1228. [Google Scholar]
  17. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  18. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  19. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial hyperspectral image segmentation using subspace multinomial logistic regression and Markov random fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  20. Meng, Z.; Jiao, L.; Liang, M.; Zhao, F. A Lightweight SpectralSpatial Convolution Module for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5504205. [Google Scholar]
  21. Zhang, C.; Zheng, Y. Hyperspectral remote sensing image classification based on combined SVM and LDA. In Proceedings of the SPIE Asia Pacific Remote Sensing 2014, Beijing, China, 13–16 October 2014; p. 92632P. [Google Scholar]
  22. Licciardi, G.; Marpu, P.; Chanussot, J.; Benediktsson, J. Linear versus nonlinear pca for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef]
  23. Villa, A.; Chanussot, J.; Jutten, C.; Benediktsson, J.; Moussaoui, S. On the use of ICA for hyperspectral image analysis. In Proceedings of the 2009 IEEE International Geoscience and Remote Sensing Symposium, Cape Town, South Africa, 12–17 July 2009; pp. IV-97–IV-100. [Google Scholar]
  24. Li, S.; Jia, X.; Zhang, B. Superpixel-based Markov random field for classification of hyperspectral images. In Proceedings of the 2013 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Melbourne, Australia, 21–26 July 2013; pp. 3491–3494. [Google Scholar]
  25. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral image classification in the presence of noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 851–865. [Google Scholar] [CrossRef]
  26. Chne, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  27. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sensors 2015, 2015, 258619. [Google Scholar] [CrossRef]
  28. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  29. Yang, J.; Zhao, Y.; Chan, J.C.; Yi, C. Hyperspectral image classification using two-channel deep convolutional neural network. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5079–5082. [Google Scholar]
  30. Li, X.; Ding, M.; Pižurica, A. Deep Feature Fusion via Two-Stream Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2615–2629. [Google Scholar] [CrossRef]
  31. Zhu, M.; Jiao, L.; Yang, S.; Wang, J. Residual Spectral-spatial Attention Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 449–462. [Google Scholar] [CrossRef]
  32. Zhang, X.; Shang, S.; Tang, X.; Feng, J.; Jiao, L. Spectral Partitioning Residual Network with Spatial Attention Mechanism for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5507714. [Google Scholar] [CrossRef]
  33. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  34. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  35. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67–87. [Google Scholar] [CrossRef]
  36. Lin, J.; Mou, L.; Zhu, X.; Ji, X.; Wang, Z.J. Attention-Aware Pseudo-3-D Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7790–7802. [Google Scholar] [CrossRef]
  37. Zhao, S.; Li, W.; Du, Q.; Ran, Q. Hyperspectral classification based on Siamese neural network using spectral–spatial feature. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 2087–2090. [Google Scholar]
  38. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  39. Gao, H.; Miao, Y.; Cao, X.; Li, C. Densely Connected Multiscale Attention Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2563–2576. [Google Scholar] [CrossRef]
  40. Xue, Y.; Zeng, D.; Chen, F.; Wang, Y.; Zhang, Z. A new dataset and deep residual spectral spatial network for hyperspectral image classification. Symmetry 2020, 12, 561. [Google Scholar] [CrossRef]
  41. Safari, K.; Prasad, S.; Labate, D. A multiscale deep learning approach for high-resolution hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 167–171. [Google Scholar] [CrossRef]
  42. Song, W.; Li, S.; Fang, L.; Lu, T. Hyperspectral image classification with deep feature fusion network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3173–3184. [Google Scholar] [CrossRef]
  43. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep & dense convolutional neural network for hyperspectral image classification. Remote Sens. 2018, 10, 1454. [Google Scholar]
  44. Li, Z.; Zhao, X.; Xu, Y.; Li, W.; Shi, X. Hyperspectral Image Classification With Multiattention Fusion Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5503305. [Google Scholar] [CrossRef]
  45. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  46. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  47. Gao, H.; Yang, Y.; Li, C.; Gao, L.; Zhang, B. Multiscale Residual Network with Mixed Depthwise Convolution for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3396–3408. [Google Scholar] [CrossRef]
  48. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  49. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A.; Liu, W.; et al. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  50. Jie, H.; Li, S.; Gang, S. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  51. Gao, H.; Zhang, Y.; Chen, Z.; Li, C. A Multiscale Dual-Branch Feature Fusion and Attention Network for Hyperspectral Images Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 8180–8192. [Google Scholar] [CrossRef]
  52. Xu, Q.; Xiao, Y.; Wang, D.; Luo, B. CSA-MSO3DCNN: Multiscale Octave 3D CNN with Channel and Spatial Attention for Hyperspectral Image Classification. Remote Sens. 2020, 12, 188. [Google Scholar] [CrossRef]
  53. Acito, N.; Matteoli, S.; Rossi, A.; Diani, M.; Corsini, G. Hyperspectral airborne “Viareggio 2013 Trial” data collection for detection algorithm assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2365–2376. [Google Scholar] [CrossRef]
  54. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  55. Zhang, X.; Wang, T.; Yang, Y. Hyperspectral image classification based on multi-scale residual network with attention mechanism. arXiv 2020, arXiv:2004.12381. [Google Scholar]
  56. Ahmad, M.; Shabbir, S.; Raza, R.A.; Mazzara, M.; Distefano, S.; Khan, A.M. Hyperspectral Image Classification: Artifacts of Dimension Reduction on Hybrid CNN. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  57. Xiang, J.; Wei, C.; Wang, M.; Teng, L. End-to-End Multilevel Hybrid Attention Framework for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5511305. [Google Scholar] [CrossRef]
  58. Wang, X.; Fan, Y. Multiscale Densely Connected Attention Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 1617–1628. [Google Scholar] [CrossRef]
Figure 1. The overall pipeline of the proposed decompressed spectral-spatial multiscale semantic feature network (DSMSFNet).
Figure 1. The overall pipeline of the proposed decompressed spectral-spatial multiscale semantic feature network (DSMSFNet).
Remotesensing 15 04642 g001
Figure 2. Architecture of a decompressed spectral-spatial feature extraction module (DSFEM).
Figure 2. Architecture of a decompressed spectral-spatial feature extraction module (DSFEM).
Remotesensing 15 04642 g002
Figure 3. Different convolutional operations.
Figure 3. Different convolutional operations.
Remotesensing 15 04642 g003
Figure 4. The structure of a compressed residual block (CRB).
Figure 4. The structure of a compressed residual block (CRB).
Remotesensing 15 04642 g004
Figure 5. Architecture of a multiscale semantic feature extraction module (MSFEM).
Figure 5. Architecture of a multiscale semantic feature extraction module (MSFEM).
Remotesensing 15 04642 g005
Figure 6. The visual maps of comparison methods for the BOW dataset.
Figure 6. The visual maps of comparison methods for the BOW dataset.
Remotesensing 15 04642 g006
Figure 7. The visual maps of comparison methods for the IP dataset.
Figure 7. The visual maps of comparison methods for the IP dataset.
Remotesensing 15 04642 g007
Figure 8. The visual maps of comparison methods for the Houston2013 dataset.
Figure 8. The visual maps of comparison methods for the Houston2013 dataset.
Remotesensing 15 04642 g008
Figure 9. The generalization performance.
Figure 9. The generalization performance.
Remotesensing 15 04642 g009
Figure 10. The influence of different spatial sizes.
Figure 10. The influence of different spatial sizes.
Remotesensing 15 04642 g010
Figure 11. The influence of diverse training percentages.
Figure 11. The influence of diverse training percentages.
Remotesensing 15 04642 g011
Figure 12. The influence of different principal component numbers.
Figure 12. The influence of different principal component numbers.
Remotesensing 15 04642 g012
Figure 13. The influence of diverse compressed ratios in the MSFEM.
Figure 13. The influence of diverse compressed ratios in the MSFEM.
Remotesensing 15 04642 g013
Figure 14. The influence of different L2 regularization parameters.
Figure 14. The influence of different L2 regularization parameters.
Remotesensing 15 04642 g014
Figure 15. The influence of diverse convolutional kernel numbers of DDBs.
Figure 15. The influence of diverse convolutional kernel numbers of DDBs.
Remotesensing 15 04642 g015
Figure 16. The influence of various numbers of DDBs in the DSFEM.
Figure 16. The influence of various numbers of DDBs in the DSFEM.
Remotesensing 15 04642 g016
Figure 17. The ablation result of the designed DSMSFNet model.
Figure 17. The ablation result of the designed DSMSFNet model.
Remotesensing 15 04642 g017
Table 1. The information on the land-cover category of the BOW dataset.
Table 1. The information on the land-cover category of the BOW dataset.
No.ColorClassTrainTest
1Remotesensing 15 04642 i001Water1085
2Remotesensing 15 04642 i002Hippo grass27241
3Remotesensing 15 04642 i003Floodplain grasses 119162
4Remotesensing 15 04642 i004Floodplain grasses 231274
5Remotesensing 15 04642 i005Reeds25223
6Remotesensing 15 04642 i006Riparian32282
7Remotesensing 15 04642 i007Fires car21182
8Remotesensing 15 04642 i008Island interior26233
9Remotesensing 15 04642 i009Acacia woodlands27242
10Remotesensing 15 04642 i010Acacia shrub lands27242
11Remotesensing 15 04642 i011Acacia grasslands22193
12Remotesensing 15 04642 i012short mopane26225
13Remotesensing 15 04642 i013Mixed mopane1190
14Remotesensing 15 04642 i014Exposed soils27243
Total3312917
Table 2. The information on the land-cover category of the IP dataset.
Table 2. The information on the land-cover category of the IP dataset.
No.ColorClassTrainTest
1Remotesensing 15 04642 i015Alfalfa1036
2Remotesensing 15 04642 i016Corn-notill2861142
3Remotesensing 15 04642 i017Corn-mintill166664
4Remotesensing 15 04642 i018Corn48189
5Remotesensing 15 04642 i019Grass-pasture97386
6Remotesensing 15 04642 i020Grass-trees146584
7Remotesensing 15 04642 i021Grass-pasture-mowed622
8Remotesensing 15 04642 i022Hay-windrowed96382
9Remotesensing 15 04642 i023Oats416
10Remotesensing 15 04642 i024Soybean-notill195777
11Remotesensing 15 04642 i025Soybean-mintill4911964
12Remotesensing 15 04642 i026Soybean-clean119474
13Remotesensing 15 04642 i027Wheat41164
14Remotesensing 15 04642 i028Woods2531012
15Remotesensing 15 04642 i029Buildings-Grass-Tree78308
16Remotesensing 15 04642 i030Stone-Steel-Towers1974
Total20558194
Table 3. The information on the land-cover category of the Houston 2013 dataset.
Table 3. The information on the land-cover category of the Houston 2013 dataset.
No.ColorClassTrainTest
1Remotesensing 15 04642 i031Healthy grass2391125
2Remotesensing 15 04642 i032Stressed grass1261128
3Remotesensing 15 04642 i033Synthetic grass70627
4Remotesensing 15 04642 i034Trees1251119
5Remotesensing 15 04642 i035Soil1251117
6Remotesensing 15 04642 i036Water33292
7Remotesensing 15 04642 i037Residential1271141
8Remotesensing 15 04642 i038Commercial1251119
9Remotesensing 15 04642 i039Road1261126
10Remotesensing 15 04642 i040Highway1231104
11Remotesensing 15 04642 i041Railway1241111
12Remotesensing 15 04642 i042Parking Lot 11241109
13Remotesensing 15 04642 i043Parking Lot 247422
14Remotesensing 15 04642 i044Tennis Court43385
15Remotesensing 15 04642 i045Running Track66594
Total151013519
Table 4. Results of classification for the BOW dataset.
Table 4. Results of classification for the BOW dataset.
No.SVMRFKNNGaussianNBHybridSNMSRN_A3D_2D_CNNRSSANMSRN_BDMCNMSDANDSMSFNet
1100.0097.8999.5998.3787.8296.0592.75100.0098.7891.0195.2997.59
298.1198.8192.1367.74100.00100.0096.77100.00100.0088.24100.00100.00
378.6590.2593.6280.58100.00100.00100.00100.00100.00100.00100.00100.00
4100.0083.6487.2565.0298.47100.0099.47100.00100.0096.4896.41100.00
580.5972.6682.3371.9088.2497.0595.9087.0892.37100.0096.54100.00
650.0076.3460.0057.2397.78100.0097.5193.53100.00100.0097.10100.00
7100.0098.6799.5597.00100.00100.00100.00100.00100.0099.57100.00100.00
884.9088.0277.5382.8499.44100.00100.00100.0091.9294.49100.00100.00
968.4880.1478.2371.4398.26100.00100.0097.45100.00100.00100.00100.00
1075.6276.8388.0267.8398.67100.0098.6798.2296.9697.80100.00100.00
1186.2489.5391.4988.8597.5196.4899.6499.27100.0099.63100.00100.00
1289.6091.5793.4991.6197.59100.00100.0097.4446.5596.41100.00100.00
1390.7779.7693.0670.97100.00100.00100.0094.88100.0098.7797.97100.00
14100.0098.8097.5993.62100.0097.70100.0095.3183.3397.18100.00100.00
OA (%)82.0585.9887.0478.8396.9599.0198.5397.1591.5097.5798.6699.79
AA (%)81.8286.9587.8781.0695.8799.1298.1396.1692.2197.0298.7199.82
Kappa × 10080.5384.8185.9677.1096.6998.9298.4096.9290.8197.3698.5599.78
Complexity (G)0.01020.00110.00050.00020.00030.00450.00250.0003
Parameter (M)9.22520.19650.25790.11590.16372.22921.26380.1628
The black bold highlights which mechanic works best.
Table 5. Results of classification for the IP dataset.
Table 5. Results of classification for the IP dataset.
No.SVMRFKNNGaussianNBHybridSNMSRN_A3D_2D_CNNRSSANMSRN_BDMCNMSDANDSMSFNet
10.0086.6736.3631.0797.06100.00100.0097.3090.32100.00100.0094.59
261.5182.0250.3845.5498.8699.7395.7998.0097.4597.4698.9598.70
384.0478.6661.9535.9297.04100.0095.9999.5498.7493.5099.54100.00
446.4372.8753.2615.3198.8698.3892.9499.4699.3996.8198.8597.42
588.8290.1684.713.5798.4797.7299.4798.2292.5498.6998.7099.48
676.7282.6178.0867.87100.00100.00100.0099.8399.65100.0099.49100.00
70.0083.3368.42100.00100.00100.00100.00100.00100.00100.0086.96100.00
883.4987.1688.5583.7896.46100.00100.0099.4880.0898.7099.74100.00
90.00100.0040.0011.0276.19100.00100.00100.000.00100.00100.00100.00
1070.8983.6169.4027.0799.7497.7297.4899.4888.9399.8798.46100.00
1158.5175.1669.4960.6098.7798.9497.4899.1997.5799.6999.74100.00
1259.3866.7462.1323.9598.3492.4091.1998.1391.5292.7491.3098.95
1382.2392.5386.7084.38100.0089.6299.3899.3994.5896.9197.02100.00
1487.3989.7891.7675.0899.9099.5197.4799.80100.0099.4099.90100.00
1586.3072.0064.12753.1794.1298.0990.8898.72100.0092.9295.0099.68
1698.36100.00100.0098.4498.67100.00100.0097.3394.3791.1498.5398.67
OA (%)70.2189.9170.9550.8898.5898.5096.8599.0795.5697.8698.6199.62
AA (%)53.0666.7762.3952.6596.8796.5694.3496.5386.2593.4794.8599.26
Kappa × 10065.0778.0166.6344.0798.3998.2996.4198.9494.9497.5798.4199.57
Complexity (G)0.01020.00110.00050.00020.00030.00450.00250.0003
Parameter (M)9.22580.19810.25820.11640.16422.22951.26400.1327
The black bold highlights which mechanic works best.
Table 6. Results of classification for the Houston 2013 dataset.
Table 6. Results of classification for the Houston 2013 dataset.
No.SVMRFKNNGaussianNBHybridSNMSRN_A3D_2D_CNNRSSANMSRN_BDMCNMSDANDSMSFNet
182.3895.6498.2990.7897.6498.8598.1698.7599.1199.0199.20100.00
298.4695.4495.7098.8099.7399.6599.1997.9899.7398.1799.5699.82
397.72100.0097.2993.0999.68100.0099.8499.52100.0098.12100.00100.00
498.7699.5598.1199.0193.2599.9199.8299.2999.4698.8999.28100.00
586.8693.3693.0073.9699.91100.00100.0099.73100.0099.6499.3799.73
6100.00100.00100.0031.00100.00100.00100.0098.29100.00100.00100.00100.00
764.9179.1587.8363.0697.77100.0096.3296.9698.7999.4395.6699.91
886.0387.9582.0570.0398.32100.0096.6193.64100.0097.5290.71100.00
961.3875.5976.0742.6793.8299.1195.9689.7595.5892.7091.2299.82
1051.3684.4379.240.0098.5796.7697.6894.3598.2295.76100.0099.91
1145.1676.5079.7634.4297.9999.7398.9296.75100.0093.2896.3999.46
1260.8272.1670.6721.0898.9299.9198.8490.8499.7391.7998.83100.00
13100.0079.7288.8915.61100.0097.3299.2095.5599.0196.5098.41100.00
1479.3996.6895.1767.4098.97100.0099.74100.00100.0099.7499.23100.00
1599.6699.6499.1399.0899.8399.00100.00100.0099.83100.0099.8399.83
OA (%)75.1787.4787.5160.8297.9199.3699.4196.3199.1796.8297.3399.88
AA (%)74.9186.0985.7763.1097.3399.0798.0396.3798.8395.8897.2199.83
Kappa × 10073.1186.4386.4757.7397.7499.3198.2896.0199.1096.5697.1199.87
Complexity (G)0.01020.00110.00050.00020.00030.00450.00250.0003
Parameter (M)5.12200.19730.25800.11620.16392.22931.26390.1557
The black bold highlights which mechanic works best.
Table 7. The effect of the constructed CConv layer.
Table 7. The effect of the constructed CConv layer.
DatasetsCaseLocation of DDBMetrics
DDB1DDB2DDB3DDB4DDB5Parameters (M)Time (s)OA (%)
BOWcase1×××××0.5228500.8499.59
case2××××0.4510930.8298.97
case3×××0.3893810.7799.42
case4××0.3076690.7799.42
case5×0.2359570.7699.49
case60.1627850.7299.79
IPcase1×××××0.3196091.2299.54
case2××××0.2793611.1998.89
case3×××0.2404931.1599.67
case4××0.2001491.1298.77
case5×0.1598051.0495.86
case60.1183691.0299.62
Houston 2013case1×××××0.3569892.2699.68
case2××××0.3166502.2099.64
case3×××0.2764022.1099.55
case4××0.2361541.9999.56
case5×0.1959061.8899.62
case60.1556581.7899.88
The black bold highlights which mechanic works best.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, D.; Li, Q.; Li, M.; Zhang, J. A Decompressed Spectral-Spatial Multiscale Semantic Feature Network for Hyperspectral Image Classification. Remote Sens. 2023, 15, 4642. https://doi.org/10.3390/rs15184642

AMA Style

Liu D, Li Q, Li M, Zhang J. A Decompressed Spectral-Spatial Multiscale Semantic Feature Network for Hyperspectral Image Classification. Remote Sensing. 2023; 15(18):4642. https://doi.org/10.3390/rs15184642

Chicago/Turabian Style

Liu, Dongxu, Qingqing Li, Meihui Li, and Jianlin Zhang. 2023. "A Decompressed Spectral-Spatial Multiscale Semantic Feature Network for Hyperspectral Image Classification" Remote Sensing 15, no. 18: 4642. https://doi.org/10.3390/rs15184642

APA Style

Liu, D., Li, Q., Li, M., & Zhang, J. (2023). A Decompressed Spectral-Spatial Multiscale Semantic Feature Network for Hyperspectral Image Classification. Remote Sensing, 15(18), 4642. https://doi.org/10.3390/rs15184642

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop