Next Article in Journal
Urban Building Extraction and Modeling Using GF-7 DLC and MUX Images
Previous Article in Journal
Two-Stream Convolutional Long- and Short-Term Memory Model Using Perceptual Loss for Sequence-to-Sequence Arctic Sea Ice Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Stage Convolutional Broad Learning with Block Diagonal Constraint for Hyperspectral Image Classification

1
Engineering Research Center of Intelligent Control for Underground Space, China University of Mining and Technology, Ministry of Education, Xuzhou 221116, China
2
School of Information and Control Engineering, China University of Mining and Technology, Xuzhou 221116, China
3
School of Computer Science and Engineering, South China University of Technology, Guangzhou 510006, China
4
Department of Computer and Information Science, Faculty of Science and Technology, University of Macau, Macau 999078, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(17), 3412; https://doi.org/10.3390/rs13173412
Submission received: 29 July 2021 / Revised: 17 August 2021 / Accepted: 23 August 2021 / Published: 27 August 2021

Abstract

:
By combining the broad learning and a convolutional neural network (CNN), a block-diagonal constrained multi-stage convolutional broad learning (MSCBL-BD) method is proposed for hyperspectral image (HSI) classification. Firstly, as the linear sparse feature extracted by the conventional broad learning method cannot fully characterize the complex spatial-spectral features of HSIs, we replace the linear sparse features in the mapped feature (MF) with the features extracted by the CNN to achieve more complex nonlinear mapping. Then, in the multi-layer mapping process of the CNN, information loss occurs to a certain degree. To this end, the multi-stage convolutional features (MSCFs) extracted by the CNN are expanded to obtain the multi-stage broad features (MSBFs). MSCFs and MSBFs are further spliced to obtain multi-stage convolutional broad features (MSCBFs). Additionally, in order to enhance the mutual independence between MSCBFs, a block diagonal constraint is introduced, and MSCBFs are mapped by a block diagonal matrix, so that each feature is represented linearly only by features of the same stage. Finally, the output layer weights of MSCBL-BD and the desired block-diagonal matrix are solved by the alternating direction method of multipliers. Experimental results on three popular HSI datasets demonstrate the superiority of MSCBL-BD.

1. Introduction

With the rapid development of remote sensing technology, the spectral and spatial resolutions of hyperspectral images (HSIs) are increasing. HSIs can be used to identify subtle differences among ground objects, demonstrating strong discriminative ability [1,2], which has been applied in many fields such as crop monitoring [3], environmental analysis and prediction [4], climate detection [5], and forest surveys [6]. Pixel-wise classification is one of the common tasks in these applications. The commonly used classification methods include support vector machine (SVM) [7,8], and k-nearest neighbor [9]. However, due to the large number of bands and the strong correlation between adjacent bands, band redundancy exists in HSI data [10,11]. Direct classification with the original HSI will result in lower classification accuracy. Therefore, before HSI classification, performing feature extraction [12] (mapping the original data to another feature space through the mapping matrix) or feature selection (directly selecting several bands from the original band according to certain criteria or strategies) [13] can effectively improve the classification accuracy. Many works have focused on feature learning, such as manifold learning-based feature extraction [14,15] and some extended versions [16,17], metric learning-based dimensionality reduction [18,19], filtering-based feature learning [20,21] and multi-feature-based feature fusion [22,23].
The recent proposed broad learning system (BLS) [24,25] can be viewed as a three-layer forward neural network consisting of an input layer, an intermediate layer, and an output layer. The intermediate layer includes two parts: the mapped feature (MF) part, which is obtained by mapping the input with randomly generated or sparse-autoencoder-optimized weights; the enhancement node (EN) part achieves width expansion by mapping MF with randomly generated weights. The BLS has strong function approximation ability with a very flexible structure [26]. For example, Liu et al. [24] extended it to radial basis function network, which was effective and efficient in classification. By combining BLS and the Takagi–Sugeno fuzzy subsystem, a fuzzy BLS method was proposed [27], which can achieve better performance in terms of regression and classification tasks than the commonly used neuro-fuzzy and non-fuzzy methods. Chen et al. [26] provided a mathematical proof of the general approximation of BLS and various structural changes of BLS. Kong et al. [28] proposed a semi-supervised BLS by using the class probability framework and further applied it to the HSI classification task.
In recent years, deep learning [29] has been widely used in HSI classification because of its strong nonlinear mapping ability and end-to-end working mode, which can automatically learn the features with strong robustness [30,31] and the underlying regularities of samples [32]. For example, Chen et al. [33] used a multilayer autoencoder for HSI classification for the first time. Subsequently, Chen et al. [34] proposed an HSI classification method based on the spectral-spatial deep belief network (SS-DBN). Based on the autoencoder and deep belief network (DBN), many improved methods have been proposed, such as spatial updated deep autoencoder [35], diversified DBN model [36], and group belief network [37]. Boththe deep autoencoder and DBN belong to the fully connected neural networks, which have a large amount of parameters, and the input data generally require a 1D vector. However, the original HSI data are in the form of a 3D tensor. Vectorizing the original 3D HSI data into a 1D form not only destroys the inherent structure of HSI data but also leads to an increase in dimensionality. As an important deep learning model, the convolutional neural network (CNN) can greatly reduce the number of parameters and reduce the difficulty of network training by utilizing the local connection and weight-sharing mechanism. To this end, many CNN-based HSI classification methods have been proposed. Romero et al. [38] proposed an unsupervised layer-wise pre-trained CNN to extract the multi-layer sparse features from HSIs. Makantasis et al. [39] first applied kernel principal component analysis to the original HSI data to obtain its low-dimension tensor representation, and then used it as the input of CNN to realize spectral-spatial feature extraction and the classification of HSIs. Chen et al. [40] extracted the spectral, spatial, and spectral-spatial features of HSIs using 1D, 2D, and 3D CNNs, respectively. Zhao and Du [41] proposed the spectral-spatial feature-based classification method to extract the spectral and spatial features from HSIs by using balanced local discriminant embedding and CNN, respectively.
The training of a CNN requires a large number of labeled samples. However, the acquisition of labeled HSI samples is expensive, and only a subset with a quite small sample size can be used for training. Therefore, training CNNs with a small number of labeled HSI samples is the primary problem to be solved. To this end, numerous works have been conducted. For example, Ghamisi et al. [42] proposed a self-improving CNN by combining fractional-order Darwinian particle swarm optimization and CNN to address the problems of dimensionality and limited labeled samples. Santara et al. [43] designed an end-to-end CNN called band-adaptive spectral-spatial feature learning neural network (BASS-Net), which consists of three parts named band selection and segmentation, parallel network, and classification. Compared with the conventional CNN, BASS-Net combines the feature selection and parallel connection of the network, which greatly reduces the amount of parameters of the network. Therefore, BASS-Net can achieve better performance than CNN in situations with limited labeled samples. Chan et al. proposed a simplified CNN structure named principal component analysis network (PCANet) [44], which has attracted attention due to its simple training process and not needing a large number of iterative processes. Furthermore, Pan et al. [45] used kernel PCA to replace PCA for convolutional kernel learning to solve the problem of the insufficient nonlinear mapping ability in PCANet. By combining the rolling guidance filter and vertex component analysis network, Pan et al. [46] proposed a simplified deep-learning model for HSI classification. In addition to the above work, sample expansion is another commonly used solution to address the issue of insufficient labeled samples. For example, Li et al. [47] proposed a CNN with pixel-pair features (CNN-PPF), constructing a new training set with many more labeled samples than the original training set by comparing the labels between every two samples.
In summary, CNN has strong feature representation ability, but when labeled samples are limited, this ability is limited to some extent. The structure of BLS is simple and flexible, but the used linear mapped features cannot fully express HSI data. To this end, we constructed a novel multi-stage convolutional broad learning with a block-diagonal constraint (MSCBL-BD) method for HSI classification by taking full advantage of CNN and BLS. The main contributions of our work are summarized as follows: (1) By combing CNN and BLS, their advantages can be simultaneously utilized. When labeled samples are limited, the training set cannot characterize the complete distribution of HSIs. Although features extracted by CNN can provide strong discriminative ability, they may overfit to the training set. The concatenation of convolutional features and broad features can be seen as the combination of fine- and coarse-designed features, which have stronger generalization ability than either of them alone. (2) After the multi-layer mapping of the CNN, some information of the original HSI is inevitably lost. Therefore, multi-stage convolutional features are utilized and expanded stage-by-stage to mitigate the information loss to a certain extent. (3) Due to the use of width expansion and multi-stage features, the similarity between the features of different stages may be improved accordingly. This results in the redundancy of features. Therefore, we use a block-diagonal matrix to impose constraints on the multi-stage convolutional broad features to enhance the independence between the convolutional broad features of different stages, which is helpful to seek diversity in features and learn a more accurate HSI classification model.
The rest of this paper is organized as follows: We elaborate the proposed MSCBL-BD for HSI classification in Section 2. Experiments on three popular hyperspectral datasets are described in Section 3, followed by the discussion of the proposed method in Section 4. The conclusions are provided in Section 5.

2. MSCBL-BD for HSI Classification

2.1. Structure of MSCBL-BD

The structure of MSCBL-BD is shown in Figure 1, which mainly includes the following parts: (1) obtaining the low-dimensional neighboring region representation, and using PCA to perform band reduction on the original HSI and constructing the low-dimensional neighboring region representation; (2) extracting the multi-stage CBFs. Firstly, use the limited labeled samples to pre-train a three-stage CNN and extract the multi-stage convolutional features (CFs) of the HSI. Secondly, perform channel-wise global average pooling on the features of the first two stages. Thirdly, perform stage-wise width expansion on the CFs to obtain the broad features (BFs), then further combine the CFs and BFs to obtain the convolutional broad features (CBFs). (3) Imposing the block-diagonal constraints on the multi-stage CBFs. Map the CBFs of all three stages through a block-diagonal matrix to obtain the block-diagonal-constrained CBFs, which ensures that each CBF is only linearly represented by features of the same stage. The optimal solutions of the output layer weight and block-diagonal matrix can be found by the alternating direction method of multipliers (ADMM).

2.2. CNN Pre-Training

The original hyperspectral image is presented in the shape of a 3D cube. If vectorization is performed directly on the HSI, it not only leads to an increase in dimensions, but also destroys the inherent structure of HSI data. The neighboring region representation is a common spectral-spatial representation method in HSIs [48], which is constructed by selecting several pixels around the target pixel. Figure 2 shows a schematic diagram of selecting the surrounding 24 pixels to form a 5 × 5 × B -size neighboring region representation, where B denotes the number of bands in the original HSI. According to this representation, not only the band information of the target pixel but also the information of neighboring pixels can be obtained. Furthermore, if the original high-dimensional neighboring region representation is directly used as the input of the CNN, redundant information will exists in the band and the number of network parameters will increase dramatically, thereby affecting the final performance of the CNN. A common method is utilizing a dimensionality reduction technique, such as principal component analysis (PCA), to reduce the number of bands in the neighboring region representation of an HSI. Define a low-dimensional neighboring region representation χ R d 1 × d 2 × d 3 as the input data of the CNN, where d 1 , d 2 , and d 3 are the width, height, and number of bands, respectively.
Due to the large amount of parameters of the fully connected layer, the CNN used in our work only includes the convolutional, pooling, nonlinea, and SoftMax layers. The input is connected to the convolutional layer by the convolutional kernel to obtain the output feature maps. The calculation formula is:
F C = I C K C + b C
where b C is the bias, F C represents the output features of the convolution layer, and I C is the input of the convolutional layer. For the first layer, the input data is the neighboring region representation of the HSI, which is denoted as I 1 C = χ . K C indicates the convolutional kernel of the convolutional layer and * represents the convolutional operation. In general, a pooling layer is added after a convolutional layer, which aims to quickly reduce the dimensions and enhance the invariance of the extracted features. The features F P obtained by the pooling layer are:
F P = down ( I P )
where down ( · ) indicates a max-pooling operation and I P is the input data of the pooling layer, which is also the output of the previous convolutional layer. To achieve nonlinear mapping, a convolutional layer or a pooling layer is typically connected to a nonlinear layer. Here, the activation function of the nonlinear layer is the sigmoid function:
F N = 1 / 1 + exp ( I N )
where I N is the input data of the nonlinear layer, which is also the output of the previous pooling layer; F N denotes the output nonlinear features of the nonlinear layer. The SoftMax function is generally used as the last layer of the CNN. The number of neurons in the SoftMax layer is equal to the number of classes. The SoftMax loss function is defined as [49]:
J ( W S , b S ) = 1 N i = 1 N j = 1 C 1 { y i = j } log e W j S I i S + b j S l = 1 C e W l S I i S + b l S
where N denotes the number of training samples; C is the number of classes; y i is the label of the ith sample; W S and b S are the weight and bias of the SoftMax layer, respectively; I S is the input data of the SoftMax layer, which is obtained by the previous multiple convolutional, pooling, and nonlinear calculation procedures; and 1 { · } is the indicator function, so that 1 { a t r u e s t a t e m e n t } = 1 , a n d 1 { a f a l s e s t a t e m e n t } = 0 . The CNN training process consists of two parts: forward and backward calculations. In the forward calculation process, the output of each layer is calculated according to the current parameters. In the backward calculation process, the weight and bias of each layer are updated by minimizing the loss function. The batch stochastic gradient descent algorithm is used for weight and bias update.
Here, one convolutional layer plus one pooling layer plus two nonlinear layers plus one 1 × 1 convolutional layer is defined as one stage, and the pre-trained CNN can obtain the multi-stage convolutional feature:
F i stage = f χ | θ i
where F i stage represents the ith stage feature, i = 1 , , s ; and s is the total number of stages, θ i = { K 1 C , b 1 C , , K i C ; b i C } denotes the learnable parameter set from the input to the ith stage of the CNN, including convolutional kernels and biases; and f · denotes the nonlinear mapping procedure of th CNN under θ i .

2.3. MSCBL-BD

The conventional BLS can be regarded as a three-layer neural network, including an input layer, an intermediate layer (composed of an MF and an EN), and an output layer. The MF is obtained by mapping the input with the weights fine-tuned by the linear autoencoder. It is worth noting that the linear features obtained by linear mapping cannot fully express the complex spectral-spatial features of an HSI, thus affecting the final classification performance. The EN is obtained by mapping the MF with random weights to achieve the width expansion of the MF. The output layer is connected to both the MF and the EN. By minimizing the error between the output vector and the ground truth label vector, the objective function of BLS is constructed by:
min W B m Z B | H B W B m Y 2 2 + λ 1 W B m 2 2
where the first term is the empirical risk, which aims to calculate the error between the model output vector and the real label vector; Y denote the labels of the samples; · 2 is the l 2 -norm; the second term is the structural risk, which is used to improve the generalization ability of the model; Z B and H B , respectively, represent the features of the MF and EN; W B m indicates the connection weights of the output layer; and λ 1 is the coefficient of structural risk term. Equation (6) can be solved by ridge regression theory.
Since the linear sparse feature extracted by BLS cannot fully characterize the complex spectral-spatial features of an HSI, the CFs are used to replace the linear sparse features in the MF to achieve more complex nonlinear mapping. Furthermore, in order to reduce the information lost by the CNN in the multi-layer mapping process, the multi-stage features extracted by the CNN are utilized here and separately expanded in width. Given the multi-stage features F 1 stage | | F s 1 stage | F s stage , the channel-wise global average pooling is first performed on the CFs of the first two stages:
P stage = dow n G F stage
where dow n G ( · ) represents the channel-wise global average pooling, which can summarize the global information of the feature map in each channel, to some extent. After the pooling operation, several 1D vectors P stage = P 1 stage | | P s 1 stage are obtained and, together with the CFs of the last stage, constitute the multi-stage CFs (MSCFs) F stage = P 1 stage | | P s 1 stage | F s stage . Furthermore, we expand the MSCFs with random weights to obtain the multi-stage BFs (MSBFs) H i stage :
H i stage = ϕ ( P i stage W i E + b i E )
where ϕ ( · ) is the nonlinear function, such as the tansig. Splice the MSCFs and the MSBFs to obtain the MSCBFs A and rewrite them as:
A = P 1 stage | H 1 stage | | P s stage | H s stage = A 1 stage | | A s stage
Enhancing the linear independence among features can help to produce a more accurate classification model [25,26], so we introduce the block-diagonal constraint [50]. In the commonly used block-diagonal representation method, each sample is represented only by samples with the same class, which can enhance the mutual independence among different classes of samples [51,52]. Here, we use a block matrix D = d i a g d 11 , , d s s to map the MSCBFs into a subspace in which each feature is only represented by those of the same stage. MSCBL-BD optimizes the following objective function:
min W m , D A D W m Y 2 2 + λ 1 W m 2 2 s . t . A = A D , D = d i a g d 11 , , d s s
where W m denotes the output layer weights of MSCBL-BD. Further consider the error term E and rewrite Equation (10) as:
min W m , D A D W m Y 2 2 + λ 1 W m 2 2 s . t . A = A D + E , D = d i a g d 11 , , d s s
Since the absolute block diagonal structure is difficult to learn, the work of [50] was consulted. Assuming that the components on the non-block diagonal are as small as possible, the incoherent extra stage is boosted, and the coherent intra-stage representation is further enhanced at the same time [50]. Two terms are constructed to achieve the above objectives: (1) P D F 2 is used to minimize the elements on the non-block diagonal, P = 1 D 1 D T Y ˜ , Y ˜ = 1 d 1 1 d 1 T 0 0 1 d s 1 d s T , where ⊙ is the Hadamard product, · F denotes the Frobenius norm, 1 D is a D-dimensional vector whose elements are all 1; (2) construct a sparse term Q D 0 to enhance the coherent intra-stage representation, where · 0 represents the l 0 -norm, which can calculate the number of non-zero elements in a matrix, Q i j = x i x j 2 2 . Here, sparsity here is used to calculate the number of elements that are equal to 0. By minimizing the sparse term, we can make as many elements on the non-diagonal as possible trend to 0. Since the optimization of l 0 -norm is NP-hard, a relaxed term Q D 1 is used here, where · 1 denotes the l 1 -norm. Then, Equation (11) can be rewritten as:
min W m , D , E 1 2 A D W m Y 2 2 + λ 1 2 W m 2 2 + λ 2 2 P D F 2 + λ 3 Q D 1 + λ 4 E 2 , 1 + λ 5 D s . t . A = A D + E
where λ 2 λ 5 are the balancing coefficients, D is used to explore the potential correlation patterns [53], and · denotes the nuclear norm. Due to the difficulty in solving the l 1 -norm and the nuclear norm problem, auxiliary variables M and N are introduced, and Equation (12) can be rewritten as:
min W m , D , E , N , M 1 2 A D W m Y 2 2 + λ 1 2 W m 2 2 + λ 2 2 P D F 2 + λ 3 Q M 1 + λ 4 E 2 , 1 + λ 5 N s . t . A = A D + E , M = D , N = D
Equation (13) can be solved by ADMM [50], and the augmented Lagrangian expression is:
L W m , D , E , N , M , C 1 , C 2 , C 3 = 1 2 A D W m Y F 2 + λ 1 2 W m F 2 + λ 2 2 P D F 2 + λ 3 Q M 1 + λ 4 E 2 , 1 + λ 5 N + C 1 , A A D E + C 2 , M D + C 3 , N D + μ 2 A A D E F 2 + M D F 2 + N D F 2
where C 1 , C 2 , and C 3 are Lagrangian multipliers, C 1 , A A D E = t r C 1 T A A D E , and μ > 0 is the penalty parameter. Each variable is updated alternately to find the desired solution and the detailed calculation process of each variable is provided below.
(1) Update W m . Fixing the other variables, the update process for W m is equivalent to solving the following objective function:
L = min W m 1 2 A D W m Y F 2 + λ 1 2 W m F 2
Calculate the derivative of Equation (15) with respect to W m and make it zero, then we can obtain the closed solutions:
W m , ( t + 1 ) = D T , ( t ) A T , ( t ) A D t + λ 1 I 1 D T , ( t ) A T Y
(2) Update D . When the remaining variables are fixed, the expression of Equation (14) about D is:
L = min D 1 2 A D W m , ( t + 1 ) Y F 2 + λ 2 2 P t D F 2 + C 1 t , A A D E t + C 2 t , M t D + C 3 t , N t D + μ t 2 A A D E t F 2 + M t D F 2 + N t D F 2 = 1 2 A D W m , ( t + 1 ) Y F 2 + λ 2 2 P t D F 2 + μ t 2 A A D E t + C 1 t μ t F 2 + M t D + C 2 t μ t F 2 + N t D + C 3 t μ t F 2 = 1 2 A D W m , ( t + 1 ) Y F 2 + λ 2 2 D R F 2 + μ t 2 A A D E t + C 1 t μ t F 2 + M t D + C 2 t μ t F 2 + N t D + C 3 t μ t F 2
where R = Y ˜ D t . Calculate the derivative of Equation (17) with respect to D and make it zero; thus, the optimal solution of D is:
D ( t + 1 ) = 1 μ t A T A W m , t + 1 W m , t + 1 , T + λ 2 μ t + 2 I + A T A 1 1 μ t A T Y W m , t + 1 , T + λ 2 μ t R + A T S 1 + S 2 + S 3
where S 1 = A E + C 1 ( t ) / μ ( t ) , S 2 = N ( t + 1 ) + C 2 t / μ ( t ) , and S 3 = M ( t ) + C 3 ( t ) / μ ( t ) .
(3) Update N . Fix the other variables and abbreviate Equation (14) as an expression only for variable N :
N t + 1 = min N λ 5 N + C 3 t , N D + μ t 2 N D t + 1 F 2 = λ 5 N + μ t 2 N D t + 1 C 3 t μ t F 2
According to [54], the solution can be obtained by using the singular value threshold operation:
N t + 1 = U h λ 5 / μ t Σ V T
where U Σ V T is the singular value decomposition of D t + 1 C 3 t μ t , and h λ 5 / μ t · is the soft threshold operation, which is defined as:
h λ x = x λ , if x > λ x + λ , if x < λ 0 , otherwise .
where λ is a threshold value.
(4) Update M . The update process of variable M is equivalent to solving the following problem:
L = min M λ 3 Q M 1 + C 2 t , M D t + 1 + μ t 2 M D t + 1 F 2 = λ 3 Q M 1 + μ t 2 M D t + 1 C 2 t μ t F 2
which can be updated by the point multiplication mechanism, and the optimal solution is calculated as follows:
M i , j t + 1 = arg min M i , j λ 3 Q i , j M i , j + μ t 2 Q i , j K i , j 2 = h λ 3 Q i , j / μ t K i , j
where K i , j = D i , j t + 1 C 3 t i , j / u t .
(5) Update E . When fixing the other variables, we rewrite the objective function of Equation (14) on variable E and obtain:
L = min E λ 4 E 2 , 1 + C 1 t , A A D t + 1 E + μ t 2 A A D t + 1 E F 2 = λ 4 E 2 , 1 + μ t 2 E A A D t + 1 + C 1 t μ t F 2
According to [50], let G = A A D t + 1 + C 1 t μ t and
E t + 1 i , : = G i 2 λ 4 u t G i 2 G i , if G i 2 > λ 4 u t 0 , if G i 2 λ 4 u t
The above steps are alternated until a predetermined maximum number of iterations is reached, thereby obtaining the desired output layer weight W m and the block diagonal matrix D , and further calculating the prediction label vector Y :
Y = A D W m
Consequently, sparse signal recovery problems can be solved by a dozen different methods, such as orthogonal matching pursuit, K-SVD, ADMM, etc. Among them, ADMM was designed for general decomposition methods and decentralized algorithms in optimization problems. Furthermore, many state-of-the-art algorithms for l 1 -norm-involved problems can be derived by ADMM. In addition, the problem in Equation (13) can be seen as a special case of the classical ADMM problem. Therefore, ADMM is used here to solve the problem in Equation (13).

3. Experiments and Analysis

To verify the validity of the proposed method, we selected three popular real HSI datasets: Indian Pines, Pavia University, and Salinas. The ground-truth maps and sample information of the three HSI datasets are shown in Figure 3. The spatial resolutions of the three datasets are 145 × 145 , 610 × 340 , and 512 × 217 , respectively. All experiments were performed on the MATLAB 2016a platform;the used computer was configured as: CPU Intel I7-4790, 16 G memory, and GPU GTX980. In order to eliminate the randomness, all experiments were repeated 10 times to obtain the mean values. A total of 200 labeled samples per class were randomly selected for training and the others were used for testing. Four evaluation indicators were selected to evaluate the experimental results, which were the classification accuracy of each class of ground object, the average classification accuracy (AA), the overall classification accuracy (OA) and the Kappa coefficient.

3.1. Parameter Setting

Except for the number of feature maps of the last convolutional layer (the numbers of feature maps are 9, 9, and 16 for the Indian Pines, Pavia University, and Salinas datasets, respectively), the network structures of the CNNs for three HSI datasets were the same. The detailed structure of the CNN is shown in Table 1. It can be seen that: (1) there were total 5 convolutional layers, i.e., C1∼C5. The convolutional kernel sizes of C2 and C4 were 1 × 1 , which is helpful for obtaining a deeper network structure at the cost of a small increase in network parameters. The convolutional kernel sizes of C1 and C3 were 4 × 4 , the convolutional kernel size of C5 was 2 × 2 . (2) There were two max-pooling layers, i.e., P1 and P2, both step sizes of which were 2. (3) A total of 4 nonlinear layers denoted as N1∼N4 were included. The activation functions used in all of the nonlinear layers were the Sigmoid function.
The CNN training process consists of two steps: forward calculation and back propagation. The former aims to calculate the classification result based on the current network parameters, and the latter updates the network parameters. Here, we used a stochastic gradient descent algorithm to update the CNN with a batch size of 100, a learning rate of 0.1, and an iteration of 1000. The CNN training process is based on the Matconvnet toolbox [55].
The parameter settings of MSCBL-BD on the three datasets are shown in Table 2, where λ 1 λ 5 are tuned to achieve the best performance via fivefold cross-validations from 0.01 , 0.1 , 1 , 5 , 10 .

3.2. Comparative Experiments

In order to verify the validity of the proposed MSCBL-BD, six benchmark methods and two special cases of MSCBL-BD (CBL and MSCBL) were selected for comparative experiments:
(1)
Traditional classification methods: SVM [8], whose optimal super-parameters are selected through the fivefold cross-validation method;
(2)
Deep learning methods: SS-DBN [34], CNN-PPF [47], and CNN [38]. To set the parameters of SS-DBN and CNN-PPF, we referred to the corresponding articles. The configuration of CNN is shown in Table 1.
(3)
Special cases of MSCBL-BD: CBL (the CBFs of the last stage are connected to the output layer and not constrained with the block diagonal matrix); MSCBL (the MSCBFs are connected to the output layer and not constrained with the block-diagonal matrix).
Table 3, Table 4 and Table 5 show the classification performance comparisons of the different methods on the three HSI datasets and the best values are shown in bold.

4. Discussion

From the experimental results, we can see that: (1) On all of three HSI datasets, for three evaluation indexes (AA, OA, and Kappa coefficients), MSCBL-BD, MSCBL, and CBL achieved higher performance than the other methods. Taking the Indian Pines dataset as an example, CBL outperformed BLS by 11.52% in terms of OA. The main reasons are two-fold: the features learned by BLS are linear sparse features and only the spectral information is utilized. Therefore, the features of the HSI cannot be fully represented. In addition, CBL, MSCBL, and MSCBL-BD outperformed CNN by 1.68%, 2.7%, and 3.91% on the OA, respectively, which verifies the superiority of width expansion. (2) On the three HSI datasets, MSCBL outperformed CBL by 1.02%, 0.24%, and 0.45%, respectively, because MSCBL utilizes the MSCFs and performs feature expansion for each stage, which makes the features learned by MSCBL more discriminative. Furthermore, MSCBL-BD outperformed MSCBL by 1.21%, 0.15%, and 0.6% on the three datasets, respectively, because the MSCBFs are mapped through a block-diagonal matrix, so that each obtained feature is linearly represented only by the features of its own stage, thereby enhancing the linear independence of the MSCBFs of each stage. This can help with learning a more accurate HSI classification model, which in turn improves the final classification accuracy. (3) Both SS-DBN and CNN are deep spectral-spatial classification methods. Compared with SS-DBN, CNN yielded higher OAs on all HSI datasets. This is because SS-DBN takes features by stacking the spectral vector and the vectorized neighboring region representation as the input of the DBN, which not only leads to an increase in the dimension of the input data, but also destroys the inherent structure of the data. In addition, in the case of a small number of labeled samples, as a kind of fully-connected neural network, over-fitting may occur in the training process of SS-DBN. Figure 4, Figure 5 and Figure 6 show the classification maps achieved by different methods. It can be seen that the classification maps obtained by MSCBL-BD on all of three HSI datasets are smoother and more detailed. Taking the Indian Pines dataset as an example, the benchmark methods misclassify more Soybean-clean, Soybean-notill, and Corn-notill as Soybean-mintill; misclassify more Grass-trees into Woods; and misclassify more Corn-notill into Corn-minitill.

5. Conclusions

A novel BLS based method (MSCBL-BD) for HSI classification was proposed in this paper. By replacing the linear sparse features with convolutional features, the nonlinear mapping ability of the model is improved and the complex spectral-spatial features of HSI can be better represented. Therefore, the CBL has higher classification accuracy than BLS. Furthermore, in order to reduce the information loss occurring in the multi-layer mapping process of CNN, the MSCBFs are utilized. Moreover, in order to train a more accurate HSI classification model, the block-diagonal constraint is introduced to the MSCBL. The MSCBFs are mapped by a block-diagonal matrix, and the obtained features are only linearly represented by those of their own stages. Therefore, the linear independence of the MSCBFs is enhanced and the classification accuracy is improved ultimately. The experimental results demonstrated the superiority of MSCBL-BD compared to some competitive methods. In our future work, we will increase the robustness of our method, so that satisfactory performance can be achieved when bands are missing.

Author Contributions

All of the authors provided significant contributions to the work. Y.K. and X.W. conceived and designed the experiments; Y.K. and X.W. performed the experiments; Y.K. and Y.C. analyzed the data; Y.K. and Y.C. wrote the original paper; C.L.P.C. reviewed and edited the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grants 62006232, 61976215, 61772532. This research was also funded by the Natural Science Foundation of Jiangsu Province under grant BK20200632.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found: http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes accessed on 15 July 2018.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADMMAlternating direction method of multipliers
BASS-NetBand-adaptive spectral-spatial feature learning neural network
BFBroad feature
BLSBroad learning system
CBFConvolutional broad feature
CFConvolutional feature
CNNConvolutional neural network
CNN-PPFConvolutional neural network with pixel-pair features
DBNDeep belief network
ENEnhancement node
HSIHyperspectral image
MFMapped feature
MSBFMulti-stage broad feature
MSCBFMulti-stage convolutional broad feature
MSCBL-BDBlock-diagonal constrained multi-stage convolutional broad learning
MSCFMulti-stage convolutional feature
PCAPrincipal component analysis
SS-DBNSpectral-spatial deep belief network
SVMSupport vector machine

References

  1. Luo, F.; Du, B.; Zhang, L.; Zhang, L.; Tao, D. Feature learning using spatial-spectral hypergraph discriminant analysis for hyperspectral image. IEEE Trans. Cybern. 2019, 49, 2406–2419. [Google Scholar] [CrossRef] [PubMed]
  2. Zhou, Y.; Wei, Y. Learning hierarchical spectral-spatial features for hyperspectral image classification. IEEE Trans. Cybern. 2016, 46, 1667–1678. [Google Scholar] [CrossRef] [PubMed]
  3. Meroni, M.; Fasbender, D.; Balaghi, R.; Dali, M.; Haffani, M.; Haythem, I.; Hooker, J.; Lahlou, M.; Lopez-Lozano, R.; Mahyou, H.; et al. Evaluating NDVI data continuity between SPOT-VEGETATION and PROBA-V missions for operational yield forecasting in North African countries. IEEE Trans. Geosci. Remote Sens. 2016, 54, 795–804. [Google Scholar] [CrossRef]
  4. Brunet, D.; Sills, D. A generalized distance transform: Theory and applications to weather analysis and forecasting. IEEE Trans. Geosci. Remote Sens. 2017, 55, 1752–1764. [Google Scholar] [CrossRef]
  5. Islam, T.; Hulley, G.C.; Malakar, N.K.; Radocinski, R.G.; Guillevic, P.C.; Hook, S.J. A Physics-based algorithm for the simultaneous retrieval of land surface temperature and emissivity from VIIRS thermal infrared data. IEEE Trans. Geosci. Remote Sens. 2017, 55, 563–576. [Google Scholar] [CrossRef]
  6. Matsuki, T.; Yokoya, N.; Iwasaki, A. Hyperspectral tree species classification of Japanese complex mixed forest with the aid of lidar data. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2177–2187. [Google Scholar] [CrossRef]
  7. Liu, L.; Huang, W.; Wang, C. Hyperspectral image classification with kernel-based least-squares support vector machines in sum space. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2018, 11, 1144–1157. [Google Scholar] [CrossRef]
  8. Tan, K.; Zhang, J.; Du, Q.; Wang, X. GPU parallel implementation of support vector machines for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 4647–4656. [Google Scholar] [CrossRef]
  9. Ma, L.; Crawford, M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  10. Bioucas-Dias, J.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  11. Sun, W.; Halevy, A.; Benedetto, J.; Czaja, W.; Liu, C.; Wu, H.; Shi, B.; Li, W. UL-Isomap based nonlinear dimensionality reduction for hyperspectral imagery classification. ISPRS J. Photogramm. Remote Sens. 2014, 89, 25–36. [Google Scholar] [CrossRef]
  12. Sun, W.; Yang, G.; Du, B.; Zhang, L.; Zhang, L. A sparse and low-rank near-isometric linear embedding method for feature extraction in hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4032–4046. [Google Scholar] [CrossRef]
  13. Zhang, L.; Zhang, Q.; Du, B.; Huang, X.; Tang, Y.Y.; Tao, D. Simultaneous spectral-spatial feature selection and extraction for hyperspectral images. IEEE Trans. Cybern. 2018, 48, 16–28. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Li, W.; Liu, J.; Du, Q. Sparse and low-rank graph for discriminant analysis of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4094–4105. [Google Scholar] [CrossRef]
  15. Huang, H.; Shi, G.; He, H.; Duan, Y.; Luo, F. Dimensionality reduction of hyperspectral imagery based on spatial-spectral manifold learning. IEEE Trans. Cybern. 2020, 50, 2604–2616. [Google Scholar] [CrossRef] [Green Version]
  16. Pan, L.; Li, H.; Li, W.; Chen, X.; Wu, G.; Du, Q. Discriminant analysis of hyperspectral imagery using fast kernel sparse and low-rank graph. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6085–6098. [Google Scholar] [CrossRef]
  17. Pan, L.; Li, H.; Deng, Y.; Zhang, F.; Chen, X.; Du, Q. Hyperspectral dimensionality reduction by tensor sparse and low-rank graph-based discriminant analysis. Remote Sens. 2017, 9, 452. [Google Scholar] [CrossRef] [Green Version]
  18. Dong, Y.; Du, B.; Zhang, L.; Zhang, L. Dimensionality reduction and classification of hyperspectral images using ensemble discriminative local metric learning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2509–2524. [Google Scholar] [CrossRef]
  19. Dong, Y.; Du, B.; Zhang, L.; Zhang, L. Exploring locally adaptive dimensionality reduction for hyperspectral image classification: A maximum margin metric learning aspect. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2017, 10, 1136–1150. [Google Scholar] [CrossRef]
  20. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J. PCA-based edge-preserving features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  21. Jia, S.; Shen, L.; Zhu, J.; Li, Q. A 3-D Gabor phase-based coding and matching framework for hyperspectral imagery classification. IEEE Trans. Cybern. 2018, 48, 1176–1188. [Google Scholar] [CrossRef]
  22. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  23. Li, J.; Huang, X.; Gamba, P.; Bioucas-Dias, J.; Zhang, L.; Benediktsson, J.; Plaza, A. Multiple feature learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1592–1606. [Google Scholar] [CrossRef] [Green Version]
  24. Liu, Z.; Chen, C.L.P. Broad learning system: Structural extensions on single-layer and multi-layer neural networks. In Proceedings of the International Conference on Security, Pattern Analysis, and Cybernetics, Shenzhen, China, 15–18 December 2017; pp. 136–141. [Google Scholar]
  25. Chen, C.L.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 10–24. [Google Scholar] [CrossRef] [PubMed]
  26. Chen, C.L.P.; Liu, Z.; Feng, S. Universal approximation capability of broad learning system and its structural variations. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 1191–1204. [Google Scholar] [CrossRef] [PubMed]
  27. Feng, S.; Chen, C.L.P. Fuzzy broad learning system: A novel neuro-fuzzy model for regression and classification. IEEE Trans. Cybern. 2020, 50, 414–424. [Google Scholar] [CrossRef] [PubMed]
  28. Kong, Y.; Wang, X.; Cheng, Y.; Chen, C.L.P. Hyperspectral imagery classification based on semi-supervised broad learning system. Remote Sens. 2018, 10, 685. [Google Scholar] [CrossRef] [Green Version]
  29. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  31. Zhu, X.; Tuia, D.; Mou, L.; Xia, G.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, T.; Su, G.; Qing, C.; Xu, X.; Cai, B.; Xing, X. Hierarchical lifelong learning by sharing representations and integrating hypothesis. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 1004–1014. [Google Scholar] [CrossRef]
  33. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  34. Chen, Y.; Zhao, X.; Jia, X. Spectral-spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  35. Ma, X.; Wang, H.; Geng, J. Spectral-spatial classification of hyperspectral image based on deep auto-encoder. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  36. Zhong, P.; Gong, Z.; Li, S.; Schönlieb, C.B. Learning to diversify deep belief networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3516–3530. [Google Scholar] [CrossRef]
  37. Zhou, X.; Li, S.; Tang, F.; Qin, K.; Hu, S.; Liu, S. Deep learning with grouped features for spatial spectral classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 97–101. [Google Scholar] [CrossRef]
  38. Romero, A.; Gatta, C.; Camps-Valls, G. Unsupervised deep feature extraction for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1349–1362. [Google Scholar] [CrossRef] [Green Version]
  39. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium, Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  40. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  41. Zhao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  42. Ghamisi, P.; Chen, Y.; Zhu, X. A self-improving convolution neural network for the classification of hyperspectral data. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1537–1541. [Google Scholar] [CrossRef] [Green Version]
  43. Santara, A.; Mani, K.; Hatwar, P.; Singh, A.; Garg, A.; Padia, K.; Mitra, P. BASS net: Band-adaptive spectral-spatial feature learning neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5293–5301. [Google Scholar] [CrossRef] [Green Version]
  44. Chan, T.; Jia, K.; Gao, S.; Lu, J.; Zeng, Z.; Ma, Y. PCANet: A simple deep learning baseline for image classification. IEEE Trans. Image Process. 2015, 24, 5017–5032. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Pan, B.; Shi, Z.; Zhang, N.; Xie, S. Hyperspectral image classification based on nonlinear spectral-spatial network. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1782–1786. [Google Scholar] [CrossRef]
  46. Pan, B.; Shi, Z.; Xu, X. R-VCANet: A new deep-learning-based hyperspectral image classification method. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2017, 10, 1975–1986. [Google Scholar] [CrossRef]
  47. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  48. Gao, Y.; Wang, X.; Cheng, Y.; Wang, Z. Dimensionality reduction for hyperspectral data based on class-aware tensor neighborhood graph and patch alignment. IEEE Trans. Neural Netw. Learn. Sys. 2015, 26, 1582–1593. [Google Scholar]
  49. Jin, J.; Fu, K.; Zhang, C. Traffic sign recognition with hinge loss trained convolutional neural networks. IEEE Trans. Intell. Transp. Syst. 2014, 15, 1991–2000. [Google Scholar] [CrossRef]
  50. Zhang, Z.; Xu, Y.; Shao, L.; Yang, J. Discriminative block-diagonal representation learning for image recognition. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 3111–3125. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Wang, Q.; He, X.; Li, X. Locality and structure regularized low rank representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 911–923. [Google Scholar] [CrossRef] [Green Version]
  52. Wang, J.; Wang, X.; Tian, F.; Liu, C.H.; Yu, H. Constrained low-rank representation for robust subspace clustering. IEEE Trans. Cybern. 2017, 47, 4534–4546. [Google Scholar] [CrossRef] [Green Version]
  53. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Cai, J.-F.; Candès, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  55. Vedaldi, A.; Lenc, K. Matconvnet-convolutional neural networks for matlab. In Proceedings of the ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 689–692. [Google Scholar]
Figure 1. Structure diagram of MSCBL-BD.
Figure 1. Structure diagram of MSCBL-BD.
Remotesensing 13 03412 g001
Figure 2. 5 × 5 × B -size neighboring region representation of an HSI.
Figure 2. 5 × 5 × B -size neighboring region representation of an HSI.
Remotesensing 13 03412 g002
Figure 3. Ground-truth maps and sample information of different hyperspectral datasets: (a) Indian Pines ( 145 × 145 spatial resolution); (b) Pavia University ( 610 × 340 spatial resolution); (c) Salinas ( 512 × 217 spatial resolution). The number of samples contained in each class is shown after the class name.
Figure 3. Ground-truth maps and sample information of different hyperspectral datasets: (a) Indian Pines ( 145 × 145 spatial resolution); (b) Pavia University ( 610 × 340 spatial resolution); (c) Salinas ( 512 × 217 spatial resolution). The number of samples contained in each class is shown after the class name.
Remotesensing 13 03412 g003
Figure 4. Classification maps obtained by different methods on the Indian Pines dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Figure 4. Classification maps obtained by different methods on the Indian Pines dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Remotesensing 13 03412 g004
Figure 5. Classification maps obtained by different methods on the Pavia University dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Figure 5. Classification maps obtained by different methods on the Pavia University dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Remotesensing 13 03412 g005
Figure 6. Classification maps obtained by the different methods on the Salinas dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Figure 6. Classification maps obtained by the different methods on the Salinas dataset: (a) SVM; (b) BLS; (c) SS-DBN; (d) CNN-PPF; (e) CNN; (f) CBL; (g) MSCBL; (h) MSCBL-BD.
Remotesensing 13 03412 g006
Table 1. Configuration of the CNN.
Table 1. Configuration of the CNN.
LayerInputNumber of FiltersWidth of FiltersHeight of FiltersStep SizeOutput
WidthHeightChannelWidthHeightChannelDimension
I1 1717154335
C1171715304411414305880
P1141430 22277301470
N17730////77301470
C277303011177301470
N277303011177301470
C37730304414430480
P34430/2222230120
N32230////2230120
C42230301112230120
N42230////2230120
C522309/16221119/169/16
Softmax119/16////119/169/16
Table 2. Parameter settings of MSCBL-BD.
Table 2. Parameter settings of MSCBL-BD.
Number of
Iterations
Number of
Nodes in EN
λ 1 λ 5
Indian Pines 500 { 0.1 , 10 , 1 , 5 , 1 }
Pavia University110400 { 0.01 , 5 , 0.01 , 10 , 0.1 }
Salinas 500 { 0.1 , 5 , 0.1 , 1 , 1 }
Table 3. Comparison of the classification performance on the Indian Pines dataset.
Table 3. Comparison of the classification performance on the Indian Pines dataset.
SVM [8]BLS [25]SS-DBN [34]CNN-PPF [47]CNN [39]CBLMSCBLMSCBL-BD
A1 (%)75.4982.5477.2092.2389.9892.8494.1496.58
A2 (%)73.7882.2982.7696.6997.7598.4399.0699.32
A3 (%)95.2695.2694.4299.8698.8099.4099.82100
A4 (%)98.4599.1098.3099.4399.2199.6899.8599.96
A5 (%)99.7199.7199.8699.8699.93100100100
A6 (%)78.2186.6887.2395.1094.6996.2297.2099.02
A7 (%)66.1869.6576.0289.4889.5992.2894.1295.96
A8 (%)83.7793.3990.5997.0098.1998.9399.4299.87
A9 (%)98.2798.9194.9399.8999.0299.5499.8199.81
AA (%)85.4689.7389.0496.6296.3597.4898.1698.95
OA (%)79.8084.2684.6194.5194.1095.7896.8098.01
Kappa0.76110.81420.81780.93440.92960.94950.96170.9762
Table 4. Comparison of the classification performance on the Pavia University dataset.
Table 4. Comparison of the classification performance on the Pavia University dataset.
SVM [8]BLS [25]SS-DBN [34]CNN-PPF [47]CNN [39]CBLMSCBLMSCBL-BD
A1 (%)83.3576.3281.4697.4096.6897.2197.7898.25
A2 (%)86.1290.1392.9697.4098.9399.1099.2799.39
A3 (%)83.2382.5988.3792.2997.9198.4098.7898.79
A4 (%)94.6695.4995.5597.6398.5498.6298.8098.80
A5 (%)99.5599.6299.9199.9399.8999.9710099.99
A6 (%)88.8791.2588.9297.8999.6499.9699.9699.92
A7 (%)90.4894.6790.3597.3399.1199.5899.7499.83
A8 (%)84.3984.5485.9393.9798.3198.2798.7398.96
A9 (%)99.9099.6599.4999.5799.6999.8099.8499.85
AA (%)90.0690.4791.4497.0598.7498.9999.2199.31
OA (%)87.0788.2190.2997.0598.5898.8299.0699.21
Kappa0.83010.84450.87080.96260.98090.98420.98730.9893
Table 5. Comparison of the classification performance on the Salinas dataset.
Table 5. Comparison of the classification performance on the Salinas dataset.
SVM [8]BLS [25]SS-DBN [34]CNN-PPF [47]CNN [39]CBLMSCBLMSCBL-BD
A1 (%)99.1499.6098.9499.8699.9799.98100100
A2 (%)99.4599.6598.9599.5999.9599.98100100
A3 (%)99.4699.4679.6099.8399.4699.8599.9399.90
A4 (%)99.5899.4399.5899.6899.5199.9299.9199.93
A5 (%)98.7999.2899.7198.4198.9499.4099.5399.76
A6 (%)99.7999.8199.9199.78100100100100
A7 (%)99.6799.5798.9099.8599.8799.9599.9899.99
A8 (%)84.4083.1887.0483.2885.6191.0592.4194.24
A9 (%)99.3799.7597.9397.4499.3199.4799.5599.63
A10 (%)94.6595.7695.7995.8498.9899.4499.6499.72
A11 (%)98.8798.6698.6499.7599.6099.8899.9499.90
A12 (%)99.9510099.7310099.9099.9599.9499.98
A13 (%)99.5299.1699.5099.38100100100100
A14 (%)97.9198.1299.9599.5299.7999.96100100
A15 (%)69.3573.8285.9281.1895.1295.0295.9097.17
A16 (%)99.0298.9499.7598.5199.8499.8999.9199.97
AA (%)96.1996.5196.2496.9998.4998.9899.1699.39
OA (%)91.6792.1893.7692.9895.9497.2297.6798.27
Kappa0.90670.91250.93020.92150.95460.96890.97400.9807
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kong, Y.; Wang, X.; Cheng, Y.; Chen, C.L.P. Multi-Stage Convolutional Broad Learning with Block Diagonal Constraint for Hyperspectral Image Classification. Remote Sens. 2021, 13, 3412. https://doi.org/10.3390/rs13173412

AMA Style

Kong Y, Wang X, Cheng Y, Chen CLP. Multi-Stage Convolutional Broad Learning with Block Diagonal Constraint for Hyperspectral Image Classification. Remote Sensing. 2021; 13(17):3412. https://doi.org/10.3390/rs13173412

Chicago/Turabian Style

Kong, Yi, Xuesong Wang, Yuhu Cheng, and C. L. Philip Chen. 2021. "Multi-Stage Convolutional Broad Learning with Block Diagonal Constraint for Hyperspectral Image Classification" Remote Sensing 13, no. 17: 3412. https://doi.org/10.3390/rs13173412

APA Style

Kong, Y., Wang, X., Cheng, Y., & Chen, C. L. P. (2021). Multi-Stage Convolutional Broad Learning with Block Diagonal Constraint for Hyperspectral Image Classification. Remote Sensing, 13(17), 3412. https://doi.org/10.3390/rs13173412

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop