Next Article in Journal
Changes in the End-of-Summer Snow Line Altitude of Summer-Accumulation-Type Glaciers in the Eastern Tien Shan Mountains from 1994 to 2016
Next Article in Special Issue
Multiscale Decomposition Prediction of Propagation Loss in Oceanic Tropospheric Ducts
Previous Article in Journal
Retracking Cryosat-2 Data in SARIn and LRM Modes for Plateau Lakes: A Case Study for Tibetan and Dianchi Lakes
Previous Article in Special Issue
Boundary-Assisted Learning for Building Extraction from Optical Remote Sensing Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Continuous Particle Swarm Optimization-Based Deep Learning Architecture Search for Hyperspectral Image Classification

1
School of Automation, China University of Geosciences (Wuhan), Wuhan 430074, China
2
Hubei Key Laboratory of Advanced Control and Intelligent Automation for Complex Systems, China University of Geosciences (Wuhan), Wuhan 430074, China
3
Hubei Key Laboratory of Intelligent Geo-Information Processing, China University of Geosciences (Wuhan), Wuhan 430078, China
4
School of Computer Science, China University of Geosciences (Wuhan), Wuhan 430074, China
5
Beibu Gulf Big Data Resources Utilisation Laboratory, Beibu Gulf University, Qinzhou 535011, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(6), 1082; https://doi.org/10.3390/rs13061082
Submission received: 31 December 2020 / Revised: 20 February 2021 / Accepted: 27 February 2021 / Published: 12 March 2021
(This article belongs to the Special Issue Advanced Artificial Intelligence and Deep Learning for Remote Sensing)

Abstract

:
Deep convolutional neural networks (CNNs) are widely used in hyperspectral image (HSI) classification. However, the most successful CNN architectures are handcrafted, which need professional knowledge and consume a very significant amount of time. To automatically design cell-based CNN architectures for HSI classification, we propose an efficient continuous evolutionary method, named CPSO-Net, which can dramatically accelerate optimal architecture generation by the optimization of weight-sharing parameters. First, a SuperNet with all candidate operations is maintained to share the parameters for all individuals and optimized by collecting the gradients of all individuals in the population. Second, a novel direct encoding strategy is devised to encode architectures into particles, which inherit the parameters from the SuperNet. Then, particle swarm optimization is used to search for the optimal deep architecture from the particle swarm. Furthermore, experiments with limited training samples based on four widely used biased and unbiased hyperspectral datasets showed that our proposed method achieves good performance comparable to the state-of-the-art HSI classification methods.

1. Introduction

A hyperspectral image (HSI) has a high spectral and spatial resolution, which provides rich information about ground truths [1]. Hyperspectral sensors mounted on different space platforms, i.e., imaging spectrometers, are used to capture images in the ultraviolet, visible, near-infrared, and mid-infrared regions of the electromagnetic spectrum, where tens to hundreds of continuous and subdivided spectral bands simultaneously image the target area. The exploitation of HSI signatures is a powerful tool for observing the surface of the Earth. Therefore, there are a wide range of processing techniques to explore HSI data [2,3,4], and HSI classification is one of the most popular techniques for HSI data exploitation. HSI classification is widely applied in many fields, such as agricultural applications [5], forestry and environmental management [6], water and maritime resource management [7], and military and defense applications [8].
Compared to traditional HSI classification algorithms, convolutional neural networks (CNNs) can not only explore spectral information, but can also combine spatial information with spectral information [9]. Moreover, because of the share of parameters to tackle the curse of dimensionality [10] and the local receptive field to learn spatial information [11], CNNs show good performance in HSI band selection [12,13], feature extraction and classification [14]. There are many CNN models used as HSI classifiers, including the spectral CNN [15,16], the spatial CNN [17,18], and the spectral–spatial CNN [19,20]. With the advance of remote sensing technology, better classifiers are being developed to further improve the classification performance and the complexity of CNNs. On the one hand, some new types of CNNs have been introduced for HSI classification, such as the dense CNN using shortcut connections between layers [21] and the deep residual network with skip connections [22]. On the other hand, some other techniques have been combined with CNNs, such as the attention mechanism [23], sparse representation [24], multi-scale fusion [25], and discriminant embedding [26].
Although the aforementioned CNN models have achieved remarkable progress in HSI classification tasks in recent years, there are two major problems that should be considered when applying these CNN models for HSI classification. On the one hand, designing state-of-the-art successful CNN architectures requires substantial professional knowledge of human experts and consume a significant amount of time for repeated debugging. For example, ResNet [27] and DesNet [21] were carefully handcrafted by taking into consideration the manual domain knowledge. On the other hand, manually optimizing a CNN when applying it to different data sources may take days or even weeks and should always be fine-tuned; additionally, this process results in a heavy computational burden, and is not affordable for most researchers and developers [28].
To overcome the problems of handcrafting CNN models, neural architecture search (NAS) [29] is an import method to automatically design neural architectures. Since Zoph et al. [30] successfully discovered a neural network architecture that can achieve comparable performance to handcrafted CNNs, there has been a growing interest in designing a robust and well-performing neural architecture from the predefined search space by NAS. There are many different search strategies that can be used to design a proper CNN architecture, including evolutionary algorithms (EAs), reinforcement learning (RL), and gradient-based methods. Recently, Chen et al. [31] proposed the automatic design of CNNs for HSI classification, based on the gradient descent method, and the automatic CNNs achieved better performance compared to some state-of-the-art deep learning methods. However, the method based on gradient descent easily converges to the local optimal, and it cannot be used when there is a vanishing gradient problem or an exploding gradient problem. EAs search globally, and population-based optimization affords EAs a very good global search ability. To address the NAS problem, a global search technique is required, which has always been a topic of interest for discovering the neural architecture by the EA search strategy. Junior et al. [32] found good chain CNN architectures based on the particle swarm optimization (PSO) search strategy using nine datasets, which can only search simple chain-CNNs. Real et al. [33] proposed AmoebaNet-A and found comparable cell-based neural architectures to some state-of-the-art ImageNet models using the CIFAR-10 dataset, showing that they consumed a significant time without weight-sharing. An alternative architecture optimization method is neuroevolution, which is inspired by the evolution of natural brains. Stanley et al. [34] proposed a classic neuroevolutionary approach, NEAT, Neural Networks through Augmenting Topologieswhich can modify weights and add nodes and connections in neural networks. HyperNEAT [35,36] is an improvement of NEAT, which uses an indirect encoding method named CPPNs [37] that can generate the patterns of weights in neural networks themselves to evolve into bigger network.
To automatically search for CNN architectures for HSI classification, we propose an NAS method named PSO-Net, which is based on PSO and consumes a significant amount of time to search for the optimal architecture. To further accelerate optimal architecture generation, we propose CPSO-Net with weight-sharing parameter optimization based on PSO-Net. We propose two cell-based CNN architecture search methods by PSO, capable of searching globally, compared to the gradient descent method. The main contributions of this paper are summarized as follows:
(1)
Two methods based on PSO are explored to automatically design the CNN architecture for HSI classification.
(2)
A novel encoding strategy is devised that can be used to encode architectures into arrays with the information of the connections and basic operations types between the nodes in computation cells.
(3)
To improve the search efficiency, CPSO-Net maintains continuous SuperNet sharing parameters for all particles and optimizes by collecting the gradients of all individuals in the population.
(4)
PSO-Net and CPSO-Net are tested on four biased and unbiased hyperspectral datasets with limited training samples, showing comparable performance to the state-of-the-art CNN classification methods.
The remainder of this paper is structured as follows. Detailed related work about neural architecture search and PSO is presented in Section 2. Section 3 introduces the proposed PSO-Net and CPSO-Net in detail. Section 4 summarizes and analyzes a series of experiments performed for HSI classification. Section 5 provides the main conclusions and perspectives of this work.

2. Related Work

In this section, we introduce two methods related to the method we proposed, namely neural architecture search and particle swarm optimization.

2.1. Neural Architecture Search

NAS aims to design a well-performing neural network architecture by selecting and combining basic operations and connections from a predefined search space.
There are three main parts in NAS, namely the search space, the search strategy, and the performance estimation strategy. The search space is a set of all possible architectures. The search strategy then automatically finds the optimal architecture from the search space. The estimation of performance aims to find the architectures with high predictive performance by training and validation based on data.
The search space should be predefined with the information of all basic operations and connection nodes. The search space mainly includes two types. The first is to directly construct the global search space of the entire neural network architecture, including the chain architecture search space [38] and the multi-branch architecture search space [30]; the second is to construct parts of the neural network architecture by repeating certain specific structures, such as the cell-based search space [39].
The search strategy contains gradient-based search and RL- and EA-based approaches. Gradient-based approaches [40,41,42] convert the discrete architecture space into a continuous space and find the optimal architecture using the gradient method in the continuous space. RL-based methods [30,39] set the state–action space as the NAS search space and set the policy function as the recurrent network to finish the search process. The validation accuracy of each individual is used as the fitness to evolve the next generation. In EA-based approaches [43,44], the population is composed of a group of candidate network structures, and individuals generate new network architectures through evolutionary algorithms. To improve the search efficiency of EA-based approaches, Yang et al. [45] maintained a SuperNet that is trained by the gradient descent method and shares parameters for all architectures.
The core of the estimation of performance is the weight parameter optimization step, which can be divided into two categories, namely individual and weight-sharing optimization. Individual weight parameter optimization often learns each network separately, typically incurring a heavy computational burden and taking a significant amount of time. Therefore, weight-sharing parameter optimization methods are used to accelerate training [43,46], which can be divided into two categories according to the search strategies, i.e., weight-sharing heuristic search strategies and weight-sharing differentiable search strategies. Weight-sharing heuristic search strategies [47,48] always train one SuperNet that includes all candidate operations first, and then focuses on each individual inherit parameter from the SuperNet. Rather than training all individuals, the SuperNet is optimized only once during one iteration, which dramatically reduces the computational complexity by separately learning each network. Weight-sharing differentiable search strategies [40] relax one architecture to one SuperNet, with the weights of all candidate operations being optimized by the gradient descent method.

2.2. Particle Swarm Optimization

PSO is a kind of evolutionary computation, which is derived from research on the predation behavior of birds [49]. The main idea of particle swarm optimization is to find the optimal solution through the cooperation and information-sharing among individuals from the swarm. In PSO, each candidate solution can be thought of as a bird in the solution space, which is called a particle. The optimal solution of itself in the past is the p B e s t , and the current global optimal solution in the swarm is the g B e s t . At every iteration, particles update themselves by tracking two extremes: p B e s t and g B e s t . Each particle is updated according to the following rule of PSO:
v i , j ( t + 1 ) = w × v i , j ( t ) + c p × r p × ( p B e s t i , j x i , j ( t ) ) + c g × r g × ( g B e s t j x i , j ( t ) )
x i , j ( t + 1 ) = x i , j ( t ) + v i , j ( t + 1 )
where v i , j is the velocity of the i-th particle, x i , j is the current particle position of the i-th particle, w is a constant called momentum that controls how much the previous velocity affects the velocity at the current time step, c p , and c g are yjr constants defined beforehand, representing the weight of yjr particles approaching the p B e s t and g B e s t , and r p and r g are random numbers in [0, 1).
One of the positive characteristics of PSO is that it is capable of fast convergence [50] when compared to other evolutionary approaches, such as genetic algorithms. Due to the fact that searching CNNs takes a lot time, even on the most powerful computers, a faster converging characteristic can have a positive impact in terms of obtaining the optimal CNN architecture.

3. Proposed Method

In this section, we propose a cell-based CNN architecture search method by particle swarm optimization at first, named PSO-Net. To make the search stage more efficient and to reduce the time consumed, we developed another novel continuous PSO approach for searching neural architectures, namely CPSO-Net.
The two proposed methods are all based on the PSO architecture search step, including four procedures to search for the optimal CNN architecture: Construction of the search space, initialization of the architectures, weight optimization of the neural network constructed by the architectures, fitness evaluation of the individual particles, and particle update. CPSO-Net is an improvement based on PSO-Net, and the two methods have different procedures for the parameter optimization of architectures. The weight optimization of the former is the individual weight optimization of all particles, and the latter is weight-sharing by the inherited weights from the SuperNet.
Figure 1 shows the framework of the stages of PSO-Net. From the framework, one can see that network architectures are encoded into arrays. Then, the gradient descent method is used separately to obtain the weight parameters for all architectures, and PSO is used as the architecture search method to optimize the architectures by the fitness of all particles. Finally, the optimal architecture is found after reaching the number of iterations.
Figure 2 shows the framework of the stages of CPSO-Net, which is proposed-based PSO-Net. The main difference of Figure 2 is the procedure of the parameter optimization of architectures. To accelerate the optimal architecture generation, a SuperNet is maintained to share the parameters for all particles and is optimized by collecting the gradients of all individuals in the population. Rather than training all the individuals of PSO-Net, the SuperNet of CPSO-Net is optimized only once during one iteration, which dramatically reduces the computational complexity by separately learning each network.
All the operations of PSO-Net and CPSO-Net are detailed in the following subsections.

3.1. Construction of the Search Space

The search space contains two parts, namely all candidate basic operations and connection nodes. On the one hand, there are three main basic operation types: Convolution operations, pooling operations, and nonlinear operations. All candidate basic operations are placed into the operations list named O P to prepare for the next encoding strategy. On the other hand, the cell-based search space is used to design CNN architectures in the algorithm we proposed. Each neural network architecture is built by stacking two types of k number computation cells, which includes normal cells and one reduction cell. Each cell consists of an ordered sequence with two input nodes, n intermediate nodes, and one output node. The input of the input nodes are the cell outputs in the previous two cells, and the output of the cell is a cascade by the n intermediate nodes, each of which is calculated by the sum of two computations. Additionally, each computation is based on a random one operation of the basic operations O P computing on a random one node of its previous nodes. Each computation is defined as an element, which is composed of the basic operation and the computed node. Therefore, there are 2n elements in each cell with n intermediate nodes. Figure 3 shows the general architecture stacking by the automatically designed computation cells by PSO-Net and CPSO-Net.

3.2. Initialization of the Swarm

Particle encoding is the core of the initialization of architectures. A novel encoding strategy was devised that can be used to encode architectures into arrays with the information of all of the basic operations and connections between the nodes in computation cells—where the first and second input node is numbered the 0th and 1th in all nodes, and the n intermediate nodes are numbered in cells successively following the number of two input nodes. First, an array with a shape of (2, 2n, 2) is built, which represents an architecture. The first dimension of the array represents two types of computation cells, namely normal cells and reduction cells. The second dimension of the array represents the 2n elements (defined in Section 3.1) in each cell with n intermediate nodes. The third dimension of the array is composed of the index of basic operations list O P and the computed node. Next, P particles of the swarm are initialized by the above encoding strategy. Figure 4 shows an example of an encoded architecture, including an encoded normal cell and a reduction cell with four intermediate nodes.

3.3. The Weight Optimization of the Neural Network Constructed by Architectures

The weight of the neural network constructed by architectures should be optimized, when receiving information of the connection and basic operation type between the nodes in computation cells.

3.3.1. Individual Parameter Optimization of PSO-Net

The parameter optimization of PSO-Net involves training individual particles by the gradient descent method. Each network N i we searched can be represented by the particle P i and a set of full precision parameters W i ,   P i , i { 1 , , P } , where P is the population size. The training dataset X is denoted as the input data, and N i ( X ) is the prediction of the i-th network N i . The loss can be expressed as L i = H N i ( X ) , Y , where H is the criterion and Y is the target. The parameter W i of N i can be optimized by:
d W i = L i W i

3.3.2. Weight-Sharing Parameter Optimization of CPSO-Net

To further accelerate optimal network generation, we built a SuperNet N that includes all candidate operations and sharing parameters W for all particles. Each particle is regarded as a subnet and inherits the weight parameters from the fixed SuperNet, which can be optimized by the gradient descent method. Therefore, there is no need to train each particle from the beginning to the end, and the search time can therefore be greatly reduced. After the parameters have converged, the architecture could alternately be optimized by the PSO algorithm.
Sampling from the SuperNet N , the different network N i is a part of N , and the parameter W i of N i can be inherited from W. Therefore, W i can be represent as W i = W P i , i { 1 , , P } , where ⊙ is the mask operation that retains the parameters of the corresponding nodes of particle P i . Therefore, the gradient of W i also can be calculated as:
d W i = L i W i = L i W P i
The SuperNet shares W for all different architectures, so the gradient of parameter W should be calculated by collecting the gradients of all individuals in the population:
d W = 1 P i = 1 P d W i = 1 P i = 1 P L i W P i
Inspired by the mini-batch samples idea of stochastic gradient descent for updating parameters, we used mini-batch architectures to accumulate the gradients of all individuals for updating the shared weight W.

3.4. Fitness Evaluation

Fitness evaluation is performed using the HSI classification accuracy of the optimized neural network on the validation set, and aimed to estimate the performance of the architectures. Each particle is evaluated as a solution by the fitness evaluation function. Its optimal solution in the past is the p B e s t , and the current global optimal solution in the swarm is the g B e s t . By the fitness of PSO-Net and CPSO-Net, we can find the p B e s t and p B e s t .

3.5. Particle Update

The difference and the velocity operator are devised based on PSO to update the particles. The particle P is updated element by element until all elements of the particle have been updated, so the particle update problem is transformed into the element update problem. Before being able to update an element, the difference between two elements with the same index in two particles needs to be measured, and the velocity of the element needs to be computed.
The difference ( E 2 E 1 ) is calculated by comparing E 2 and E 1 . If both elements are the same, the difference is “None”, which means no difference. If the two elements are different, the result of ( E 2 E 1 ) is E 2 .
The velocity of any given element of a particle P _ E is based on the two differences: ( g B e s t _ E P _ E ) and ( p B e s t _ E P _ E ) , where g B e s t , p B e s t , and P have the same index in a particle. Each element velocity is chosen from the two differences based on the decision factor c g and a number r obtained at random from [0,1). If r c g , the algorithm selects from the difference ( g B e s t _ E P _ E ) . Otherwise, the algorithm selects from the difference ( p B e s t _ E P _ E ) .
Each element is updated according to its element velocity. If its element velocity is “None”, the element remains the same, otherwise it should be displaced by its velocity. Particles are modified element by element according to the element velocity. Figure 5 shows the update of an element of a particle.
The essence of the role of a particle update is that a particle update can be converted to an element update in the particle. We assumed that the elements of p B e s t and g B e s t can obtain good fitness, so that p B e s t and g B e s t can achieve good fitness. Therefore, we compared P to p B e s t and g B e s t element by element, and determined whether the element P needs to be replaced by p B e s t or g B e s t . Figure 4 shows the update process of particles.

4. Experimental Results and Analysis

In this section, we explain the experimental setting of the method proposed and a comparative algorithm, followed by the empirical results, to confirm the effectiveness of the proposed method for automatically designing CNNs for HSI classification based on four HSI datasets with biased and unbiased data. The experimental conditions were as follows: A NVIDIA GTX1080 Ti GPU, a E5-2620 CPU, and a memory of 32 GB.

4.1. Datasets

In this section, we used four standard hyperspectral datasets to evaluate the methods we proposed. As shown in Table 1, the four datasets were Salinas Valley in California in the USA (Salinas), a mixed vegetation site over the Indian Pines test area in northwestern Indiana in the USA (Indian Pines), an urban site over the University of Pavia in Italy (Pavia), and the Kennedy Space Center (KSC) in Florida in the USA. The detailed training samples of each class are shown in Table 1.
The first dataset, Salinas, was collected by the Airborne Visible Infrared Imaging Spectrometer (AVIRIS) with 224-band over Salinas Valley, California, which is of a high spatial resolution (3.7 m/pixels). The available dataset consists of 204 bands with 512 × 217 pixels after discarding the 20 water absorption bands and the 16 land cover classes.
The second dataset, Indian Pines, was captured by the AVIRIS sensor with 220-band over the Indian Pines test area. The available dataset consists of 200 bands with 145 × 145 pixels after removing the water absorption bands. The ground reference map covers 16 classes of interest.
The third dataset, University of Pavia, was collected by the ROSIS-3 sensor with 115-band over the University of Pavia, which is of a high spatial resolution (1.3 m/pixels) to avoid mixed pixels. The available dataset consists of 103 bands with 610 × 340 pixel vectors after removing the noisy bands. The ground reference map covers nine classes of interest.
The fourth dataset, KSC, was captured by the airborne AVIRIS sensor over KSC, Florida, at an altitude of approximately 20 km with a spatial resolution of 18 m. The available dataset consists of 176 bands with 512 × 614 pixels after removing the water absorption and noisy bands. The ground reference map covers 13 classes of interest.

4.2. Comparative Experiment

To evaluate the performance of the two methods we proposed, some CNN classification methods, including handcrafted CNN models and automatically designed CNN models based on spatial–spectral information, were employed to achieve a comprehensive comparison with the PSO-Net and CPSO-Net methods. We chose several handcrafted CNN models that have reported results in the spectral–spatial information hyperspectral datasets to test the proposed algorithms. The CNN [51] was conducted, which is a sample CNN designed to tackle the problem of HSI spectral–spatial classification. The spectral–spatial residual network (SSRN) [22] and ResNet [27] are residual-based methods, which can achieve good classification performance. DesNet [21] using shortcut connections between layers was used for comparison. Moreover, an automatically designed CNN method, namely Auto-CNN, was introduced, which explores the automatic design of CNNs for HSI classification by the gradient descent and achieved good performance. To demonstrate the performance of the two methods we proposed, we made a comparison in the classification accuracy and the computational complexity with handcrafted and automatically designed CNN models. To further illustrate the search effectiveness of PSO-Net and CPSO-Net, we compared them with a peer, the automatically designed method Auto-CNN, in terms of time complexity.

4.3. Experimental Settings

In this section, we explain the detailed parameter setting for our experiments. For the automatically designed algorithms, including Auto-CNN, PSO-Net, and CPSO-Net, the search space should be predefined. We set the number of computation cells k to 3 and the intermediate nodes n to 4, and the basic operation list O P includes the following candidate basic operations: None; 3 × 3 max pooling, abbreviated as max_3 × 3; 3 × 3 average pooling, abbreviated as avg_3 × 3; skip connection, abbreviated as skip; 3 × 3 separable convolution, abbreviated as sep_3 × 3; 5 × 5 separable convolution, abbreviated as sep_5 × 5; 3× 3 dilated convolution, abbreviated as dil_3 × 3; 5 × 5 dilated convolution, abbreviated as dil_5 × 5. For HSI classification, the general architecture of automatically designed CNNs includes three parts. First, hyperspectral datasets as an input pass one 1 × 1 bottleneck convolution layers to set the number of HSI bands to 10 for the next computation cells. Then, two normal cells with padding and setting stride to 1, and reduction cells with padding and setting stride to 2 are stacked to complete the CNN. Finally, the final labels can be obtained by the fully connected layer.
For all the experiments in this work, we split the labeled samples of the HSI datasets into three subsets by sampling without replacement, namely the training, validation, and test sets. To ensure that all classes are included in the training, validation, and test dataset, we choose samples including all classes based on the proportion of every class in whole dataset and samples in each class are randomly selected with the same probability in this class. Based on the rule above, we chose 200 samples, including all classes as the training set and 600 samples as the validation set for each dataset. To demonstrate the stability of the two proposed methods, we conducted the experiments five times with random split datasets. Details of the training samples of each class of the four datasets are shown in Table 1. All the remaining labeled samples served as the test set to evaluate the capability of the network. The training set was used to train the network, and the validation set was used to evaluate the performance of the architectures found and to search for the optimal architecture for all the automatic design algorithms.
For biased datasets, we chose 7 × 7 × B neighborhoods of a pixel as input, where B is the number of HSI bands. For unbiased datasets, samples with 32 × 32 slides were not enough to divide into the training and validation datasets, so we chose 7 × 7 neighborhoods of a pixel, and none of the samples were overlapping. For all the CNN models designed by hand, i.e., 3-D CNN, SSRN, ResNet, DesNet, we set the training epochs to 100 epochs. The learning rate for the weights of all these models was 0.025. For 3-D Auto-CNN, we set the training epochs in the architecture search and test process both to 100 epochs. The learning rate for the weights in the search and test process was 0.025 and 0.05, respectively. For PSO-Net, we set the swarm size to 100, the number of iterations to 40, c g to 0.5, the training epochs in the architecture search and test process both to 100 epochs, and the learning rate for the weights in the architecture search and test process to 0.025. For CPSO-Net, only the training epochs in the architecture search were different from above, inspired by the mini-batch architecture idea of stochastic gradient descent for updating parameters, with only one epoch achieving good performance. For the test process, the experimental settings for the different methods can be found in Table 2.

4.4. Results Analysis

In this section, we analyze the classification results and compare our algorithm with other methods in terms of classification accuracy, parameters, and time complexity on biased hyperspectral datasets to illustrate the effectiveness of our algorithms. We also demonstrate the optimal architectures searched by our methods, and analyze the convergence to illustrate the feasibility of the methods we proposed. Finally, we conduct experiments on unbiased hyperspectral datasets to further show our algorithms’ advantages.

4.4.1. Classification Accuracy

Table 3, Table 4, Table 5 and Table 6 show the detailed classification results on test datasets of the two methods we proposed and the other comparative algorithm. As shown in Table 3, Table 4, Table 5 and Table 6, it is obvious that PSO-Net and CPSO-Net achieved a big improvement classification accuracy over the handcraft methods, including CNN, SSRN, ResNet, and DesNet. Therefore, we mainly compared the methods with Auto-CNN, which is an automatically designed method as well. For the University of Pavia dataset, CPSO-Net improved the OA, AA, and Kappa of Auto-CNN by 1.34%, 0.64%, and 0.0185, respectively, and PSO-Net exhibited the best OA, AA, and Kappa, with improvements of 1.78%, 2.31%, and 0.0221 over Auto-CNN. respectively. For the Salinas dataset, PSO-Net and CPSO-Net performed almost the same, and CPSO-Net exhibited the best OA, AA and Kappa, with improvements of 1.22%, 1.95% and 0.0221 over Auto-CNN, respectively. For the Indian Pines dataset, both PSO-Net and CPSO-Net achieved higher classification results than the other algorithms. CPSO-Net exhibited the best AA and Kappa, while PSO-Net exhibited the best OA, with improvements of 1.88%, 0.0171, and 1.37% over Auto-CNN, respectively. For the KSC dataset, similar results to the other datasets were obtained. Obviously, CPSO-Net dramatically accelerated the optimal network generation and achieved quality performance comparable to PSO-Net. For the Salinas, Indian Pines, and KSC datasets, CPSO-Net achieved even better classification performance than PSO-Net, with some improvement over PSO-Net in terms of AA, OA, and Kappa. There may be some significant differences in every class accuracy. For example, CPSO-Net achieved particularly improvement of 12.98% over Auto-CNN in the first-class classification on Salinas. In addition, Table 7 shows the detailed average rankings of the classification accuracy on four datasets through the minimize performance measure based Friedman test [52], which is a statistical test for the homogeneity of multiple samples. From Table 7, we can see that the proposed approaches get almost similar ranking and rank higher than other comparative algorithms in terms of OA, AA, and Kappa.
To sum up from the classification accuracy analysis above, we can draw two conclusions. On the one hand, the PSO-Net and CPSO-Net algorithms made significant improvements over the state-of-the-art CNN models designed by hand, including CNN, SSRN, ResNet, and DesNet. Additionally, the proposed approaches are able to find better architectures than Auto-CNN optimized by the gradient descent. On the other hand, the two methods proposed herein achieved almost the same classification performance, while CPSO-Net dramatically reduced the time computation.
In addiction, taking CPSO-Net as example, we made a sensitivity hyperparameter analysis of the cell numbers with an additional two and four cells, intermediate nodes with an additional three and five nodes, and a population size with an additional 60 and 140. From Figure 6a,b, it can be seen that fewer number of cells or intermediate nodes achieved a lower OA, but the models with a higher number of cells or intermediate nodes did not necessarily lead to better performance. The main reason is that the models with simple architectures may not be able to extract deep features, and complex models may cause an overfitting problem with limited training samples. From Figure 6c, we can see that a small population may make architectures fall into the local optimal, but a large population will not greatly improve the classification performance when spending much more time.

4.4.2. Complexity Analysis

To analyze the complexity of the method we proposed, the results of several trainable parameters compared to both human-designed and automatically designed CNNs are shown in the bottom row of Table 3, Table 4, Table 5 and Table 6, and the time consumed compared to the peer automatically designed method are shown in Figure 7. From Table 3, Table 4, Table 5 and Table 6, it is obvious that the models searched by PSO-Net and CPSO-Net achieved better results with the state-of-the-art CNNs designed by human experts while having significantly less parameters, which means less time and computational complexity to train the CNN models. There are two main reasons for the above results. First, a CNN model with a high number of trainable parameters may cause an overfitting problem with limited training samples. Second, automatically designed methods can search CNNs more suitably for specific HSI datasets. For the automatically designed methods Auto-CNN, PSO-Net, and CPSO-Net, even though they had the same search space, CPSO-Net and PSO-Net performed better than Auto-CNN. However, there was no consistent pattern between the number of parameters of the models searched by the three automatically designed methods.
To further compare the search effectiveness of these automatically designed methods, Figure 7 shows the time consumed in the different stages of the peer automatically designed CNN methods based on four biased HSI datasets. PSO-Net consumed approximately 15 h to converge, which is a significant time compared to the other two automatically designed methods. Therefore, Figure 7a only shows the search time of Auto-CNN, CPSO-Net, and CPSO-Net convergence. From Table 3, Table 4, Table 5 and Table 6 and Figure 7a, we can see that CPSO-Net and PSO-Net achieved better performance with a big improvement for Auto-CNN, but CPSO-Net dramatically reduced the search time by the weight-sharing parameters from the SuperNet compared to PSO-Net. The time of the total iteration of CPSO-Net consumed more time than Auto-CNN, while it converged at six, eight, nine, and eight iterations (the detailed convergence analysis can be seen is Section 4.4.3) based on Pavia, Salinas, Indian Pines, and KSC, respectively. CPSO-Net only consumed approximately 15 min to converge, which is almost close to and even less than Auto-CNN. From Figure 7b, we can see that the training time of the CNNs searched by CPSO-Net was less than that of PSO-Net and Auto-CNN based on the Pavia, Salinas, and Indian Pines datasets. From Figure 7c, the three automatic-designed method, i.e., Auto-CNN, PSO-Net, and CPSO-Net, consumed almost the same time to test all the data of the CNNs searched.
To sum up, from Figure 7, we can see that PSO-Net consumed a significant amount of time to converge, while CPSO-Net dramatically accelerated the search generation. The time CPSO-Net consumed to converge was almost same as and even less than Auto-CNN.

4.4.3. Convergence Analysis

To dramatically accelerate optimal network generation, CPSO-Net shared the weights of SuperNet for all particles based on PSO-Net and achieved quality performance comparable to PSO-Net. Therefore, we mainly analyze the convergence of CPSO-Net in this section. The convergence of the architectures was different from the other solutions, and it is not enough only to consider the fitness of particles. The fitness was determined by the HSI classification accuracy of the optimized neural network, which was composed of the architectures and weight parameters. However, the number of parameters of the architectures and the position of the particles were only related to architecture. Therefore, the architecture convergence analysis of CPSO-Net took the number of parameters of the architectures and the position of all the particles as important criteria.
The number of parameters of the architectures were closely related to the operations in said architectures. With the alter of the operations in the architectures, the number of parameters changed accordingly, representing the architectures converging when the number of parameters remained the same during the generations. Figure 8 shows the accuracy and number of parameters of g B e s t during iterations based on four datasets by CPSO-Net. From the number of parameters of g B e s t , we can see that the architectures converged at six, eight, nine, and eight iterations based on the Pavia, Salinas, Indian Pines, and KSC datasets, respectively. The g B e s t accuracy of the validation dataset still increased after the convergence of architectures. The main reason is that the parameters of the architectures inherited from the SuperNet continued to be optimized, while SuperNet was trained until reaching the maximum iterations.
We further visualized the position of 20 particles selected randomly and g B e s t during the iterations based on four datasets by CPSO-Net in Figure 9. The position of the particles was projected from the array (encoding from the architectures) into one-dimensional space using principal component analysis (PCA). As we can see, all particles divided globally at initialization, gradually tended toward focusing on global exploration with the iteration, toward the best particle position found thus far, and the position of g B e s t converged at the same iteration as the number of parameters of g B e s t , as shown in Figure 8.

4.4.4. Optimal Architecture

To find the optimal architecture is the main purpose of the method we proposed, the optimal architecture can be searched at the end of the algorithms. After searching, the optimal architecture can be used to classify all the data. Herein, we took PSO-Net based on biased the University of Pavia dataset and CPSO-Net based on the biased Salinas dataset as examples. The architectures of the optimal CNN architectures found by the proposed PSO-Net and CPSO-Net are shown in Figure 10 and Figure 11. Compared to the handcraft CNNs, PSO-Net and CPSO-Net designed the optimal architectures in a predefined search space without professional knowledge of human experts and without spending significant time on repeated debugging. The search space is constructed by incorporating typical properties of successful architectures designed by hand. Therefore, the optimal architectures searched includes some typical properties of handcraft CNNs. For example, inspired by the ResNet, the optimal architectures in Figure 10 and Figure 11 introduced skip connections.

4.4.5. Classification Maps

To clearly represent the classification results, we plotted the whole image classification maps of all the models. The classification maps obtained by the different models based on four biased datasets are shown in Figure 12, Figure 13, Figure 14 and Figure 15. We evaluated the classification accuracies from the noise scatter in the visual perspective. From the results, we can see that CPSO-Net achieved less scatter in the class with the largest area compared to the other models, which means that CPSO-Net can obtain more precise classification accuracy in this class. From the resulting images, we figured out how the different classification methods affected the classification results. Obviously, the automatically designed CNNs had less noise than the handcrafted CNN model, which demonstrates the feasibility of designing CNN models automatically. It is obvious that the CNN search by the methods we proposed had less scatter than the automatically design method Auto-CNN for HSI classification, which further demonstrates the effectiveness of PSO-Net and CPSO-Net.

4.5. Unbiased HSI Classification

There may be pixel overlapping and training-test information leakage problems on biased datasets which lead to over-optimistic results for spectral-spatial methods. Therefore, to further demonstrate the effectiveness of PSO-Net and CPSO-Net, we conducted some more experiments on the unbiased hyperspectral datasets and tested the two methods in comparison to CNN, SSRN, ResNet, DesNet, and Auto-CNN based on the four biased HSI datasets of Pavia, Salinas, Indian Pines, and KSC. There are some methods to split training and test datasets without pixels overlapping and training-test information leakage, such as patch-based [53] and set-to-sets distance based [54] methods.
For the unbiased datasets, samples with 32 × 32 slides were not enough for dividing into the training and test datasets, so we chose 7 × 7 neighborhoods of a pixel according to the patch-based [53] method, and none of the samples were overlapping. When the All other experimental settings were the same with the aforementioned biased datasets.
From Table 8, Table 9, Table 10 and Table 11, it is obvious to see that PSO-Net and CPSO-Net performed much better than the other models, and the improvements of PSO-Net and CPSO-Net were significant. In particular, the performance of CPSO-Net was better and consumed much less time than PSO-Net. However, the results of different unbiased spatial-based HSI classification methods were all lower than the performance of the biased methods.

5. Conclusions

In this paper, we proposed two novel evolutionary-based search methods, PSO-Net and CPSO-Net, which use particle swarm optimization as the search strategy to search for the optimal architecture. For PSO-Net, the gradient descent method was used to obtain the weight parameters for all architectures separately, and PSO was used to optimize the architectures and search for the optimal architecture until reaching the maximum number of iterations. To accelerate the optimal network generation, CPSO-Net made an improvement based on the PSO-Net parameter optimization of architectures. For CPSO-Net, we maintained a SuperNet with all candidate operations and shared the weight parameters for all individuals. There was no need to train each particle from the beginning to the end, so the search time was reduced greatly. Our results showed that PSO-Net and CPSO-Net can achieve better classification accuracies compared to the state-of-the-art algorithms of the automatic design of CNNs for HSI classification, and CPSO-Net can obtain almost the same results with much less time consumed than PSO-Net.
For future works, we will consider multiple complementary objectives, such as the number of parameters and the classification accuracy to achieve the Pareto optimal front. To further demonstrate the effectiveness of the methods we proposed herein, we will test these methods on the same common image classification datasets.

Author Contributions

Conceptualization, X.L. and C.Z.; methodology, X.L.; software, X.L. and C.Z.; validation, X.L., C.Z., and Z.C.; formal analysis, X.L.; investigation, C.Z.; resources, X.L. and Z.C.; data curation, C.Z.; writing—original draft preparation, C.Z.; writing—review and editing, X.L., Z.C., J.Y., Z.Z., and X.G.; visualization, C.Z.; supervision, X.L.; project administration, X.L. and Z.C.; funding acquisition, X.L. All authors read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Nature Science Foundation of China (grant numbers 61973285, 62076226, 61873249, and 61773355) and the Open Research Project of the Hubei Key Laboratory of Intelligent Geo-Information Processing (grant number KLIGIP-2019A04).

Acknowledgments

The authors would like to thank M. Gra n ¯ a, M.A. Veganzons, and B. Ayerdi for the hyperspectral image datasets used in this paper, and Y. Chen for the Auto-CNN source codes.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  2. Yi, C.; Zhao, Y.Q.; Chan, J.C.W. Hyperspectral image super-resolution based on spatial and spectral correlation fusion. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4165–4177. [Google Scholar] [CrossRef]
  3. Kang, X.; Zhang, X.; Li, S.; Li, K.; Li, J.; Benediktsson, J.A. Hyperspectral anomaly detection with attribute and edge-preserving filters. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5600–5611. [Google Scholar] [CrossRef]
  4. Haut, J.M.; Bernabé, S.; Paoletti, M.E.; Fernandez-Beltran, R.; Plaza, A.; Plaza, J. LowChigh-power consumption architectures for deep-learning models applied to hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2018, 16, 776–780. [Google Scholar] [CrossRef]
  5. Teke, M.; Deveci, H.S.; Haliloglu, O.; Gürbüz, S.Z.; Sakarya, U. A short survey of hyperspectral remote sensing applications in agriculture. In Proceedings of the 2013 6th International Conference on Recent Advances in Space Technologies (RAST), Istanbul, Turkey, 16 August 2013; pp. 171–176. [Google Scholar]
  6. Shang, X.; Chisholm, L.A. Classification of australian native forest species using hyperspectral remote sensing and machine-learning classification algorithms. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2481–2489. [Google Scholar] [CrossRef]
  7. Han, Y.; Li, J.; Zhang, Y.; Hong, Z.; Wang, J. Sea ice detection based on an improved similarity measurement method using hyperspectral data. Sensors 2017, 17, 1124. [Google Scholar]
  8. El-Sharkawy, Y.H.; Elbasuney, S. Hyperspectral imaging: Anew prospective for remote recognition of explosive materials. Remote Sens. Appl. Soc. Environ. 2019, 13, 31–38. [Google Scholar] [CrossRef]
  9. Li, Y.; Zhang, H.; Shen, Q. Spectral-spatial classification of hyperspectral imagery with 3d convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  10. Luo, Y.; Zou, J.; Yao, C.; Zhao, X.; Li, T.; Bai, G. Hsi-cnn: A novel convolution neural network for hyperspectral image. In Proceedings of the International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 16–17 July 2018; pp. 464–469. [Google Scholar]
  11. Santara, A.; Mani, K.; Hatwar, P.; Singh, A.; Garg, A.; Padia, K.; Mitra, P. BASS Net: Band-adaptive spectral-spatial feature learning neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5293–5301. [Google Scholar] [CrossRef] [Green Version]
  12. Cai, Y.; Liu, X.; Cai, Z. BS-Nets: An end-to-end framework for band selection of hyperspectral image. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1969–1984. [Google Scholar] [CrossRef] [Green Version]
  13. Cai, Y.; Zhang, Z.; Cai, Z.; Liu, X.; Jiang, X.; Yan, Q. Graph convolutional subspace clustering: A robust subspace clustering framework for hyperspectral image. IEEE Trans. Geosci. Remote Sens. 2020. [Google Scholar] [CrossRef]
  14. Feng, J.; Wang, L.; Yu, H.; Jiao, L.; Zhang, X. Divide-and-conquer dual-architecture convolutional neural network for classification of hyperspectral images. Remote Sens. 2019, 11, 484. [Google Scholar] [CrossRef] [Green Version]
  15. Charmisha, K.; Sowmya, V.; Soman, K. Dimensionally reduced features for hyperspectral image classification using deep learning. In Proceedings of the International Conference on Communications and Cyber Physical Engineering (ICCCE), Hyderabad, India, 24–25 January 2018; pp. 171–179. [Google Scholar]
  16. Du, J.; Li, Z. A hyperspectral target detection framework with subtraction pixel pair features. IEEE Access. 2018, 6, 45562–45577. [Google Scholar] [CrossRef]
  17. Jia, P.; Zhang, M.; Yu, W.; Shen, F.; Shen, Y. Convolutional neural network based classification for hyperspectral data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5075–5078. [Google Scholar]
  18. Haut, J.; Paoletti, M.; Plaza, J.; Plaza, A. Hyperspectral image classification using random occlusion data augmentation. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1751–1755. [Google Scholar] [CrossRef]
  19. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial-spectral feature learning by 3d convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  20. Sellami, A.; Farah, M.; Farah, I.R.; Solaiman, B. Hyperspectral imagery classification based on semi-supervised 3-d deep neural network and adaptive band selection. Expert Syst. Appl. 2019, 129, 246–259. [Google Scholar] [CrossRef]
  21. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep&Dense convolutional neural network for hyperspectral image classification. Remote Sens. 2018, 10, 1454. [Google Scholar]
  22. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral-spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  23. Fang, B.; Li, Y.; Zhang, H.; Chan, J. Hyperspectral images classification based on dense convolutional networks with spectral-wise attention mechanism. Remote Sens. 2019, 11, 159. [Google Scholar] [CrossRef] [Green Version]
  24. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  25. Safari, K.; Prasad, S.; Labate, D. A multiscale deep learning approach for high-resolution hyperspectral image classification. IEEE Geosci. Remote Sens Lett. 2021, 18, 167–171. [Google Scholar] [CrossRef]
  26. Shao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. He, X.; Zhao, K.; Chu, X. AutoML: A survey of the state-of-the-art. Knowl. Based Syst. 2021, 212, 1–27. [Google Scholar] [CrossRef]
  29. Elsken, T.; Metzen, J.H.; Hutter, F. Neural architecture search: A survey. J. Mach. Learn. Res. 2019, 20, 1–21. [Google Scholar]
  30. Zoph, B.; Le, Q.V. Neural architecture search with reinforcement learning. arXiv 2017, arXiv:1611.01578. [Google Scholar]
  31. Chen, Y.; Zhu, K.; Zhu, L.; He, X.; Ghamisi, G.; Benediktsson, J.A. Automatic design of convolutional neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7048–7066. [Google Scholar] [CrossRef]
  32. Junior, F.; Erivaldo, F.; Yen, G. Particle swarm optimization of deep neural networks architectures for image classification. Swarm Evol. Comput. 2019, 49, 62–74. [Google Scholar] [CrossRef]
  33. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the Conference on Artificial Intelligence (AAAI), Honolulu, HI, USA, 27 January–1 February 2019; pp. 4780–4789. [Google Scholar]
  34. Stanley, K.O.; Miikkulainen, R. Evolving neural networks through augmenting topologies. IEEE Trans. Evol. Comput. 2002, 10, 99–127. [Google Scholar] [CrossRef]
  35. Gauci, J.; Stanley, K.O. Autonomous evolution of topographic regularities in artificial neural networks. Neural Comput. 2010, 22, 1860–1898. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Stanley, K.O.; D’Ambrosio, D.B.; Gauci, J. A hypercube-based indirect encoding for evolving large-scale neural networks. Artif. Life 2009, 15, 185–212. [Google Scholar] [CrossRef]
  37. Stanley, K.O. Compositional pattern producing networks: A novel abstraction of development. Genet. Program Evol. Mach. 2007, 8, 131–162. [Google Scholar] [CrossRef] [Green Version]
  38. Baker, B.; Gupta, O.; Naik, N.; Raskar, R. Designing Neural Network Architectures Using Reinforcement Learning. arXiv 2017, arXiv:1611.02167. [Google Scholar]
  39. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8697–8710. [Google Scholar]
  40. Liu, H.; Simonyan, K.; Yang, Y. Darts: Differentiable architecture search. arXiv 2019, arXiv:1806.09055. [Google Scholar]
  41. Xie, S.; Zheng, H.; Liu, C.; Lin, C. Snas: Stochastic neural architecture search. arXiv 2019, arXiv:1812.09926. [Google Scholar]
  42. Wu, B.; Dai, X.; Zhang, P.; Wang, Y.; Sun, F.; Wu, Y.; Tian, Y.; Vajda, P.; Jia, Y.; Keutzer, K. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10734–10742. [Google Scholar]
  43. Pham, H.; Guan, M.Y.; Zoph, B.; Le, Q.V.; Dean, J. Efficient neural architecture search via parameter sharing. In Proceedings of the Internation Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 6522–6531. [Google Scholar]
  44. Xie, L.; Yuille, A. Genetic cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1388–1397. [Google Scholar]
  45. Yang, Z.; Wang, Y.; Chen, X.; Shi, B.; Xu, C.; Xu, C.; Tian, Q.; Xu, C. CARS: Continuous evolution for efficient neural architecture search. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1829–1838. [Google Scholar]
  46. Elsken, T.; Metzen, J.H.; Hutter, F. Efficient multi-objective neural architecture search via lamarckian evolution. arXiv 2019, arXiv:1804.09081. [Google Scholar]
  47. Brock, A.; Lim, T.; Ritchie, J.M.; Weston, N. Smash: One-shot model architecture search through hypernetworks. arXiv 2017, arXiv:1708.05344. [Google Scholar]
  48. Saxena, S.; Verbeek, J. Convolutional neural fabrics. In Proceedings of the Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 4060–4068. [Google Scholar]
  49. Eberhart, R.; Kennedy, J. A new optimizer using particle swarm theory. In Proceedings of the 6th International Symposium on Micro Machine and Human Science (MHS), Nagoya, Japan, 4–6 October 1995; pp. 39–43. [Google Scholar]
  50. Sahu, A.; Panigrahi, S.K.; Pattnaik, S. Fast convergence particle swarm optimization for functions optimization. Procedia Technol. 2012, 4, 319–324. [Google Scholar] [CrossRef] [Green Version]
  51. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  52. Friedman, M. A comparison of alternative tests of significance for the problem of m rankings. Ann. Math. Stat. 1940, 30, 86–92. [Google Scholar] [CrossRef]
  53. Nalepa, J.; Myller, M.; Kawulok, M. Validating hyperspectral image segmentation. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1264–1268. [Google Scholar] [CrossRef] [Green Version]
  54. Cao, X.; Ren, M.; Zhao, J.; Lu, H.; Jiao, L. Non-overlapping classification of hyperspectral imagery based on set-to-sets distance. Neurocomputing 2020, 378, 422–434. [Google Scholar] [CrossRef]
Figure 1. Framework of the stages of particle swarm optimization PSO-Net. The constructed network architectures from the search space are encoded into arrays, and populations are initialized. The gradient descent method is used to obtain the weight parameters for all architectures separately, and PSO is used to optimize the architectures and to search for the optimal architecture until reaching the number of iterations. The output is the optimal architecture.
Figure 1. Framework of the stages of particle swarm optimization PSO-Net. The constructed network architectures from the search space are encoded into arrays, and populations are initialized. The gradient descent method is used to obtain the weight parameters for all architectures separately, and PSO is used to optimize the architectures and to search for the optimal architecture until reaching the number of iterations. The output is the optimal architecture.
Remotesensing 13 01082 g001
Figure 2. Framework of CPSO-Net. A continuous SuperNet that contains all candidate operations and can be trained by gradient descent is maintained first. Then, all architectures are encoded into an array, inheriting the parameters from the SuperNet, and search for the optimal architecture by PSO. Finally, the optimal architecture is found.
Figure 2. Framework of CPSO-Net. A continuous SuperNet that contains all candidate operations and can be trained by gradient descent is maintained first. Then, all architectures are encoded into an array, inheriting the parameters from the SuperNet, and search for the optimal architecture by PSO. Finally, the optimal architecture is found.
Remotesensing 13 01082 g002
Figure 3. The general architecture stacking by the automatically designed computation cells by PSO-Net and CPSO-Net. The bottleneck convolution layer and the fully connected layer in Figure 3 are designed for data preprocessing and classification with human knowledge.
Figure 3. The general architecture stacking by the automatically designed computation cells by PSO-Net and CPSO-Net. The bottleneck convolution layer and the fully connected layer in Figure 3 are designed for data preprocessing and classification with human knowledge.
Remotesensing 13 01082 g003
Figure 4. An example of an encoded architecture, including an encoded normal cell and a reduction cell with four intermediate nodes. There are eight elements in each cell with four intermediate nodes. (a) An example of an encoded normal cell. The 0th intermediate node, as well the 2th node in the cell, are calculated by the sum of two elements O P [2] computed based on the 0th node in the cell and O P [4] computed based on the 1th node in the cell, where the 0th and 1th nodes in the cell are the two input nodes. All the remaining intermediate nodes are encoded into arrays in the same way. (b) An example of an encoded reduction cell.
Figure 4. An example of an encoded architecture, including an encoded normal cell and a reduction cell with four intermediate nodes. There are eight elements in each cell with four intermediate nodes. (a) An example of an encoded normal cell. The 0th intermediate node, as well the 2th node in the cell, are calculated by the sum of two elements O P [2] computed based on the 0th node in the cell and O P [4] computed based on the 1th node in the cell, where the 0th and 1th nodes in the cell are the two input nodes. All the remaining intermediate nodes are encoded into arrays in the same way. (b) An example of an encoded reduction cell.
Remotesensing 13 01082 g004
Figure 5. The update process for an element in a particle. This example shows an update of an element of the 2 n d intermediate node. Before being able to update the element of a particle, the difference between ( g B e s t _ E P _ E ) or ( p B e s t _ E P _ E ) by the different, and the velocity of the element needs to be computed using the velocity operator.
Figure 5. The update process for an element in a particle. This example shows an update of an element of the 2 n d intermediate node. Before being able to update the element of a particle, the difference between ( g B e s t _ E P _ E ) or ( p B e s t _ E P _ E ) by the different, and the velocity of the element needs to be computed using the velocity operator.
Remotesensing 13 01082 g005
Figure 6. Classification results of different models searched by CPSO-Net with different numbers of cells, intermediate nodes, and population sizes based on four biased hyperspectral image (HSI) datasets.
Figure 6. Classification results of different models searched by CPSO-Net with different numbers of cells, intermediate nodes, and population sizes based on four biased hyperspectral image (HSI) datasets.
Remotesensing 13 01082 g006
Figure 7. Time consumed in the different stages of the different automatically designed CNNs methods based on four biased HSI datasets.
Figure 7. Time consumed in the different stages of the different automatically designed CNNs methods based on four biased HSI datasets.
Remotesensing 13 01082 g007
Figure 8. The accuracy and parameters of g B e s t during the iterations based on four biased datasets by CPSO-Net.
Figure 8. The accuracy and parameters of g B e s t during the iterations based on four biased datasets by CPSO-Net.
Remotesensing 13 01082 g008
Figure 9. Positions of 20 particles and g B e s t during the iterations based on four biased HSI datasets by CPSO-Net.
Figure 9. Positions of 20 particles and g B e s t during the iterations based on four biased HSI datasets by CPSO-Net.
Remotesensing 13 01082 g009
Figure 10. Optimal architecture of PSO-Net for the biased Pavia dataset.
Figure 10. Optimal architecture of PSO-Net for the biased Pavia dataset.
Remotesensing 13 01082 g010
Figure 11. Optimal architecture of CPSO-Net for the biased Salinas dataset.
Figure 11. Optimal architecture of CPSO-Net for the biased Salinas dataset.
Remotesensing 13 01082 g011
Figure 12. Classification Maps on biased Pavia dataset.
Figure 12. Classification Maps on biased Pavia dataset.
Remotesensing 13 01082 g012
Figure 13. Classification maps based on the biased Salinas dataset.
Figure 13. Classification maps based on the biased Salinas dataset.
Remotesensing 13 01082 g013
Figure 14. Classification maps based on the biased Indian Pines dataset.
Figure 14. Classification maps based on the biased Indian Pines dataset.
Remotesensing 13 01082 g014
Figure 15. Classification maps based on the biased KSC dataset.
Figure 15. Classification maps based on the biased KSC dataset.
Remotesensing 13 01082 g015
Table 1. Number of samples in the Salinas Valley (Salinas), Indian Pines, University of Pavia (Pavia), and Kennedy Space Center (KSC) hyperspectral datasets.
Table 1. Number of samples in the Salinas Valley (Salinas), Indian Pines, University of Pavia (Pavia), and Kennedy Space Center (KSC) hyperspectral datasets.
Salinas Valley (Salinas)Indian Pines
No.ClassSamplesTrainingNo.ClassSamplesTraining
1Broccoli_green_weeds_1200981Alfalfa461
2Broccoli_green_weeds_23726142Corn–notill142828
3Fallow197683Corn–mintill83016
4Fallow_rough_plow139464Corn2374
5Fallow_smooth2678105Grass–pasture48310
6Stubble3959146Grass–trees73014
7Celery3579147Grass–pasture–mowed281
8Grapes_untrained11,271408Hay–windrowed47810
9Soil_vineyard_develop6203229Oats201
10Corn_senesced_green_weeds32781210Soybean–notill97218
11Lettuce_romaine_4wk1068411Soybean–mintill245547
12Lettuce_romaine_5wk1927812Soybean–clean59312
13Lettuce_romaine_6wk916413Wheat2054
14Lettuce_romaine_7wk1070414Woods126524
15Vineyard_untrained72682615Buildings–grass–trees–drives3868
16Vineyard_vertical_trellis1807616Stone–steel–towers932
Total43,980200Total10,249200
University of Pavia (Pavia)Kennedy Space Center (KSC)
No.ClassSamplesTrainingNo.ClassSamplesTraining
1Asphalt6631341Scrub76124
2Meadows18,649702Willow swamp24310
3Gravel2099123Cp hammock25610
4Trees3064144Slash pine25210
5Painted metal sheets134565Oak/broadleaf26110
6Bare soil5029306Hardwood22910
7Bitumen1330147Swamp1056
8Self-blocking bricks3682168Graminoid marsh43120
9Shadows94749Spartina marsh52020
10Cattail marsh40418
11Salt marsh41918
12Mud flats50320
13Water92724
Total42,776200Total5211200
Table 2. Experimental settings for the different methods. CNN, convolutional neural network.
Table 2. Experimental settings for the different methods. CNN, convolutional neural network.
AlgorithmArchitectureCNN TrainingPSO
ParametersCellsNodesTrain_SamplesValid_SamplesEpochs_SearchEpochs_TestPop_SizeIterations
Hand-CNNs//200600/100//
Auto-CNNs34200600/100//
PSO-Net3420060010010010040
CPSO-Net34200600110010040
Table 3. Classification accuracy with bolding the best accuracy based on the biased Pavia dataset, run five times.
Table 3. Classification accuracy with bolding the best accuracy based on the biased Pavia dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)91.79 ± 0.9891.65 ± 2.6488.67 ± 0.6591.86 ± 0.2392.58 ± 0.8294.36 ± 0.5593.96 ± 1.23
AA (%)84.59 ± 2.9789.00 ± 3.2780.86 ± 1.1284.86 ± 1.4987.75 ± 1.2390.06 ± 1.0288.39 ± 2.11
K × 10089.42 ± 1.5889.16 ± 3.3785.60 ± 1.3489.63 ± 1.8790.47 ± 1.3092.68 ± 1.0392.32 ± 1.07
Asphalt92.11 ± 2.4095.08 ± 2.4691.90 ± 2.4894.32 ± 1.1695.16 ± 1.3895.68 ± 1.3595.31 ± 1.48
Meadows98.88 ± 0.7296.20 ± 4.2097.16 ± 0.2398.38 ± 0.8498.17 ± 0.7999.05 ± 0.4898.45 ± 0.59
Gravel59.99 ± 13.8569.92 ± 18.3679.43 ± 8.2781.87 ± 9.4882.65 ± 12.9370.93 ± 14.3984.23 ± 13.21
Trees93.35 ± 4.3590.84 ± 6.9352.57 ± 12.7649.15 ± 8.8263.49 ± 14.8276.46 ± 6.3271.24 ± 5.92
Painted metal sheets97.10 ± 4.3099.25 ± 2.0199.33 ± 0.3599.48 ± 0.3896.72 ± 2.4996.45 ± 1.0598.28 ± 3.16
Bare Soil93.97 ± 3.2489.34 ± 10.8893.42 ± 2.4996.26 ± 2.8398.76 ± 0.3898.84 ± 0.4799.54 ± 0.46
Bitumen54.23 ± 12.4882.36 ± 14.0187.16 ± 1.8494.24 ± 2.8595.82 ± 1.3797.95 ± 1.2597.43 ± 1.85
Self-blocking bricks88.77 ± 5.6479.67 ± 18.6895.42 ± 1.5697.80 ± 0.3796.42 ± 2.4897.22 ± 1.3798.42 ± 0.76
Shadows82.95 ± 21.6098.34 ± 2.1131.55 ± 10.2652.23 ± 11.7262.74 ± 25.6477.02 ± 8.2952.61 ± 16.47
Parameters141 K216.5 K25.6 M7.3 M92.4 K85.7 K37.7 K
Table 4. Classification accuracy with bolding the best accuracy based on the biased Salinas dataset, run five times.
Table 4. Classification accuracy with bolding the best accuracy based on the biased Salinas dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)90.68 ± 1.2891.16 ± 1.6491.57 ± 0.7492.97 ± 0.8295.42 ± 1.4896.18 ± 1.0896.64 ± 1.48
AA (%)88.18 ± 1.2593.59 ± 2.3592.80 ± 1.0294.94 ± 1.1495.37 ± 2.0596.52 ± 1.2497.32 ± 1.24
K × 10089.56 ± 1.4590.16 ± 1.8290.51 ± 0.9592.11 ± 0.8994.09 ± 1.8595.84 ± 1.5796.23 ± 1.15
Brocoli_green_weeds_181.76 ± 6.6899.59 ± 0.7298.73 ± 0.3695.44 ± 3.5881.69 ± 8.2984.80 ± 8.3594.67 ± 5.33
Brocoli_green_weeds_294.88 ± 6.0799.54 ± 1.3574.36 ± 0.4489.69 ± 1.4597.52 ±0.2897.53 ± 0.4896.95 ± 1.21
Fallow91.35 ± 4.6282.03 ± 20.05100.00 ± 0.0095.11 ± 2.1392.36 ± 1.4094.22 ± 1.3797.68 ± 1.04
Fallow_rough_plow88.31 ± 7.4998.61 ± 1.5795.42 ± 2.59100.00 ± 0.0093.28 ± 1.2796.45 ± 1.2196.82 ± 1.60
Fallow_smooth95.60 ± 4.4097.13 ± 3.9896.76 ± 0.3397.31 ± 0.4297.36 ± 0.5897.62 ± 0.7096.93 ± 1.46
Stubble99.74 ± 0.6399.81 ± 0.4999.36 ± 0.6399.37 ± 0.5699.78 ± 0.1799.82 ±0.0599.87 ± 0.12
Celery97.42 ± 2.0399.62 ± 0.5899.45 ± 0.3599.48 ± 0.2799.47 ± 0.1299.49 ± 0.5998.66 ± 0.49
Grapes_untrained88.48 ± 3.1381.13 ± 8.8696.76 ± 1.4297.16 ± 0.6795.26 ± 1.2196.12 ± 2.4898.03 ± 3.37
Soil_vineyard_develop98.96 ± 1.7098.96 ± 2.3097.39 ± 1.3599.37 ± 0.5499.36 ± 1.1099.38 ± 0.3495.42 ± 3.91
Corn_senesced_green_weeds95.20 ± 5.5791.58 ± 3.7293.82 ± 4.4899.21.00 ± 0.9399.81 ± 0.2599.22 ± 0.6499.86 ± 0.52
Lettuce_romaine_4wk77.33 ± 8.4383.38 ± 28.6798.02 ± 0.8399.86 ± 0.4398.85 ± 0.1299.92 ± 0.1395.89 ± 1.14
Lettuce_romaine_5wk89.66 ± 9.0498.72 ± 3.0296.12 ± 0.6599.45 ± 0.5792.63 ± 0.6499.82 ± 1.0899.92 ± 1.89
Lettuce_romaine_6wk87.92 ± 9.4395.30 ± 5.0978.30 ± 6.3478.81 ± 6.8795.42 ± 0.7296.61 ± 1.4695.42 ± 1.79
Lettuce_romaine_7wk90.53 ± 8.7196.95 ± 3.1597.02 ± 0.8596.50 ± 1.4593.78 ± 1.0394.77 ± 1.1297.44 ± 3.35
Vineyard_untrained81.78 ± 11.2479.72 ± 7.4771.30 ± 10.2974.63 ± 8.2795.62 ± 2.3495.89 ± 1.0196.55 ± 1.56
Vineyard_vertical_trellis61.12 ± 14.3895.37 ± 2.8491.96 ± 2.4897.65 ± 3.8193.79 ± 4.9692.70 ± 5.2397.12 ± 6.67
Parameters193 K370.3 K27.5 M7.6 M101.8 K61.6 K47.1 K
Table 5. Classification accuracy with bolding the best accuracy based on the biased Indian Pines dataset, run five times.
Table 5. Classification accuracy with bolding the best accuracy based on the biased Indian Pines dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)85.29 ± 1.4580.66 ± 2.9181.61 ± 1.2387.77 ± 1.1788.12 ± 1.4289.52 ± 0.9889.32 ± 1.27
AA (%)69.94 ± 4.8365.35 ± 6.8566.71 ± 5.6472.13 ±4.7673.05 ±6.2974.25 ± 4.0474.42 ± 7.26
K × 10082.70 ± 1.7577.83 ± 3.3778.02 ± 1.8485.34 ± 1.4985.87 ± 1.9487.52 ± 1.5987.58 ± 1.53
Alfalfa35.95 ± 35.5212.46 ± 22.2825.39 ± 46.3446.72 ± 41.4736.63 ± 42.2738.63 ± 45.4238.32 ± 39.62
Corn–notill82.49 ± 14.9479.05 ± 10.5987.37 ± 4.2696.21 ± 5.3686.02 ± 4.9389.90 ± 2.2189.95 ± 3.83
Corn–mintill64.88 ± 14.9473.92 ± 16.9762.38 ± 18.9462.86 ± 22.6379.63 ± 19.5479.07 ± 26.9867.46 ± 17.62
Corn45.67 ± 20.5138.96 ± 21.6951.37 ± 26.4349.06 ± 20.5254.19 ± 38.1672.49 ± 18.6468.27 ± 39.48
Grass–pasture58.91 ± 16.7275.46 ± 16.0067.39 ± 20.2358.33 ± 23.8474.63 ± 18.9274.82 ± 24.8568.63 ± 29.94
Grass–trees89.89 ± 8.4894.84 ± 5.2087.57 ± 5.4294.41 ± 4.2694.89 ± 4.3789.63 ± 3.6286.62 ± 7.76
Grass–pasture–mowed48.26 ± 42.0733.70 ± 43.8418.54 ± 43.3634.83 ± 44.3256.38 ±45.2969.73 ± 43.7473.84 ± 56.16
Hay–windrowed92.54 ± 7.2899.53 ± 0.6989.54 ± 4.7392.41 ± 2.6298.05 ± 0.7598.34 ± 1.3796.35 ± 3.64
Oats24.29 ± 31.898.48 ± 13.1335.84 ± 35.7452.63 ± 31.4113.86 ± 26.9519.92 ± 43.8721.92 ± 43.82
Soybean–notill86.89 ± 4.6575.52 ± 8.8480.13 ± 2.6586.35 ± 2.7391.64 ± 3.2390.42 ± 3.2594.87 ± 2.78
Soybean–mintill92.72 ± 1.5782.76 ± 6.9085.75 ± 5.8390.61 ± 2.8492.76 ± 5.5487.50 ± 7.5394.36 ± 1.35
Soybean–clean68.58 ± 16.9168.04 ± 19.1471.36 ± 22.7585.32 ± 19.3172.27 ± 21.9371.89 ± 19.2173.05 ± 16.95
Wheat92.99 ± 13.3695.41 ± 8.3486.73 ± 2.5671.39 ± 3.7594.02 ± 5.8583.21 ± 4.8682.83 ± 8.42
Woods98.89 ± 1.2896.38 ± 3.1593.79 ± 2.6595.36 ± 2.7997.46 ± 1.3897.62 ± 1.7495.49 ± 1.35
Buildings–grass–trees75.44 ± 13.8060.64 ± 19.6064.52 ± 32.2871.16 ± 21.8453.72 ± 31.5754.37 ± 27.4668.42 ± 32.49
Stone–steel–towers57.62 ± 26.5550.40 ± 42.3059.72 ± 32.4866.39 ± 27.4872.68 ± 27.5471.42 ± 16.2970.34 ± 29.67
parameters191.4 K364.2 K27.4 M7.6 M38.4 K53.6 K55.1 K
Table 6. Classification accuracy with bolding the best accuracy based on the biased KSC dataset, fun five times.
Table 6. Classification accuracy with bolding the best accuracy based on the biased KSC dataset, fun five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)95.98 ± 0.8396.77 ± 0.8395.76 ± 0.7295.92 ± 1.5296.04 ± 0.9097.43 ± 1.1397.56 ± 1.69
AA (%)94.07 ± 1.3994.70 ± 0.8695.12 ± 1.5894.24 ± 1.4694.01 ± 1.1996.21 ± 0.6196.41 ± 1.24
K × 10095.53 ± 0.9596.41 ± 0.9395.28 ± 0.6395.46 ± 0.7395.64 ± 1.1097.07 ± 0.7797.28 ± 1.15
Scrub99.72 ± 0.2997.72 ± 2.1699.16 ± 1.0497.89 ± 1.7894.42 ± 1.5493.28 ± 2.3196.22 ± 137
Willow swamp82.18 ± 9.5890.35 ± 9.9595.40 ± 3.6383.12 ± 5.0283.50 ± 5.2686.37 ± 6.1888.64 ± 7.31
Cp hammock92.80 ± 7.6095.88 ± 2.5293.42 ± 2.5790.53 ± 1.4889.84 ± 6.4590.91 ± 5.3594.53 ± 6.34
Slash pine78.38 ± 7.1991.68 ± 5.9795.23 ± 3.2984.94 ± 3.3595.03 ± 3.9196.18 ± 4.6794.84 ± 5.39
Oak/broadleaf91.57 ± 7.5181.88 ± 9.9795.76 ± 4.2783.86 ± 1.5885.27 ± 2.5796.38 ± 1.3987.58 ± 2.58
Hardwood88.02 ± 8.5491.55 ± 8.0297.38 ± 2.5697.96 ± 2.5688.47 ± 4.8496.67 ± 2.1297.55 ± 1.79
Swamp96.64 ± 7.5589.91 ± 10.7689.27 ± 5.73100.00 ± 0.0098.67 ± 1.4698.00±1.23100.00 ± 0.00
Graminoid marsh95.81 ± 5.4296.96 ± 4.3184.37 ± 8.2698.12 ± 2.4997.95 ± 2.6298.31 ± 1.0896.00 ± 4.57
Spartina marsh98.46 ± 3.1988.86 ± 0.7096.23 ± 2.1499.58 ± 1.12100.00 ± 0.0098.23 ± 1.77100.00 ± 0.00
Cattail marsh100.00 ± 0.0098.16 ± 3.7794.96 ± 1.5093.37 ± 1.4892.45 ± 1.2399.65 ± 0.6898.56 ± 2.49
Salt marsh100.00 ± 0.0098.28 ± 2.8396.25 ± 1.7796.50 ± 1.94100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
Mud flats99.33 ± 1.0799.97 ± 1.6399.13 ± 0.9299.25 ± 1.8498.69 ± 1.0999.36 ± 0.6999.44 ± 0.56
Water100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
Parameters178.9 K327.2 K26.7 M7.5 M100.4 K69.4 K60.1 K
Table 7. Average rankings of Friedman test on four datasets in terms of OA, AA and Kappa, bold the best results.
Table 7. Average rankings of Friedman test on four datasets in terms of OA, AA and Kappa, bold the best results.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
Ranking (OA)2.52.51.753.54.756.56.5
Ranking (OA)23.52.53.53.756.256.5
Ranking (Kappa)2.51.751.75456.56.5
Table 8. Classification accuracy with bolding the best accuracy based on the unbiased Pavia dataset, run five times.
Table 8. Classification accuracy with bolding the best accuracy based on the unbiased Pavia dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)83.12 ± 2.6982.46 ± 2.5879.93 ± 0.9382.99 ± 1.3286.05 ± 1.4386.70 ± 1.2687.13 ± 0.76
AA (%)76.35 ± 2.4678.63 ± 3.5973.82 ± 1.5776.25 ± 1.4679.50 ± 1.7881.46 ± 1.3481.98 ± 1.45
K × 10078.42 ± 2.8378.462 ± 3.4873.35 ± 1.8677.33 ± 1.8381.44 ± 1.4781.94 ± 1.3982.92 ± 1.14
Asphalt79.12 ± 2.6983.45 ± 2.3484.12 ± 2.5886.71 ± 1.7893.99 ± 1.9691.31 ± 1.6794.14 ± 2.79
Meadows85.24 ± 1.5393.75 ± 2.7394.35 ± 2.2195.50 ± 1.9395.00 ± 1.3695.74 ±1.3895.70 ± 1.39
Gravel53.14 ± 16.5859.84 ± 18.4380.37 ± 8.3571.29 ± 9.6851.91 ± 17.4350.17 ± 15.7250.63 ± 13.23
Trees83.57 ± 5.1882.69 ± 6.5478.45 ± 11.3585.92 ± 8.6987.17 ± 5.8288.60 ± 5.3589.37 ± 5.24
Painted metal sheets86.95 ± 5.5189.95 ± 2.4174.38 ± 3.1571.76 ± 5.2894.81 ± 3.7899.58 ± 1.0597.34 ± 3.57
Bare soil62.32 ± 3.5869.78 ± 10.7352.50 ± 2.3352.52 ± 2.5666.44 ± 3.2366.10 ± 4.5967.31 ± 5.35
Bitumen63.74 ± 16.7262.32 ± 14.4261.77 ± 1.9770.78 ± 2.5957.68 ± 1.3771.26 ± 1.6465.80 ± 1.46
Self-blocking bricks75.63 ± 7.4471.85 ± 18.5159.82 ± 1.3465.39 ± 1.6872.10 ± 2.3171.73 ± 1.7580.07 ± 2.73
Shadows97.49 ± 19.2798.84 ± 2.2478.62 ± 11.2486.38 ± 11.7296.45 ± 15.7398.73 ± 7.9497.51 ± 8.47
Table 9. Classification accuracy with bolding the best accuracy based on the unbiased Salinas dataset, run five times.
Table 9. Classification accuracy with bolding the best accuracy based on the unbiased Salinas dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)80.36 ± 1.6981.21 ± 1.5883.13 ± 0.9781.99 ± 1.4684.41 ± 1.5388.83 ± 1.3689.75 ± 1.72
AA (%)82.86 ± 1.5784.62 ± 2.5880.37 ± 1.4881.58 ± 1.3987.21 ± 1.8388.90 ± 1.6290.53 ± 1.43
K × 10078.62 ± 1.7280.37 ± 1.4981.20 ± 1.2479.91 ± 1.3482.69 ± 1.6787.54 ± 1.3888.58 ± 1.84
Brocoli_green_weeds_192.46 ± 3.4973.63 ± 4.758.93 ± 3.5738.13 ± 19.4560.22 ± 8.2986.19 ± 8.8497.50 ± 7.39
Brocoli_green_weeds_292.73 ± 6.7391.52 ± 1.5893.03 ± 1.2996.65 ± 1.3898.66 ± 1.5299.55 ± 1.3599.36 ± 1.20
Fallow86.63 ± 4.6882.84 ± 20.5847.17 ± 10.3479.07 ± 2.4787.06 ± 1.3976.72 ± 1.6784.81 ± 1.38
Fallow_rough_plow92.52 ± 7.5791.37 ± 1.8293.26 ± 2.9392.61 ± 1.5897.39 ± 1.9796.65 ± 1.2999.60 ± 1.56
Fallow_smooth87.43 ±4.8392.59 ± 3.6894.28 ± 1.2393.59 ± 1.4894.58 ± 1.7390.78 ± 3.7292.22 ± 2.39
Stubble93.52 ± 0.3893.69 ± 0.6097.85 ± 1.4696.84 ± 1.4499.34 ± 1.2698.85 ± 1.3799.37 ± 1.05
Celery93.34 ± 2.8594.77 ± 0.5597.70 ± 1.2697.52 ± 1.2798.87 ± 1.2499.07 ± 0.9599.82 ± 0.73
Grapes_untrained70.42 ± 3.5873.37 ± 8.3478.54 ± 1.9675.71 ± 1.8465.87 ± 5.3886.60 ± 2.5984.70 ± 4.45
Soil_vineyard_develop95.34 ± 1.7394.74 ± 2.3795.10 ± 1.6896.53 ± 1.7297.01 ± 1.3798.64 ± 1.4697.33 ± 3.80
Corn_senesced_green_weeds72.74 ± 5.7974.64 ± 3.7589.46 ± 2.5787.41 ± 1.2588.36 ± 3.2191.19 ± 2.5289.42 ± 1.26
Lettuce_romaine_4wk68.63 ± 8.7084.73 ± 28.8238.58 ± 2.6734.33 ± 8.3971.34 ± 1.4263.45 ± 3.7377.94 ± 1.58
Lettuce_romaine_5wk82.57 ± 9.5288.62 ± 3.4798.37 ± 1.3698.89 ± 1.6797.03 ± 1.4597.50 ± 1.2888.96 ± 6.96
Lettuce_romaine_6wk84.97 ± 9.2395.30 ± 5.6898.53 ± 6.3498.57 ± 1.5288.77 ± 2.7184.97 ± 3.4988.95 ± 2.02
Lettuce_romaine_7wk82.47 ± 8.7883.89 ± 3.2970.77 ± 1.4679.33 ± 1.5296.68 ± 1.1493.24 ± 1.3879.60 ± 2.13
Vineyard_untrained54.69 ± 11.4457.72 ± 7.4572.91 ± 10.5872.76 ± 2.6274.32 ± 2.3768.77 ± 1.4874.45 ± 1.36
Vineyard_vertical_trellis74.27 ± 14.8379.57 ± 2.8561.53 ± 2.5967.44 ± 4.1979.86 ± 4.8590.24 ± 5.9494.45 ± 6.48
Table 10. Classification accuracy with bolding the best accuracy based on the unbiased Indian Pines dataset, run five times.
Table 10. Classification accuracy with bolding the best accuracy based on the unbiased Indian Pines dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)58.11 ± 1.9461.83 ± 1.6258.46 ± 1.2358.59 ± 1.4668.23 ± 1.7370.61 ± 0.7369.89 ± 1.48
AA (%)45.23 ± 6.9349.24 ± 6.8440.57 ± 6.6447.28 ± 7.3754.63 ± 6.5256.47 ± 5.7360.80 ± 7.42
K × 10052.14 ± 1.4655.25 ± 3.4751.35 ± 1.8452.77 ± 2.7463.55 ± 1.4266.90 ± 1.2766.79 ± 1.73
Alfalfa13.55 ± 35.4512.78 ± 22.6113.95 ± 42.3520.23 ± 43.2814.78 ± 49.5721.74 ± 43.6324.78 ± 39.35
Corn–notill32.47 ± 14.7335.84 ± 10.4834.70 ± 3.7324.93 ± 17.5353.29 ± 4.9350.83 ± 5.5265.20 ± 3.53
Corn–mintill40.54 ± 14.4842.53 ± 16.6826.06 ± 29.2534.55± 13.6340.40 ± 29.5249.75 ± 26.4251.59 ± 17.24
Corn5.41 ± 20.5330.96 ± 21.4115.13 ± 12.2537.39 ± 29.5334.46 ± 38.7453.08 ± 24.6540.17 ± 39.51
Grass–pasture55.74 ± 16.5357.96 ± 16.3831.89 ± 32.4665.19 ± 17.4576.42 ± 18.2768.62 ± 23.5381.86 ± 29.96
Grass-trees88.52 ± 8.9282.94 ± 5.7995.0 6± 3.4397.19 ± 1.3790.04 ± 4.7492.88 ± 3.7492.77 ± 7.82
Grass–pasture–mowed0.00 ± 0.008.92 ± 43.920.00 ± 0.000.00 ± 0.008.58 ± 45.932.84 ± 13.3616.42 ± 56.51
Hay–windrowed89.92 ± 7.7193.74 ± 1.7986.06 ± 1.8596.06 ± 1.9497.88 ± 1.7995.49 ± 1.5290.04 ± 3.84
Oats0.00 ± 0.001.68 ± 13.680.00 ± 0.000.00 ± 0.008.80 ± 36.950.00 ± 0.0022.00 ± 23.2
Soybean–notill49.62 ± 4.4857.05 ± 8.6425.35 ± 4.7536.95 ± 7.4375.41 ± 3.6764.65 ± 4.5265.33 ± 2.83
Soybean–mintill69.62 ± 1.7069.84 ± 6.6878.70 ± 3.8566.44 ± 9.4275.32 ± 5.5281.96 ± 7.9466.91 ± 1.37
Soybean–clean31.31 ± 16.4933.94 ± 19.5424.47 ± 25.3146.13 ± 13.3538.20 ± 21.6437.73 ± 19.4740.45 ± 16.93
Wheat78.51 ± 13.7278.81 ± 8.8954.03 ± 4.4969.40 ± 7.4242.54 ± 5.5275.41 ± 4.6383.32 ± 8.62
Woods88.52 ± 1.6985.06 ± 3.1987.56 ± 2.5785.54 ± 1.8287.88 ± 1.3195.96 ± 1.4292.76 ± 1.73
Buildings–grass–trees23.42 ± 13.6230.85 ± 19.2830.13 ± 23.1731.78 ± 27.4848.92 ± 31.6241.54 ± 27.9654.71 ± 27.63
Stone–steel–towers55.41 ± 26.3764.96 ± 42.3546.03 ± 23.4842.70 ± 21.4773.76 ± 27.2671.12 ± 16.9394.62 ± 22.36
Table 11. Classification accuracy with bolding the best accuracy based on the unbiased KSC dataset, run five times.
Table 11. Classification accuracy with bolding the best accuracy based on the unbiased KSC dataset, run five times.
MethodCNNSSRNResNetDesNetAuto-CNNPSO-NetCPSO-Net
OA (%)68.46 ± 1.5669.52 ± 1.7268.19 ± 1.8373.57 ± 1.7474.41 ± 1.3782.50 ± 1.4783.77 ± 1.74
AA (%)65.46 ± 1.4668.35 ± 1.4759.59 ± 1.9464.10 ± 1.6466.41 ± 1.7674.98 ± 1.8576.55 ± 1.87
K × 10064.31 ± 1.6868.61 ± 1.4164.01 ± 1.8470.65 ± 1.7971.46 ± 1.8680.49 ± 1.2881.96 ± 1.85
Scrub91.72 ± 2.3193.72 ± 2.5769.69 ± 1.6479.61 ± 1.9491.11 ± 1.8498.40 ± 1.4996.54 ± 2.97
Willow swamp68.18 ± 9.3571.28 ± 9.5137.73 ± 3.7433.76 ± 5.6440.41 ± 5.7576.63 ± 7.9680.25 ± 6.24
Cp hammock70.80 ± 7.0472.94 ± 2.7973.86 ± 2.9273.44 ± 2.5365.69 ± 6.7474.04 ± 6.8773.71 ± 5.29
Slash pine23.24 ± 7.5226.90 ± 5.4720.88 ± 10.3832.26 ± 12.4527.62 ± 11.6338.02 ± 15.3756.42 ± 4.73
Oak/broadleaf43.84 ± 7.1546.67 ± 4.7343.69 ± 4.9036.85 ± 24.7463.81 ± 2.4742.86 ± 8.9539.38 ± 11.33
Hardwood22.29 ± 5.4921.72 ± 6.6820.63 ± 7.9223.50 ± 5.6352.40 ± 4.6263.24 ± 1.3652.21 ± 2.73
Swamp62.69 ± 7.6969.27 ± 5.3661.30 ± 3.1768.36 ± 2.6570.29 ± 1.3576.38 ± 3.5262.67 ± 1.48
Graminoid marsh75.48 ± 5.7376.40 ± 4.3875.56 ± 8.2578.14 ± 1.4545.86 ± 12.6556.60 ± 6.4283.26 ± 1.34
Spartina marsh59.91 ± 3.2964.92 ± 1.5258.60 ± 2.4766.45 ± 1.8695.77 ± 1.4298.38 ± 1.6993.42 ± 1.83
Cattail marsh76.15 ± 2.8376.31 ± 3.2771.47 ± 1.8278.01 ± 1.8941.59 ± 7.7370.89 ± 4.8484.00 ± 2.63
Salt marsh81.16 ± 2.3786.27 ± 2.6382.62 ± 1.9487.41 ± 1.2897.42 ± 1.4596.43 ± 1.9598.52 ± 1.85
Mud flats80.82 ± 1.4882.82 ± 1.4758.75 ± 1.8280.73 ± 1.3872.73 ± 1.0984.21 ± 1.7683.82 ± 3.43
Water94.72 ± 3.5299.38 ± 1.5299.89 ± 1.2394.78.00 ± 2.5398.64 ± 1.2498.66 ± 1.9690.95 ± 1.74
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, X.; Zhang, C.; Cai, Z.; Yang, J.; Zhou, Z.; Gong, X. Continuous Particle Swarm Optimization-Based Deep Learning Architecture Search for Hyperspectral Image Classification. Remote Sens. 2021, 13, 1082. https://doi.org/10.3390/rs13061082

AMA Style

Liu X, Zhang C, Cai Z, Yang J, Zhou Z, Gong X. Continuous Particle Swarm Optimization-Based Deep Learning Architecture Search for Hyperspectral Image Classification. Remote Sensing. 2021; 13(6):1082. https://doi.org/10.3390/rs13061082

Chicago/Turabian Style

Liu, Xiaobo, Chaochao Zhang, Zhihua Cai, Jianfeng Yang, Zhilang Zhou, and Xin Gong. 2021. "Continuous Particle Swarm Optimization-Based Deep Learning Architecture Search for Hyperspectral Image Classification" Remote Sensing 13, no. 6: 1082. https://doi.org/10.3390/rs13061082

APA Style

Liu, X., Zhang, C., Cai, Z., Yang, J., Zhou, Z., & Gong, X. (2021). Continuous Particle Swarm Optimization-Based Deep Learning Architecture Search for Hyperspectral Image Classification. Remote Sensing, 13(6), 1082. https://doi.org/10.3390/rs13061082

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop