Next Article in Journal
VHR-BirdPose: Vision Transformer-Based HRNet for Bird Pose Estimation with Attention Mechanism
Next Article in Special Issue
A Multi-Stage Adaptive Copy-Paste Data Augmentation Algorithm Based on Model Training Preferences
Previous Article in Journal
Development of a Hybrid Method for Multi-Stage End-to-End Recognition of Grocery Products in Shelf Images
Previous Article in Special Issue
Gaze Estimation Method Combining Facial Feature Extractor with Pyramid Squeeze Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Spatial–Spectral Attention-Based Neural Architecture Search for Hyperspectral Image Classification

1
Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150080, China
2
Communication Construction Operation and Maintenance Center, State Grid Heilongjiang Electric Power Co., Ltd., Information and Communication Company, Harbin 150010, China
3
Department of Computer Science, Chubu University, Kasugai-shi 487-8501, Aichi, Japan
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(17), 3641; https://doi.org/10.3390/electronics12173641
Submission received: 23 July 2023 / Revised: 24 August 2023 / Accepted: 25 August 2023 / Published: 29 August 2023
(This article belongs to the Special Issue Deep Learning in Image Processing and Pattern Recognition)

Abstract

:
Convolutional neural networks (CNNs) are indeed commonly employed for hyperspectral image classification. However, the architecture of cellular neural networks typically requires manual design and fine-tuning, which can be quite laborious. Fortunately, there have been recent advancements in the field of Neural Architecture Search (NAS) that enable the automatic design of networks. These NAS techniques have significantly improved the accuracy of HSI classification, pushing it to new levels. This article proposes a Multi-Scale Spatial–Spectral Attention-based NAS, MS3ANAS) framework for HSI classification to automatically design a neural network structure for HSI classifiers. First, this paper constructs a multi-scale attention mechanism extended search space, which considers multi-scale filters to reduce parameters while maintaining large-scale receptive field and enhanced multi-scale spectral–spatial feature extraction to increase network sensitivity towards hyperspectral information. Then, we combined the slow–fast learning architecture update paradigm to optimize and iteratively update the architecture vector and effectively improve the model’s generalization ability. Finally, we introduced the Lion optimizer to track only momentum and use symbol operations to calculate updates, thereby reducing memory overhead and effectively reducing training time. The proposed NAS method demonstrates impressive classification performance and effectively improves accuracy across three HSI datasets (University of Pavia, Xuzhou, and WHU-Hi-Hanchuan).

1. Introduction

Hyperspectral remote sensing images (HSIs) capture abundant spatial–spectral information across numerous spectral bands, enabling effective differentiation of surface cover. Therefore, hyperspectral images are widely used in environmental science [1], mineral exploration [2], plant detection [3], military reconnaissance [4], and so on.
HSIs can extract feature vectors containing thousands of bands from each spatial pixel position. In practice, there are two phenomena in HSIs: (1) Different objects have similar spectral characteristics; (2) The same object at different positions has different spectral characteristics [5]. This is due to the influence of the imaging factors of atmosphere and light, which leads to the obvious spectral shift between different scenes.
HSI classification is based on pixel level. In the past decade, various traditional machine learning methods, including the K-nearest neighbor (KNN) algorithm, Support Vector Machine (SVM) [6], and others, have been widely employed for pixel-level HSI classification [7].
Compared with traditional classification algorithms, deep learning-based HSI classification methods have demonstrated their effectiveness in extracting robust features from HSIs, and they have better classification performance. HSI classification methods utilizing CNN can be classified into three categories based on the feature extraction approach employed: spectral CNN [8,9,10], spatial CNN [11,12], and spectral–spatial CNN [13,14,15,16,17]. However, many excellent deep classification network architectures need to be designed manually, such as ResNet [16], DenseNet [18], VGG [19], and GoogLeNet [20].
Designing an efficient network architecture for HSI classification is a challenging task that involves significant time, energy, and a large number of verification experiments, making it difficult to achieve manually. In addition, HSI data has obvious differences in the number of frequency bands, spectral range, and spatial resolution, so the network architectures applied to different HSI data classifications are also different [21]. It is necessary to design different network architectures for HSI classification, which requires a large amount of work to adjust parameters, resulting in a large consumption of time and resource costs. Therefore, a natural idea is to automate the neural network design process and minimize human assistance.
The need to enhance the efficiency of automatic neural network design technology is motivated by both the expense of computing resources and the burden of parameter adjustment [22]. NAS has been applied to many tasks and has achieved remarkable results, such as speech recognition [23], computer vision [24], and so on. The objective of Neural Architecture Search (NAS) is to automate the construction of high-performance neural network structures by selecting and combining various neural operations from predefined search spaces. Previous approaches to NAS have employed various methods, including reinforcement learning algorithm (RL) [25], evolutionary algorithm (EA) [26], and gradient-based methods, to perform architecture searches. The reinforcement learning algorithm regards NAS search as an agent behavior, and network construction is realized through different behaviors and rewards, which can evaluate the performance of the acquired network structure. Representation and optimization of agent strategy are two key contents of network structure search using reinforcement learning [25]. The evolutionary algorithm uses a genetic algorithm to realize selection, crossover, and compilation to initialize the search of individuals and cyberspace [27,28].
Among various NAS methods, DARTS and population-based NAS are among the most popular ones because they have unique advantages in dealing with many challenging tasks. DARTS mainly benefits from the advantage of relaxing the search space into a continuous and high search efficiency. The population-based NAS mainly benefits from the advantages of diverse candidate structures within the population and involves genetic operators to drive the search process. As the term itself indicates, the population-based NAS represents a candidate architecture for each individual. Collaboration between candidate architectures can eliminate poor and well-preserved competition and push the overall optimization. At present, most population-based NAS methods use genetic algorithms or genetic programming [25,29] to simulate natural evolution processes, which requires careful design of random crossover mutation operators. For example, the population-based NAS method [30] has designed 11 mutation operators to modify the attributes of the network, including changing the learning rate, resetting weights, inserting convolutions, etc. This study shows that designing different genetic operators has good feasibility, and neural networks can be developed and changed from an architectural perspective. Although the population-based NAS has good performance, many current methods mainly have the drawback of low computational efficiency.
Whether based on reinforcement learning or population algorithm, NAS realizes automation through resource-consuming search. To minimize resource usage, one-shot NAS techniques using hypernetworks have been devised [31,32]. DARTS, a one-shot NAS approach employing a discernible search strategy [32], incorporates weight sharing to integrate hypernetwork training and the search for the optimal candidate architecture. This integration effectively curtails computing resource wastage. At the same time, the gradient-based method is used to optimize the over-parameterized hypernetwork. However, the greater number of weight parameters in DARTS compared to architecture parameters imposes constraints on architecture optimization, leading to performance crashes [31]. Essentially, this is because DARTS’ candidate results lack diversity in gradient optimization.
However, these NAS methods have their own characteristics that DARTS jointly trains the hypernetwork and only searches for the optimal solution through the gradient, so they have defects in flexibility and stability [33]. Many population-based NASs mainly rely on random crossover/mutation search, which usually requires a lot of computational cost to evaluate the performance. Therefore, this work adopts the slow–fast learning paradigm architecture update process in DARTS, which integrates the idea of population-based NAS. In addition, it can benefit from the advantages of differentiated NAS while overcoming the shortcomings of the population-based NAS [34].
NAS technology can efficiently automate this process, not only to find the weight of specific image classification tasks but also to obtain the best network architecture. The incorporation of NAS offers an excellent solution for HSI data classification, relieving individuals from the burdensome task of network architecture design. Chen et al. pioneered the integration of the DARTS method into the HSI classification task, presenting their proposed approach as 3D-Auto-CNN with Cutout (CNAS) [35] and leveraging point-by-point convolution to reduce the spectral dimensions of HSI to several dozens. Subsequently, DARTS is utilized to search for a neural network architecture that is well-suited for the HSI dataset. This approach aims to minimize redundancy and repetition.
In recent years, the rise of attention mechanisms has brought new research directions to deep learning. Many researchers use different scales of attention mechanisms to extract the effective features of different objects in HSI data and classify them effectively. To reduce repetition, there have been several approaches proposed in the literature for HSI classification. Wang introduced the Squeeze and Excitation (SE) [36] module, which adaptively learns the weights of different spectral bands and adjacent pixels in HSI. This module helps capture relevant information for classification. Roy et al. [37] introduced a novel method named A2S2K-ResNet (Attention-based Adaptive Spectral Spatial Kernel Improved Residual Network). This approach integrates spectral attention to effectively capture discriminative features for HSI classification. These methods contribute to reducing redundancy and improving the efficiency of HSI classification tasks. However, in the field of HSI classification for NAS applications, there is limited search space utilizing multi-scale attention mechanisms. In addition, various samples in the HSI dataset exhibit long- tailed distributions, resulting in imbalanced HSI classification results. In light of the aforementioned approaches, we propose a novel method that utilizes a multi-scale search space combined with multi-scale spatial–spectral attention. This approach aims to enhance the extraction of significant spectral–spatial information while suppressing redundant information and noise. By incorporating this multi-scale attention mechanism, our method effectively improves the accuracy of classification while reducing computational complexity. Furthermore, it helps to minimize repetition and redundancy in the HSI classification process.
Building upon the aforementioned discussion, we present a novel NAS method that encompasses the automatic design of a search space for multi-scale attention mechanisms and a search strategy based on the slow–fast paradigm. The key contributions of this research can be summarized as follows:
  • To address the issue of redundancy, we have proposed a highly effective NAS classification framework called Multi-Scale Spatial–Spectral Attention-based NAS (MS3ANAS). By carefully analyzing the characteristics of HSI, we have designed a multi-scale search space that incorporates rich spatial–spectral attention mechanisms. This search space consists of seven convolutional operators, each equipped with attention mechanisms of different scales. The search process can automatically learn to add attention modules to appropriate locations in the architecture to fully explore the spectral and spatial information of HSIs for classification.
  • The slow–fast learning paradigm was quoted for NAS, which further optimizes the overall architecture vector through pseudo gradient iteration, stabilizing the optimal search process, which improves the update and rate of convergence speed of the architecture for HSI classification.
  • To better train the HSI classification model, the Lion optimizer has been introduced, which not only effectively improves the accuracy of HSI data classification but also reduces memory overhead and training time because the algorithm only tracks momentum and uses symbol operations to calculate updates.

2. Materials and Methods

2.1. Search Framework

Figure 1 illustrates the proposed NAS framework for HSI classification in this paper. Initially, the hyperspectral image is divided into patches using a sliding window approach. These patches are then fed into the automatic architecture search model, which consists of both normal cells and reduction cells. Secondly, the NAS network includes supernet architecture search and final network optimization. We designed a multi-scale attention mechanism search operator for the search space (The different colored arrows represent different candidate operations), enhancing the ability to extract multi-scale spectral–spatial features from HSI data, thereby improving classification performance. Then, we utilized the slow–fast learning paradigm to effectively update the overall architecture vector, improving the efficiency of constructing hypergrid architecture units. Finally, we introduced the Lion optimizer to accelerate model convergence and alleviate model memory overhead.

2.2. Neural Architecture Search Algorithm

MS3ANAS consists of three architecture search components: the search space of the multi-scale expanding attention mechanism, the search strategy of the slow–fast learning paradigm, and the Lion optimizer. Here is a detailed explanation of each component.

2.2.1. Multi-Scale Attention Mechanism Expanded Search Space

  • Multiple attention mechanism guided search space
Applying different attention mechanisms to different datasets may face varying challenges. For example, in some HIS datasets, it may be more challenging to use spatial attention mechanisms due to strong correlations between data. In contrast, it may be more difficult to use channel attention mechanisms due to highly variable features. Moreover, the performance of attention mechanisms can also be affected by factors such as dataset size and quality, model architecture, and hyperparameters. Therefore, it is important to choose appropriate attention mechanisms based on different HSI datasets and tasks to achieve optimal performance and effectiveness. In addition, various samples in the HSI dataset exhibit long-tailed distributions, resulting in imbalanced HSI classification results. We propose a multiple search space with rich attention for HSI classification, which can effectively improve classification accuracy and reduce computational complexity. We select four types of attention mechanisms to form it. They are the convolutional block attention module (CBAM) [38], squeeze-and-excitation (SE) module [36], triplet attention (TA) module [39], and coordinate attention mechanism (CA) [40], as shown in Figure 2.
To reduce repetition, various attention mechanisms have been proposed in the literature. One such mechanism is the CBAM module, which combines channel attention and spatial attention. It aggregates spatial features by performing max-pooling on top of global average pooling. This allows the network to focus on important spatial information. Another attention mechanism is the SE module, which enhances the network’s sensitivity to informative features. It recalibrates filter responses through squeeze-and-excitation operations, improving the learning ability of convolutional layers. The TA module is composed of three parallel branches, each serving a distinct purpose. The first branch focuses on establishing spatial attention, while the other two branches aim to capture cross-dimensional interactions between the channel and spatial dimensions. The final output is obtained by averaging the outputs from these three branches. The CA mechanism learns weights by combining features at each position with their corresponding coordinate information. This mechanism effectively captures spatial correlations, leading to improved model performance. By incorporating these attention mechanisms, models can effectively reduce redundancy and improve the performance of HSI classification tasks.
As NAS-based HSI classification requires the determination of the search space, we propose search space O . Formally, let O denote a sequence of candidate operations (multiple attention guided operations), where each operation represents a function p ( · ) to be applied to s i . For each cell q , we configure an architecture parameter α p i q for operation p i . To ensure continuity of the search space, we relax the search space to allow it to be optimized via gradient descent. Specifically, we relax the architecture parameter α p i q to be continuous, and then compute the operational probabilities for different operations by applying softmax over all α p i q .
R p i q = exp ( α p i q ) j = 1 n exp ( α p j q )
Here, n represents the number of candidate operations available. The larger value of R p i q indicates a higher likelihood of selecting the representative operation. The output of the cell is obtained by taking the weighted sum of all possible operations.
p ¯ ( y i ) = p P R p i q p i ( y i )
The notation p i ( y i ) signifies the application of operation p i on input y i . Consequently, the search process is transformed into a learning process of a set of architectural parameters { α p i q }. Furthermore, as the network weights w also need to be learned, we are required to solve the following bi-level optimization problem.
min α v a l ( w * , α )
s . t .   w * = argmin t r a i n ( w , α )
The purpose of the above formulae is to search for the architectural parameters that minimize the validation loss v a l ( w * , α ) , and the network weights w * are obtained by minimizing the training loss t r a i n ( w , α ) . It is important to note that the training loss and the validation loss are identical.
The objective of the aforementioned formulae is to search for architectural parameters that minimize the validation loss v a l ( w * , α ) , while the network weights w * are obtained by minimizing the training loss t r a i n ( w , α ) . It is crucial to emphasize that the training loss and the validation loss are indeed the same.
2.
Multi-scale attention mechanism search space
Different types of convolutions have varying computational requirements and parameter counts. Building deeper models can be challenging due to the expensive parameters and time required for high-dimensional convolutions. This can limit the efficiency and feasibility of constructing deeper models. Therefore, high-dimensional convolutions are often replaced with lower-dimensional separable convolutions to alleviate this issue. The novel operation MCS_sepConv_(b × b) (b = 3, 5, 7) in the search space combines spectral–spatial CBAM with spatial separable convolution [41], which helps to extract deeper spectral–spatial features, and it can enhance the spectral–spatial adaptive learning ability of the I data. In addition, we use small filters to reduce parameters while maintaining a large-scale receptive field. We have strengthened the ability to extract scale spectral–spatial features from data, thereby improving classification performance. MCS_sepConv(b = 3, 5, 7) with different scales are as shown in Figure 3.
MCS_sepConv_(b × b) refers to a set of convolutional operators that are commonly used in remote sensing and other image-processing tasks to extract deeper spectral–spatial features. These operators consist of two separate convolutional layers, one for processing the spectral dimension and the other for processing the spatial dimension, and they are applied in a cascaded manner. By using MCS_sepConv_(b × b), the model can capture more complex and abstract features by combining spectral and spatial information.
Convolution is a widely used operation in NAS-based methodIor HSI classification. Previous research has primarily concentrated on traditional convolution techniques, depthwise separable convolution, and dilated convolution in order to reduce redundancy. However, the high-dimensional convolutional operations can lead to a large number of parameters and time consumption, making it very difficult to construct an optimal architecture. For example, depthwise separable convolution can be represented as Sep-Conv(b × b) = Conv(1 × 1)(Conv(b × b)(x)), and the parameter and Flops(floating point operations per second) calculations are:
P 1 = C i n × b × b + 1 × C i n × C o u t
F 1 = C i n × b × b × H × W + 1 × 1 × C i n × C o u t × H × W
where C i n × H × W is the size of the input x, and C o u t × H × W corresponds to the output features, where C i n and C o u t represent the input and output spectral bands, respectively. Additionally, P1 and F1 represent the number of parameters and Flops. Incorporating a pointwise convolution following a depthwise convolution in the CNN can effectively extract both spatial and spectral features in a sequential manner. Assuming C i n = C o u t = C , the number of parameters and Flops for separable convolutions is only ( 1 / b 2 + 1 / C ) of regular convolutions. If we extend the concept of separable convolution to the spatial dimensions, we can define spatial separable convolution as Spatial_SepConv(b × b) = Conv(1 × 1Conv(1 × b)(Conv(b × 1)(x)). The number of parameters and Flops for a spatial separable convolution is only ( 2 b + C ) / ( C + b 2 ) of separable convolution:
P 4 = 2 × ( C i n × b × 1 ) + 1 × 1 × C i n × C o u t
F 4 = 2 × ( C i n × b × 1 × H × W )                 + 1 × 1 × C i n × C o u t × H × W
To reduce redundancy in a convolutional neural network (CNN), one approach is to use small filters instead of large-scale filters. This helps to decrease the number of parameters while still maintaining a large receptive field. This is because, in the CNN, the number of parameters and computations is directly related to the size of the input data and the convolution kernel. By using smaller filters, the number of parameters and computations can be significantly reduced. By using small filters, we can reduce the number of parameters while maintaining a large receptive field, which is important for capturing multi-scale features in HSI classification. This can be achieved by using a combination of small filters with different kernel sizes, which allows us to capture multi-scale features while reducing the number of parameters and computations.
To enhance the adaptive feature extraction capability of the designed convolution operation, a lightweight attention module called CBAM (Multi-scale Channel–Spatial Attention, MCS) is incorporated to enhance the spectral and spatial adaptive learning ability of the data cube. MCS is a combination of the Multi-scale Channel Attention (MS) mechanism and the Multi-scale Spatial Attention (MS) mechanism as follows:
M c ( F ) = σ ( M L P ( A v g P o o l ( F ) + M L P ( M a x P o o l ( F ) ) = ( W 1 ( W 0 ( F a v g C ) ) + W 1 ( W 0 ( F m a x C ) ) ) F = M c ( F ) F
M s ( F ) = σ ( f 7 × 7 ( [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] ) ) = ( f 7 × 7 [ F a v g S ; F m a x S ] ) F = M s ( F ) F
where M c s ( F ) = M s ( M c ( F ) ) represents the spectral–spatial CBAM, which involves element-wise multiplication , W 0 C i n / r × C i n , W 1 C i n × C i n / r , sigmoid activation σ , and the 7 × 7 convolutional f 7 × 7 .
As shown in Table 1, Conv(5 × 5) and Conv(7 × 7) can be achieved by repeating the Conv(3 × 3) operations for substitution. In addition, Conv(3 × 3) can be replaced with Spatial sepconv(3 × 3) = (Conv(1 × 3)Conv(3 × 1)). Assuming C i n = C o u t = C , when using Spatial sepconv(3 × 3) for equivalent replacement of Conv(5 × 5), the parameter count is reduced by about 25%. After replacing Conv(7 × 7), the parameters were reduced by about 33%. The amount of Flops also showed a significant decrease. It can be explained that we use small filters instead of large ones to effectively reduce parameters while maintaining a large receptive field, which has obvious effectiveness.

2.2.2. Search Strategy Based on the Slow–Fast Learning Paradigm

In order to further utilize the search space guided by multiple attention mechanisms, we apply the slow–fast learning paradigm to optimize and iteratively update the architecture vector. In addition, the HSI classification involves multiple categories, and the sample size in each category presents the long-tailed distribution, resulting in imbalanced classification results for certain categories in classification tasks. By using the slow–fast paradigm to update the architecture vector, the overall architecture vector can be updated from the perspective of pseudo gradients [42]. Different hyperspectral imbalanced data can be updated during the construction of the architecture unit to obtain the best results and effectively enhance the generalization of the model. This design is essentially aimed at designing an effective NAS method that benefits from the high efficiency of differentiable NAS and overcomes the drawbacks of high search costs in population-based NAS.
This section optimizes from the perspective of NAS search strategy and effectively updates the population of architecture vectors through pseudo gradient iteration to achieve optimal results. The objective of NAS is to initially search for architecture vectors α * O . Then, the architecture parameter C α * is used to minimize the validation loss v a l ( C α * , w α * ) , and finally, the weight w * , which is associated with the architecture, is obtained by minimizing the training loss t r a i n ( C α * , w α * ) . Therefore, the slow–fast learning is essentially the optimizer for α parameters in our NAS. Specifically, we give the architecture vector α t obtained by slow–fast learning in the tth generation, and its iterative update method is shown in Figure 4.
α t + 1 = α t + Δ α t
where the above equation satisfies v a l ( C α g + 1 , w α g + 1 * ) < v a l ( C α g , w α g * ) .
To generate pseudo gradient Δ α t effectively, this approach suggests utilizing a population of H architecture vectors { α } h = 1 H . Each generation randomly splits the population into H/2 pairs. Next, for each pair of population q, the order of validation loss values is employed to identify the fast learner α q , f t and slow learner α q , s t , with smaller loss values indicating slow learner and larger loss values indicating fast learner [ v a l ( C α q , f t , w α q , f t * ) < v a l ( C α q , s t , w α q , s t * ) ]. Then, α q , s t learns from α q , f t and updates it.
Δ α q , s t = η 1 ( α q , f t α q , s t ) + η 2 Δ α q , s t 1
Here, η 1 , η 2 [ 0 , 1 ] represents a randomly generated value obtained from a uniform distribution. Specifically, η 1 determines the step size for α q , s t to learn from α q , f t , and η 2 determines the influence of momentum Δ α q , s t 1 . Finally, all fast–slow learners are aggregated to form a new group of the t + 1th generation. The above pseudo gradient update method is derived from the second derivative of gradient descent in back propagation, so each architecture vector in the search space of the multi-attention mechanism designed in this work will move towards the optimal update direction from the vector that converges faster than them.
After the architecture update process, it is necessary to evaluate the performance of candidate architectures for decoding architecture vectors, so that for each pair of architecture vectors, fast learners can learn and update slow learners by verifying the loss. Subsequently, the candidate architecture C α g is assessed by solving the following optimization problems:
w α g * = optimize ( w α g ) = s t e p ( s t e p ( s t e p ( w α g | C α g ) | C α g ) | C α g )
In the given equation, w a g * represents the optimal weight of the candidate architecture, while s t e p ( ) denotes the iterative optimization process utilized for updating the weights of the neural network.

2.2.3. Lion Optimizer

The optimizer is utilized to update and compute the network parameters that influence model training and output, with the goal of approaching or achieving the optimal value and thereby minimizing (or maximizing) the loss function. This study utilizes the Lion optimizer, known for its simplicity, efficiency, and speed. Reducing redundancy is an important aspect of this work. At the same time, this optimizer is suitable for large-scale optimization problems involving big datasets or high-dimensional parameter spaces. The use of the Lion optimizer can not only effectively improve the accuracy of HSI data classification but also reduce memory overhead and effectively reduce training time because the algorithm only tracks momentum and uses symbolic operation to calculate updates. The Lion optimizer formula is written as follows [43]:
θ t + 1 = ω t θ t
g t = θ f ( θ t 1 )
where ω t is the decay rate, θ t is the weights, and g t = θ f ( θ t 1 ) is the gradient at θ t 1 .
To calculate the average attenuation of the current and past square gradient, the mathematical relationship can be written as follows:
c t = β 1 m t 1 + ( 1 β 1 ) g t
where c t is the 1st moment estimate, and m t 1 represents the momentum vector from the previous iteration. β 1 is the decay rate of the 1st moment. The weight reduction process for decoupling is as follows:
θ t θ t 1 η t ( s i g n ( c t ) + λ θ t 1 )
where η t is the step size. To counteract biases, the bias-corrected first and second moments are computed in the following manner:
m t β 2 m t 1 + ( 1 β 2 ) g t
The algorithm differs from various adaptive algorithms in that it only tracks momentum and uses symbolic operations to compute updates, leading to lower memory overhead and achieving a unified update magnitude across all dimensions.

3. Results

3.1. Hyperspectral Image Datasets

To assess the efficacy of the proposed NAS method, this research conducts classification experiments on three datasets: one standard hyperspectral dataset (Pavia University) and two practical datasets (Xuzhou and WHU-Hi-Hanchuan). Table 2, Table 3 and Table 4 show the number of pixels in each category, the false color composite image, and the ground truth map of the three datasets, respectively. The aim is to reduce redundancy in the experiments and ensure a comprehensive evaluation. The Pavia University dataset was gathered using AROSIS sensors that flew over Pavia in northern Italy. It consists of 103 spectral bands and has a dataset size of 610 × 340 pixels. It consists of nine categories. The Xuzhou dataset was collected in November 2014 using the HySpex SWIR-384 and HySpex VNIR-1600 imaging spectrometers. The dataset consists of 436 bands and covers a specific pixel size in the mining area. A field survey was conducted, and a total of nine categories were calibrated. In total, 68,877 marker samples were used.
The WHU-Hi-Hanchuan dataset was collected in Hanchuan, Hubei Province, using the 17 mm focus Headwall Nano Hyperspec sensor installed on the Leica Airbot X6 UAV V1 platform. The size of the dataset is 1217 × 303 pixels, with 16 categories. The survey collected data from 68,877 marker samples.

3.2. Implementation Details

Our experiments are conducted at Intel (R) Xeon (R) 4208 [email protected] Processor and Nvidia GeForce RTX 2080Ti graphics card. We conducted 10 experiments to take the average value to obtain the overall accuracy (OA), average accuracy (AA), and Kappa coefficient (K) of the experiment.
The hyperspectral image samples are extracted using the sliding window strategy with a window size and overlap rate set at 50%. For training purposes, 30 samples are randomly selected as the training dataset, while 20 samples are used for validation. The training dataset is utilized to train the weights and biases of each neuron in the model, whereas the architecture variables are optimized based on the validation dataset. Once the optimal architecture is obtained, the remaining samples are employed as the test dataset to evaluate the performance of the optimized network architecture and derive the final classification results.
To evaluate the discovered architectures, we conducted evaluations on the discovered architectures with two cells and sixteen initial channels. Each cell consisted of seven nodes, including two input nodes, four intermediate nodes, and one output node. The population size (N) and generation number were set to 1, 20, 35, and 50, respectively. The batch size for each search stage was 32, and we used Stochastic Gradient Descent (SGD) to optimize the weight parameters with a learning rate of 0.025. The cosine power annealing strategy was employed to reduce the learning rate, with a maximum learning rate of 0.025, a minimum learning rate of 0.0001, and a power curve parameter of 2. To prevent overfitting and ensure the robustness of the final model, we also used label smoothing regularization with a smoothing factor of 0.1.

3.3. Classification Comparison with State-of-the-Art Methods

In this section, we evaluate the classification performance of MS3ANAS by comparing it with several advanced methods. These methods include Extended Morphological Profile combined with Support Vector Machine (EMP-SVM) [44], 2D-CNN [11], 3D-CNN [14], Spectral–Spatial Residual Network (SSRN) [16], Residual Network (ResNet) [45], Multi-Layer Perceptron Mixer (MLP Mixer) [46], CNN model designed using NAS (CNAS) [34], and Efficient Convolutional Neural Architecture Search for LiDAR DSM Classification (AN-AS-CPA-LS) [47]. To ensure the rigor of the experiment, we randomly selected three training samples for classes with fewer than three samples. The training times for manually designed CNN models were set to 200. For CNAS and ANAS-CPA-LS, they searched for the optimal architecture based on the DARTS strategy. Therefore, their configuration and superparameter settings were the same as those of the MS3ANAS model. All experimental results are shown in Table 5, Table 6 and Table 7, and they represent the average values of ten runs with different random initializations.
As shown in Table 5, Table 6 and Table 7, our proposed MS3ANAS typically outperforms other methods in terms of OA, AA, and Kappa on three datasets. The EMP-SVM has the lowest classification accuracy, and the classification accuracy of 2D-CNN and 3D-CNN is also poor, while SSRN uses skip connections to extract depth feature information, resulting in higher classification accuracy than 2D-CNN or 3D-CNN. This is due to the lightweight spatial–spectral attention operations found in multi-scale attention search spaces that can better extract joint spectral–spatial features of hyperspectral images. Taking the Pavia dataset as an example, compared to CNAS, the MS3ANAS model proposed in this paper improves OA from 97.42% to 99.16%. Both ANAS-CPA-LS and MS3ANAS have built the search space of attention mechanism, but the search space built by our method contains attention mechanisms of multiple scales. At the same time, we use small filters to reduce parameters while maintaining a large range of receptive fields. Therefore, the ability to extract different scales of spectral–spatial features from data is enhanced, thereby improving classification performance. However, in the case of limited training samples, the ability of attention mechanisms to fit spatial structural information is not easily utilized. Compared with ANAS-CPA-LS, MS3ANAS has better classification performance.
As shown in Table 6, compared with EMP-SVM, 2D-CNN, 3D-CNN, SSRN, ResNet, CNAS, and ANAS-CPA-LS, OA obtained by our proposed method increased by 13.94%, 10.12%, 6.73%, 6.03%, 3.90%, 1.37%, and 0.72%, respectively, on the Xuzhou dataset. Taking the WHU-Hi-Hanchuan dataset as an example, OA reached 98.30%, and increased by 0.86%, 2.15%, 4.35%, 5.52%, 7.29%, 9.23%, and 10.34%, respectively, compared with ANAS-CPA-LS, CNAS, ResNet, SSRN, 3D-CNN, 2D-CNN, and EMP-SVM. Except for EMP-SVM, 2D-CNN is the second worst in performance because it only uses spatial features. Meanwhile, for other methods, MS3ANAS achieved significant improvements in both AA and Kappa results. From the classification results of the three datasets, it can be seen that the overall performance of the NAS-based method is higher than that of the handmade methods. At the same time, the search space using a multi-scale attention mechanism can enhance the model’s ability to consider spectral and spatial information, effectively improving classification performance.
Due to the design of a lightweight NAS model in this article, we compared its complexity with existing technologies in terms of required model parameters (Params) and Flops. We can see from Table 5, Table 6 and Table 7 that different search space operations and architecture update methods can lead to differences in the amount of network parameters. Flops are determined by the input data size and parameters; therefore, we will conduct an overall analysis based on the complexity and classification performance of the network model. Compared with handmade hyperspectral image classification models such as 2D-CNN and 3D-CNN, as well as NAS methods such as CNAS and ANAS-CPA-LS, the proposed method has lower occupancy rates in Params and Flops but achieves the highest OA while achieving the best classification performance on three datasets. The results further indicate that the proposed method can solve classification tasks with low cost but high performance. This mainly benefits from the effectiveness of our special multi-scale search space and the applicability of the slow–fast learning architecture update paradigm to different hyperspectral datasets.
The classification diagrams of the other seven methods on the three HSI datasets are displayed in Figure 5, Figure 6 and Figure 7. It is evident that the proposed algorithm demonstrates superior performance in general. Compared with ANAS-CPA-LS, the ANAS-CPA-LS incorrectly classified Bare Soil (Class 6) as Meadows (Class 2) in Figure 5h of the Pavia dataset classification results.
In Figure 6h of the experimental results on the Xuzhou dataset, some Coals (level 2) were mistakenly classified as Cement (level 7). Through a comparison of the classification maps obtained, it can be concluded that our method achieves more accurate classification results with reduced salt-and-pepper noise. Because our method uses the combination of multi-scale attention mechanism and common attention mechanism as the search operator, it can more effectively extract deeper spatial features and combine spectral and spatial information to capture more complex and abstract features.

4. Discussion

4.1. Optimal Architecture Analysis

Figure 8, Figure 9 and Figure 10 illustrate the cells with the optimal architecture on the three datasets, which are obtained through the proposed method. These cells consist of both normal cells and reduction cells.
In our experiment, we employed the slow–fast learning paradigms and the multi-scale attention mechanism search space to search for the optimal network architecture for each HSI dataset. We can observe that the network architecture retains more multi-scale search operators (such as MCS_ sepConv (3 × 3), MCS_ SepConv (5 × 5), and MCS_ SepConv (7 × 7)). As a result, our model comprises a higher number of learnable parameters, allowing it to maintain a wide range of receptive fields. This effectively addresses the issues of gradient disappearance and network degradation. Consequently, our method can extract scaled spectral–spatial features from the data more effectively, leading to an improvement in classification performance.

4.2. The Analysis of Slow and Fast Learning Using Optimal Combination Actions

In order to empirically analyze the slow–fast learning process, it is necessary to select different parameters for experiments. In this section, the generation is combined with different search spaces to conduct experiments on three datasets. In the experiment, the control variable method is used for each dataset. The number of experiments, training samples, validation samples, and test samples is consistent. The conclusion drawn is that the (a) Pavia University, (b) Xuzhou, and (c) WHU-Hi-Hanchuan datasets display the complete presentation of each combination, and all datasets display MA + MS and generation 50 as the best performance combination (Figure 11). Therefore, it can be proven that slow–fast learning can perform pseudo gradient architecture updates on different hyperspectral imbalanced data during the construction process of building units, effectively improving classification performance.

4.3. Ablation Experiments Analysis

To validate the effectiveness of our method in HSI data classification, we conducted numerous ablation experiments, as presented in Table 8. Specifically, when we solely employ the CNAS model for classifying the Pavia dataset with only 30 training samples, OA reached 97.42%, 97.07%, and 96.15%, respectively. When CNAS was combined with multi-scale attention expansion mechanism search space (MS), OA increased by 0.48%, 0.77%, and 0.89%, respectively. This result demonstrates that the incorporation of a multi-scale attention mechanism to expand the search space enhances the network’s sensitivity to informative features, resulting in a slight improvement in the classification performance of the model. In comparison to CNAS, our method achieves better results. the search time of MS + NAS on the three datasets does not change significantly, but it significantly reduces the parameters and achieves better classification accuracy. Obviously, when NAS was combined with the slow–fast learning paradigm search strategy, the OA of the three datasets increased by 1.02%, 0.39%, and 0.89%, respectively. This is due to the application of slow and fast learning paradigms to optimize and interactively update the architecture vector. It can update the architecture of different hyperspectral imbalanced data and obtain classification models with stronger generalization ability. At the same time, it can be observed that the search time for the three datasets was reduced by 0.249 h, 0.111 h, and 0.087 h after SF + MSNAS referenced the Lion optimizer. This is because this optimizer can effectively reduce the time consumption of NAS architecture search. In addition, the algorithm only tracks momentum, effectively reducing memory overhead.

5. Conclusions

In this paper, we introduce a neural network architecture search algorithm called MS3ANAS. This approach aims to address the limitations of conventional differentiable NAS methods in three key areas: search space, search strategy, and architecture resource optimization. Firstly, we introduce an extended search space with a multi-scale attention mechanism, which can not only enhance the robustness and receptive field capture ability of the model but also better realize path optimization. Secondly, we quoted the slow–fast architecture learning paradigm, which not only optimizes the iterative updating of architecture vectors but also effectively improves the model’s generalization ability. Finally, the Lion optimizer was introduced to improve the accuracy of HSI data classification, reduce memory overhead, and reduce computational complexity. We conducted experiments on three HSI datasets and compared MS3ANAS with seven methods. The experimental results show that our method is more competitive. Through the datasets of Xuzhou, Pavia University, and WHU-Hi-Hanchuan, the OA of our method reached 97.82%, 98.46%, and 98.41%, respectively.
In future work, we will further optimize the NAS structure continuously to make it more lightweight and easier to adapt to more complex remote sensing image classification tasks. Furthermore, it is important to consider the practical applicability of our proposed method. Meanwhile, we plan to apply our approach to real-world hyperspectral image classification tasks, such as environmental monitoring, agricultural assessment, and urban planning. By validating the performance of our method on these practical applications, we can demonstrate its effectiveness and potential impact in various domains. Moreover, the potential applications of our method extend beyond hyperspectral image classification. By exploring these new application areas, we can further broaden the scope of our research and contribute to advancements in related fields. We are excited about the opportunities that lie ahead and look forward to witnessing the practical implementation and impact of our work in the field of remote sensing and computer vision.

Author Contributions

Conceptualization, Y.S., Y.Z., Y.I., H.W. and A.W.; methodology, Y.S. and Y.Z.; software, Y.S.; validation, Y.S.; writing—review and editing, Y.Z., Y.I., H.W. and A.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the High-end Foreign Experts Introduction Program (G2022012010L), Heilongjiang Natural Science Foundation Project (LH2023F034), and Reserved Leaders of Heilongjiang Provincial Leading Talent Echelon (2021).

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Transon, J.; d’Andrimont, R.; Maugnard, A.; Defourny, P. Survey ofhyperspectral earth observation applications from space in the sentinel-2 context. Remote Sens. 2018, 10, 157. [Google Scholar] [CrossRef]
  2. Carrino, T.A.; Crósta, A.P.; Toledo, C.L.B.; Silva, A.M. Hyper-spectral remote sensing applied to mineral exploration in southern peru:A multiple data integration approach in the chapi chiara gold prospect. Int. J. Appl. Earth Obs. Geoinf. 2018, 64, 287–300. [Google Scholar]
  3. Behmann, J.; Steinrücken, J.; Plümer, L. Detection of early plant stress responses in hyperspectral images. ISPRS J. Photogramm. Remote Sens. 2014, 93, 98–111. [Google Scholar] [CrossRef]
  4. Shimoni, M.; Haelterman, R.; Perneel, C. Hypersectral imaging for military and security applications: Combining myriad processing and sensing techniques. IEEE Geosci. Remote Sens. Mag. 2019, 7, 101–117. [Google Scholar] [CrossRef]
  5. Sun, W.; Peng, J.; Yang, G.; Du, Q. Fast and latent low-ranksubspace clustering for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3906–3915. [Google Scholar] [CrossRef]
  6. Samadzadegan, F.; Hasani, H.; Schenk, T. Simultaneous feature selection and SVM parameter determination in classification of hyperspectral imagery using Ant Colony Optimization. Can. J. Remote Sens. 2012, 38, 139–156. [Google Scholar] [CrossRef]
  7. Chandra, B.; Sharma, R.K. On improving recurrent neural network for image classification. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1904–1907. [Google Scholar] [CrossRef]
  8. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutionalneural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  9. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  10. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2016, 55, 844–853. [Google Scholar] [CrossRef]
  11. Yue, J.; Zhao, W.; Mao, S.; Liu, H. Spectral–spatial classification of hyperspectral images using deep convolutional neural networks. Remote Sens. Lett. 2015, 6, 468–477. [Google Scholar] [CrossRef]
  12. Zhang, H.; Li, Y.; Zhang, Y.; Shen, Q. Spectral-spatial classification of hyperspectral imagery using a dual-channel convolutional neural network. Remote Sens. Lett. 2017, 8, 438–447. [Google Scholar] [CrossRef]
  13. Wang, B.; Shao, Q.; Song, D.; Li, Z.; Tang, Y.; Yang, C.; Wang, M. A Spectral-Spatial Features Integrated Network for Hyperspectral Detection of Marine Oil Spill. Remote Sens. 2021, 13, 1568. [Google Scholar] [CrossRef]
  14. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyper-spectral imagery with 3d convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  15. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  16. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-d deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  17. Zhang, H.; Li, Y.; Jiang, Y.; Wang, P.; Shen, Q.; Shen, C. Hyperspectral classification based on lightweight 3-d-cnn with transfer learning. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5813–5828. [Google Scholar] [CrossRef]
  18. Yang, G.; Gewali, U.B.; Ientilucci, E.; Gartley, M.; Monteiro, S.T. Dual-Channel Densenet for Hyperspectral Image Classification. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 2595–2598. [Google Scholar]
  19. Bagaskara, A.; Suryanegara, M. Evaluation of VGG-16 and VGG-19 Deep Learning Architecture for Classifying Dementia People. In Proceedings of the 2021 4th International Conference of Computer and Informatics Engineering (IC2IE), Depok, Indonesia, 14–15 September 2021; pp. 1–4. [Google Scholar]
  20. Aswathy, P.; Siddhartha; Mishra, D. Deep GoogLeNet Features for Visual Object Tracking. In Proceedings of the 2018 IEEE 13th International Conference on Industrial and Information Systems (ICIIS), Rupnagar, India, 1–2 December 2018; pp. 60–66. [Google Scholar]
  21. Xia, M.; Yuan, G.; Yang, L.; Xia, K.; Ren, Y.; Shi, Z.; Zhou, H. Few-Shot Hyperspectral Image Classification Based on Convolutional Residuals and SAM Siamese Networks. Electronics 2023, 12, 3415. [Google Scholar] [CrossRef]
  22. Jiang, Y.; Yu, S.; Wang, T.; Sun, Z.; Wang, S. Skeleton-Based Human Action Recognition Based on Single Path One-Shot Neural Architecture Search. Electronics 2023, 12, 3156. [Google Scholar] [CrossRef]
  23. Xu, H.; Yao, L.; Li, Z.; Liang, X.; Zhang, W. Auto-FPN: Auto-matic network architecture adaptation for object detection beyond classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6648–6657. [Google Scholar]
  24. Zoph, B.; Le, Q.V. Neural architecture search with reinforcement earning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017; pp. 1–16. [Google Scholar]
  25. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the Association for the Advancement of Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 4780–4789. [Google Scholar]
  26. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-Aware neural architecture search for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2820–2828. [Google Scholar]
  27. Jin, J.; Zhang, Q.; He, J.; Yu, H. Quantum Dynamic Optimization Algorithm for Neural Architecture Search on Image Classification. Electronics 2022, 11, 3969. [Google Scholar] [CrossRef]
  28. Liu, H.; Simonyan, K.; Yang, Y. DARTS: Differentiable architecture search. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 30 April–30 May 2019; pp. 1–13. [Google Scholar]
  29. Stanley, K.O.; Clune, J.; Lehman, J.; Miikkulainen, R. Designing neural networks through neuroevolution. Nat. Mach. Intell. 2019, 1, 24–35. [Google Scholar] [CrossRef]
  30. Real, E.; Moore, S.; Selle, A.; Saxena, S.; Suematsu, Y.L.; Tan, J.; Le, Q.V.; Kurakin, A. Large-scale evolution of image classififiers. In Proceedings of the 34th International Conference on Machine Learning, PMLR, Sydney, NSW, Australia, 6–11 August 2017; pp. 2902–2911. [Google Scholar]
  31. Liang, H.; Zhang, S.; Sun, J.; He, X.; Huang, W.; Zhuang, K.; Li, Z. DARTS+: Improved differentiable architecture search with early stopping. arXiv 2020, arXiv:1909.06035. [Google Scholar]
  32. Guo, Z.; Zhang, X.; Mu, H.; Heng, W.; Liu, Z.; Wei, Y.; Sun, J. Single path one-shot neural architecture search with uniform sampling. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Volume 12361, pp. 544–560. [Google Scholar] [CrossRef]
  33. Park, G.; Yi, Y. CondNAS: Neural Architecture Search for Conditional CNNs. Electronics 2022, 11, 1101. [Google Scholar] [CrossRef]
  34. Stanley, K.O.; Miikkulainen, R. Evolving neural networks through augmenting topologies. Evol. Comput. 2002, 10, 99–127. [Google Scholar] [CrossRef] [PubMed]
  35. Chen, Y.; Zhu, K.; Zhu, L.; He, X.; Ghamisi, P.; Benediktsson, J.A. Automatic design of convolutional neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7048–7066. [Google Scholar] [CrossRef]
  36. Hu, J.; Shen, L.; Sun, G. Squeeze–and–excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  37. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-Based Adaptive Spectral–Spatial Kernel ResNet for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7831–7843. [Google Scholar] [CrossRef]
  38. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  39. Misra, D.; Nalamada, T.; Arasanipalai, A.U.; Hou, Q. Rotate to Attend: Convolutional Triplet Attention Module. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3138–3147. [Google Scholar]
  40. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the Computer Vision and Pattern Recognition, Beijing, China, 29 October–1 November 2021. [Google Scholar]
  41. Cao, C.; Xiang, H.; Song, W.; Yi, H.; Xiao, F.; Gao, X. Lightweight Multiscale Neural Architecture Search with Spectral–Spatial Attention for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  42. Tan, H.; Cheng, R.; Huang, S.; He, C.; Qiu, C.; Yang, F.; Luo, P. RelativeNAS: Relative Neural Architecture Search via Slow-Fast Learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 475–489. [Google Scholar] [CrossRef]
  43. Chen, X.; Liang, C.; Huang, D.; Real, E.; Wang, K.; Liu, Y.; Pham, H.; Dong, X.; Luong, T.; Hsieh, C.-J.; et al. Symbolic Discovery of Optimization Algorithms. arXiv 2023, arXiv:2302.06675. [Google Scholar]
  44. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  45. Liu, X.; Meng, Y.; Fu, M. Classification Research Based on Residual Network for Hyperspectral Image. In Proceedings of the 2019 IEEE 4th International Conference on Signal and Image Processing (ICSIP), Wuxi, China, 19–21 July 2019; pp. 911–915. [Google Scholar]
  46. He, X.; Chen, Y. Modifications of the Multi-Layer Perceptron for Hyperspectral Image Classification. Remote Sens. 2021, 13, 3547. [Google Scholar] [CrossRef]
  47. Wang, A.; Xue, D.; Wu, H.; Gu, Y. Efficient Convolutional Neural Architecture Search for LiDAR DSM Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed MS3ANAS model.
Figure 1. The overall framework of the proposed MS3ANAS model.
Electronics 12 03641 g001
Figure 2. Multiple attention mechanism search space (a) CBAM; (d) CA; (b) SE; (c) TA.
Figure 2. Multiple attention mechanism search space (a) CBAM; (d) CA; (b) SE; (c) TA.
Electronics 12 03641 g002aElectronics 12 03641 g002b
Figure 3. Multi-scale attention mechanism guided search space (a) MCS_sepConv(3 × 3); (b) MCS_sepConv(5 × 5); (c) MCS_sepConv(7 × 7).
Figure 3. Multi-scale attention mechanism guided search space (a) MCS_sepConv(3 × 3); (b) MCS_sepConv(5 × 5); (c) MCS_sepConv(7 × 7).
Electronics 12 03641 g003
Figure 4. The slow–fast learning process.
Figure 4. The slow–fast learning process.
Electronics 12 03641 g004
Figure 5. The classification results of the Pavia University dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d)3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Figure 5. The classification results of the Pavia University dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d)3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Electronics 12 03641 g005
Figure 6. The classification results of the Xuzhou dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d) 3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Figure 6. The classification results of the Xuzhou dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d) 3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Electronics 12 03641 g006
Figure 7. The classification results of the WHU-Hi-Hanchuan dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d) 3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Figure 7. The classification results of the WHU-Hi-Hanchuan dataset. (a) Ground truth map; (b) EMP-SVM; (c) 2D-CNN; (d) 3D-CNN; (e) SSRN; (f) MLP-Mixer; (g) CNAS; (h) ANAS-CPA-LS; and (i) MS3ANAS.
Electronics 12 03641 g007
Figure 8. The searched cell architectures of the Pavia University dataset.
Figure 8. The searched cell architectures of the Pavia University dataset.
Electronics 12 03641 g008
Figure 9. The searched cell architectures of the Xuzhou dataset.
Figure 9. The searched cell architectures of the Xuzhou dataset.
Electronics 12 03641 g009
Figure 10. The searched cell architectures of the WHU-Hi-Hanchuan dataset.
Figure 10. The searched cell architectures of the WHU-Hi-Hanchuan dataset.
Electronics 12 03641 g010
Figure 11. The optimal combined action of the search space and the number of generations for each dataset. (a) Pavia University; (b) Xuzhou; (c) WHU-Hi-Hanchuan.
Figure 11. The optimal combined action of the search space and the number of generations for each dataset. (a) Pavia University; (b) Xuzhou; (c) WHU-Hi-Hanchuan.
Electronics 12 03641 g011
Table 1. Comparison of convolution parameters and Flops at different scales.
Table 1. Comparison of convolution parameters and Flops at different scales.
OperationConvolutionRepeatParametersFlops
Conv(3 × 3)Conv(3 × 3)-9 × C29 × C2 × H × W
Conv(5 × 5)Conv(3 × 3)218 × C218 × C2 × H × W
Conv(7 × 7)Conv(3 × 3)327 × C227 × C2 × H × W
Spatial_sepconv(3 × 3)Conv(1 × 3)(Conv(3 × 1))-6 × C26 × C × H × W
Spatial_sepconv(5 × 5)Conv(1 × 3)(Conv(3 × 1))212 × C212 × C × H × W
Spatial_sepconv(7 × 7)Conv(1 × 3)(Conv(3 × 1))318 × C218 × C × H × W
Table 2. Pavia University Dataset Labeled Sample Counts.
Table 2. Pavia University Dataset Labeled Sample Counts.
No.ClassColorSample NumbersFalse Color MapGround Truth Map
1AsphaltElectronics 12 03641 i0016631Electronics 12 03641 i002Electronics 12 03641 i003
2MeadowsElectronics 12 03641 i00418,649
3GravelElectronics 12 03641 i0052099
4TreesElectronics 12 03641 i0063064
5Painted metal sheetsElectronics 12 03641 i0071345
6Bare SoilElectronics 12 03641 i0085029
7BitumenElectronics 12 03641 i0091330
8Self-Blocking BricksElectronics 12 03641 i0103682
9ShadowsElectronics 12 03641 i011947
Total 42,776
Table 3. Xuzhou Dataset Labeled Sample Counts.
Table 3. Xuzhou Dataset Labeled Sample Counts.
No.ClassColorSample NumbersFalse Color MapGround Truth Map
1Bareland-1Electronics 12 03641 i00126,396Electronics 12 03641 i013Electronics 12 03641 i014
2LakesElectronics 12 03641 i0044027
3CoalsElectronics 12 03641 i0052783
4CementElectronics 12 03641 i0065214
5Crops-1Electronics 12 03641 i00713,184
6TreesElectronics 12 03641 i0082436
7Bareland-2Electronics 12 03641 i0096990
8Crops-2Electronics 12 03641 i0104777
9Red-tilesElectronics 12 03641 i0113070
Total 68,877
Table 4. WHU-Hi-Hanchuan Dataset Labeled Sample Counts.
Table 4. WHU-Hi-Hanchuan Dataset Labeled Sample Counts.
No.ClassColorSample NumbersFalse Color MapGround Truth Map
1StrawberryElectronics 12 03641 i02344,735Electronics 12 03641 i024Electronics 12 03641 i025
2CowpeaElectronics 12 03641 i00122,753
3SoybeanElectronics 12 03641 i02710,287
4SorghumElectronics 12 03641 i0285353
5Water spinachElectronics 12 03641 i0291200
6WatermelonElectronics 12 03641 i0304533
7GreensElectronics 12 03641 i0315903
8TreesElectronics 12 03641 i03217,978
9GrassElectronics 12 03641 i03344,735
10Red roofElectronics 12 03641 i0349469
11Gray roofElectronics 12 03641 i03510,516
12PlasticElectronics 12 03641 i03616,911
13Bare soilElectronics 12 03641 i0373679
14RoadElectronics 12 03641 i0389116
15Bright objectElectronics 12 03641 i03918,560
16WaterElectronics 12 03641 i0401136
Total 0
Table 5. Classification results of all methods on the Pavia University dataset.
Table 5. Classification results of all methods on the Pavia University dataset.
MethodsEMP-SVM2D-CNN3D-CNNSSRNMLP-MixerCNASANAS-CPA-LSMS3ANAS
160.77 ± 9.1176.52 ± 7.3286.51 ± 3.5094.01 ± 0.7296.17 ± 1.0796.45 ± 0.1397.70 ± 0.7099.01 ± 0.35
295.68 ± 1.1094.47 ± 3.3594.32 ± 3.6999.40 ± 0.2297.80 ± 0.8198.96 ± 0.0299.46 ± 0.4299.79 ± 0.14
388.06 ± 6.6493.23 ± 0.1369.58 ± 5.8298.26 ± 3.2894.43 ± 3.1698.17 ± 0.1191.13 ± 6.2797.09 ± 0.11
498.38 ± 1.5599.52 ± 0.1595.97 ± 3.0298.73 ± 1.3299.31 ± 1.0099.70 ± 0.0499.66 ± 0.1699.89 ± 0.01
599.46 ± 0.3699.36 ± 0.6399.58 ± 0.1199.64 ± 0.6999.78 ± 0.3399.42 ± 0.04100 ± 0.0099.87 ± 0.12
692.70 ± 1.1681.95 ± 12.1694.80 ± 1.0095.28 ± 0.1999.27 ± 4.1399.00 ± 0.2199.80 ± 0.1199.86 ± 0.07
799.77 ± 0.2242.66 ± 9.3295.07 ± 1.2996.35 ± 0.3598.47 ± 0.5399.41 ± 0.1399.16 ± 0.8399.38 ± 0.36
865.37 ± 6.3171.28 ± 5.1384.52 ± 6.7184.58 ± 0.0381.86 ± 2.7286.12 ± 0.2595.13 ± 0.6895.33 ± 1.33
949.30 ± 8.3396.79 ± 2.2299.77 ± 0.1099.74 ± 0.1699.93 ± 1.4697.47 ± 1.3399.65 ± 0.34100 ± 0.00
OA/%83.39 ± 0.8587.36 ± 1.6791.04 ± 1.7094.68 ± 1.3296.16 ± 0.7197.42 ± 0.2998.44 ± 0.5799.16 ± 0.03
AA/%83.27 ± 3.8689.14 ± 7.2091.12 ± 1.1196.22 ± 1.5896.32 ± 0.9097.18 ± 0.2597.97 ± 0.7298.91 ± 1.14
100 K81.70 ± 1.4482.99 ± 2.4687.99 ± 2.3592.94 ± 0.4194.89 ± 0.0996.76 ± 0.1897.93 ± 0.7698.89 ± 0.09
Params/M0.00320.14390.09340.08940.43050.21640.24320.1376
Flops/G0.00080.12740.46820.48430.02482.83022.72840.0194
Table 6. Classification results of all methods on the Xuzhou dataset.
Table 6. Classification results of all methods on the Xuzhou dataset.
MethodsEMP-SVM2D-CNN3D-CNNSSRNMLP-MixerCNASANAS-CPA-LSMS3ANAS
195.60 ± 1.4685.83 ± 1.8394.71 ± 0.8095.55 ± 1.9296.33 ± 0.0698.86 ± 0.6699.15 ± 0.6599.71 ± 0.13
292.32 ± 0.0490.94 ± 4.7285.10 ± 1.3494.87 ± 4.9799.94 ± 0.6199.06 ± 0.7999.50 ± 0.4999.94 ± 0.03
385.76 ± 0.8282.38 ± 1.8192.21 ± 0.0689.04 ± 6.9993.98 ± 0.4495.09 ± 2.7199.40 ± 0.0399.99 ± 0.01
498.45 ± 1.5793.67 ± 4.8196.86 ± 0.3795.68 ± 0.2895.02 ± 3.0292.82 ± 5.0399.10 ± 0.1699.91 ± 0.04
588.00 ± 0.0997.30 ± 0.8293.33 ± 1.7492.88 ± 0.3796.20 ± 0.9398.57 ± 0.1595.95 ± 0.0194.21 ± 0.62
682.41 ± 0.2483.66 ± 2.1086.41 ± 1.3695.28 ± 1.2679.38 ± 6.1191.75 ± 4.8997.56 ± 0.2497.55 ± 0.45
772.46 ± 3.5290.16 ± 3.1186.47 ± 2.4390.08 ± 1.2392.75 ± 3.0596.40 ± 1.4194.83 ± 0.1599.86 ± 0.98
859.38 ± 2.7796.93 ± 1.8098.81 ± 1.0495.95 ± 0.4589.20 ± 8.1199.60 ± 0.2497.07 ± 2.0699.72 ± 0.07
992.16 ± 1.2689.71 ± 0.3289.72 ± 0.3589.77 ± 2.4998.27 ± 0.3298.51 ± 0.5896.92 ± 0.0799.92 ± 0.01
OA/%84.50 ± 0.8188.32 ± 0.2791.71 ± 0.9392.41 ± 1.3494.54 ± 2.3297.07 ± 0.0697.72 ± 0.1998.44 ± 1.67
AA/%85.17 ± 1.3087.84 ± 0.7291.51 ± 1.8492.90 ± 2.2193.45 ± 2.5196.74 ± 1.8297.67 ± 0.4998.98 ± 0.25
100 K80.38 ± 0.7784.87 ± 0.4389.46 ± 1.3190.37 ± 1.6693.14 ± 2.8997.04 ± 1.0197.11 ± 0.2398.04 ± 2.10
Params/M0.01930.06270.10940.27610.39760.38440.24520.0598
Flops/G0.00590.13540.47170.44890.06433.02622.68100.0198
Table 7. Classification results of all methods on the WHU-Hi-Hanchuan dataset.
Table 7. Classification results of all methods on the WHU-Hi-Hanchuan dataset.
MethodsEMP-SVM2D-CNN3D-CNNSSRNMLP-MixerCNASANAS-CPA-LSMS3ANAS
185.45 ± 2.4788.15 ± 0.3287.16 ± 0.1990.11 ± 1.5291.62 ± 2.6393.71 ± 0.1395.93 ± 0.6897.99 ± 0.31
292.05 ± 2.0187.90 ± 3.1793.69 ± 0.0194.31 ± 2.5295.14 ± 0.0196.32 ± 0.2598.23 ± 0.2398.68 ± 0.14
380.76 ± 3.5382.11 ± 0.0587.66 ± 1.0990.52 ± 1.7790.47 ± 0.3294.41 ± 1.2296.09 ± 2.9298.80 ± 0.14
493.03 ± 2.7786.44 ± 1.5692.53 ± 1.2397.82 ± 0.1697.47 ± 0.7097.97 ± 0.4999.44 ± 0.0499.32 ± 0.12
577.85 ± 16.5952.13 ± 3.1961.44 ± 4.2481.64 ± 5.6078.18 ± 3.0288.30 ± 4.5089.75 ± 2.3996.02 ± 1.44
664.54 ± 7.9365.71 ± 1.6175.41 ± 1.0974.34 ± 1.9184.51 ± 0.0584.48 ± 1.6591.86 ± 0.6894.00 ± 0.01
765.95 ± 1.2469.80 ± 6.1774.36 ± 1.6382.13 ± 0.7882.35 ± 3.4484.44 ± 5.1492.51 ± 2.6493.79 ± 2.94
888.62 ± 0.4983.48 ± 0.8290.43 ± 2.1992.07 ± 1.1291.39 ± 1.0595.86 ± 0.7198.35 ± 0.0298.15 ± 0.13
975.43 ± 5.8178.55 ± 1.3487.84 ± 1.2587.51 ± 2.4788.69 ± 3.0590.85 ± 2.1394.88 ± 0.8797.35 ± 0.38
1090.48 ± 3.1293.35 ± 1.2695.38 ± 1.6995.04 ± 0.4495.66 ± 2.7297.38 ± 1.2299.22 ± 0.0499.73 ± 0.19
1185.60 ± 3.0182.62 ± 0.8589.71 ± 1.2686.97 ± 2.6791.47 ± 1.3791.39 ± 0.0296.60 ± 0.0197.53 ± 0.07
1273.08 ± 2.9074.78 ± 1.8975.03 ± 4.9180.84 ± 0.6986.70 ± 0.8989.59 ± 3.9293.69 ± 0.1994.75 ± 0.33
1367.30 ± 1.7167.97 ± 2.68 81.14 ± 1.1088.13 ± 2.9285.28 ± 1.0887.45 ± 5.3594.50 ± 1.8795.55 ± 0.26
1485.91 ± 0.7384.59 ± 2.7889.09 ± 0.1392.72 ± 1.0889.45 ± 0.8594.22 ± 2.0297.32 ± 0.2397.47 ± 0.93
1588.87 ± 6.2087.60 ± 2.6689.78 ± 4.2184.40 ± 11.690.09 ± 7.5892.99 ± 0.1597.08 ± 0.5695.82 ± 0.71
1697.72 ± 3.0297.48 ± 0.7197.79 ± 0.2699.15 ± 0.1699.11 ± 0.0199.49 ± 1.3799.66 ± 0.1499.77 ± 0.06
OA/%87.96 ± 0.3089.07 ± 0.3191.01 ± 0.3492.78 ± 0.0993.95 ± 0.0296.15 ± 0.0397.44 ± 0.0398.30 ± 0.07
AA/%82.04 ± 3.9783.04 ± 2.4585.53 ± 0.4488.61 ± 1.2489.66 ± 0.9894.43 ± 1.0795.50 ± 0.4497.17 ± 0.25
100 K85.85 ± 0.3787.85 ± 0.3689.43 ± 0.4091.53 ± 0.1191.73 ± 0.0395.32 ± 0.0296.97 ± 0.0898.01 ± 0.08
Params/M0.01700.06770.90281.06530.62110.38070.36720.1564
Flops/G0.00090.14660.47190.47800.10353.96404.08210.0742
Table 8. The classification results of all methods on each dataset.
Table 8. The classification results of all methods on each dataset.
DatasetIndexCNASMS + NASSF + MSNASSF + MSNAS + Lion
Pavia UniversityOA (%)97.42 ± 0.2997.90 ± 0.3798.92 ± 0.2099.16 ± 0.03
AA (%)97.18 ± 0.2597.84 ± 1.3698.07 ± 1.8398.91 ± 1.14
100 K96.76 ± 0.1897.95 ± 0.7998.45 ± 2.5198.89 ± 0.09
Search Cost (hours)3.0543.0132.9822.733
Params (M)0.16920.15740.15790.1564
XuzhouOA (%)97.07 ± 0.0697.84 ± 0.9498.23 ± 0.0998.44 ± 1.67
AA (%)96.74 ± 1.8297.94 ± 0.8198.70 ± 0.5998.98 ± 0.25
100 K97.04 ± 1.0197.18 ± 0.7897.84 ± 0.4298.04 ± 2.10
Search Cost (hours)3.3123.2953.2053.094
Params (M)0.06580.05700.05780.059
WHU-Hi-HanchuanOA (%)96.15 ± 0.0397.04 ± 0.9497.93 ± 0.0998.30 ± 0.07
AA (%)94.43 ± 1.0796.52 ± 0.8197.01 ± 0.5997.17 ± 0.25
100 K95.32 ± 0.0296.11 ± 0.7897.74 ± 0.4298.01 ± 0.08
Search Cost (hours)4.2014.1784.1134.026
Params (M)0.2920.1740.1760.156
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, Y.; Wang, A.; Zhao, Y.; Wu, H.; Iwahori, Y. Multi-Scale Spatial–Spectral Attention-Based Neural Architecture Search for Hyperspectral Image Classification. Electronics 2023, 12, 3641. https://doi.org/10.3390/electronics12173641

AMA Style

Song Y, Wang A, Zhao Y, Wu H, Iwahori Y. Multi-Scale Spatial–Spectral Attention-Based Neural Architecture Search for Hyperspectral Image Classification. Electronics. 2023; 12(17):3641. https://doi.org/10.3390/electronics12173641

Chicago/Turabian Style

Song, Yingluo, Aili Wang, Yan Zhao, Haibin Wu, and Yuji Iwahori. 2023. "Multi-Scale Spatial–Spectral Attention-Based Neural Architecture Search for Hyperspectral Image Classification" Electronics 12, no. 17: 3641. https://doi.org/10.3390/electronics12173641

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop