Next Article in Journal
Developing a Pixel-Scale Corrected Nighttime Light Dataset (PCNL, 1992–2021) Combining DMSP-OLS and NPP-VIIRS
Next Article in Special Issue
Quantifying the Loss of Coral from a Bleaching Event Using Underwater Photogrammetry and AI-Assisted Image Segmentation
Previous Article in Journal
Response of Evapotranspiration (ET) to Climate Factors and Crop Planting Structures in the Shiyang River Basin, Northwestern China
Previous Article in Special Issue
Multi-SUAV Collaboration and Low-Altitude Remote Sensing Technology-Based Image Registration and Change Detection Network of Garbage Scattered Areas in Nature Reserves
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Converging Channel Attention Mechanisms with Multilayer Perceptron Parallel Networks for Land Cover Classification

1
School of Automation, Guangxi University of Science and Technology, Liuzhou 545006, China
2
Guangxi Collaborative Innovation Centre for Earthmoving Machinery, Guangxi University of Science and Technology, Liuzhou 545006, China
3
National Satellite Meteorological Center, China Meteorological Administration, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(16), 3924; https://doi.org/10.3390/rs15163924
Submission received: 26 June 2023 / Revised: 1 August 2023 / Accepted: 4 August 2023 / Published: 8 August 2023
(This article belongs to the Special Issue Computer Vision-Based Methods and Tools in Remote Sensing)

Abstract

:
This paper proposes a network structure called CAMP-Net, which considers the problem that traditional deep learning algorithms are unable to manage the pixel information of different bands, resulting in poor differentiation of feature representations of different categories and causing classification overfitting. CAMP-Net is a parallel network that, firstly, enhances the interaction of local information of bands by grouping the spectral nesting of the band information and then proposes a parallel processing model. One branch is responsible for inputting the features, normalized difference vegetation index (NDVI) and normalized difference water index (NDWI) band information generated by grouped nesting into the ViT framework, and enhancing the interaction and information flow between different channels in the feature map by adding the channel attention mechanism to realize the expressive capability of the feature map. The other branch assists the network’s ability to enhance the extraction of different feature channels by designing a multi-layer perceptron network based on the utilization of the feature channels. Finally, the classification results are obtained by fusing the features obtained by the channel attention mechanism with those obtained by the MLP to achieve pixel-level multispectral image classification. In this study, the application of the algorithm was carried out in the feature distribution of South County, Yiyang City, Hunan Province, and the experiments were conducted based on 10 m Sentinel-2 multispectral RS images. The experimental results show that the overall accuracy of the algorithm proposed in this paper is 99.00% and the transformer (ViT) is 95.81%, while the performance of the algorithm in the Sentinel-2 dataset was greatly improved for the transformer. The transformer shows a huge improvement, which provides research value for developing a land cover classification algorithm for remote sensing images.

1. Introduction

Today’s remote sensing (RS) technology provides us with a large amount of earth observation data, such as satellite images and LiDAR data. These data can not only provide environmental monitoring on a global scale, but can also be used in application areas such as land cover classification and change detection and natural disaster monitoring and assessment [1,2,3,4]. Additionally, improving the accuracy of land cover classification is extremely important for fine agricultural division, earth observation, regional environmental protection, and urban planning tasks [5,6,7,8].
In land cover classification, traditional classification methods mainly use shallow features such as pixel color and texture to classify images [9,10,11,12,13]. In addition, machine learning-based algorithms can also be used for RS image classification, including support vector machines (SVMs) [14], random forests (RFs) [15], K-means clustering (K-Means) [16], K-nearest neighbor (KNN) [17], etc. SVM utilizes the constrained optimal solution problem for hyperplane classification, which is suitable for linearly divisible problems. RF obtains results by constructing multiple decision trees and voting. K-Means uses distance metrics for clustering analysis, but it is sensitive to abnormal samples and cannot deal with discrete features or guarantee global optimality. KNN uses the measurement of distances between different features to classify, but it cannot deal with multi-sample problems. However, these traditional machine learning algorithms still need to further improve the model’s generalization ability to cope with more complex RS image classification tasks. The deep learning (DL) model plays an indispensable role in remote sensing image processing. Due to its multilevel learning property, it can accurately approximate nonlinear relationships [18], thus realizing applications such as classification, fusion, and downscaling [19,20,21]. Deep learning has been successfully applied to land cover classification tasks in RS images, such as U-Net [22], which is a classical fully convolutional network that can obtain good classification results with a lower amount of training data, and thus has become one of the most widely used algorithms in RS image segmentation tasks. Deng et al. introduced a method for land use and cover classification using the U-Net network. A weighted loss function was introduced to address the category imbalance in the data, and data-enhancement techniques were used to improve the robustness of the model [23]. Wu et al. introduced a deep learning method based on an improved version of the U-Net neural network model for the task of sugarcane segmentation in multispectral unmanned aerial vehicle (UAV) images [24]. Hu et al. used the U-Net neural network model based on U-Net for oil palm tree detection in high-resolution remote sensing images [25]. Schiefer et al. used a deep learning method with a convolutional neural network and U-Net for rice fall detection [26].
For complex or heterogeneous feature types, because the feature extraction problem may cause U-Net and improved algorithms to have a large bias on the prediction results since the emergence of transformer [27], it has gradually become the mainstream of RS remote sensing classification. Transformer has a powerful sequence modeling capability and the ability to perceive the global information of the input sequences at the same time, uses the attention mechanism to model the global dependencies of input and output, and uses position coding to solve the problem of how to represent the relative or absolute positional relationships of elements in the sequence. In 2020, Google proposed vision transformer (ViT) [28], which successfully applied transformer to the image classification model and became the mainstay of landmark work of transformer application in the CV field. Hong et al. [29] proposed a new backbone network SpectralFormal (SF), which is able to learn the spectral local sequence information from the neighboring bands of HS images, thus generating the component spectral embeddings and obtaining good classification results.
Although many ViTs achieve good classification results, for example, [30] proposes the ASRC-Net architecture, which fuses CNN with ViT, adopts weight sharing, and uses the same way of processing at different image locations to help the transformer utilize the data more effectively. Considering the advantages of ViT, it can be combined with MLP. ViT has the advantage of being able to consider the relationship between all pixels in an image at both small and large scales, but the disadvantages are also obvious. ViT needs to be learned through a lot of training and because of its increased computational cost, it requires a higher investment of resources.It is sensitive to positional coding: the ViT uses positional coding to assign to each position a specific marker so that the model can recognize the relative position of each location. However, position coding can be affected by operations such as rotation, scaling, and translation, which can lead to a degradation of the model’s performance. Meanwhile, effectively combining ViT with MLP has several advantages. For better feature extraction capability, MLP has an excellent feature extraction capability in image processing, while ViT performs well in natural language processing. Combining them can further improve the accuracy of image processing; faster training speed: ViT uses a self-attention mechanism to process sequence data, which improves processing speed. When combined with MLP, different features in the image can be better captured, thus speeding up the training of the model; better adaptability: MLP and ViT have each been successful in different domains, and in combining them, the model can be made more generalized. This also means that the model can be applied to tasks in different domains, thus further improving the usefulness of the model.
Therefore, in this paper, combining MLP and ViT can utilize the strengths of both to improve the feature extraction capability and training speed, while also providing better adaptability and lower computational cost. The main contributions of the paper include:
(1)
The problem of multispectral image classification is revisited from the perspective of improving the feature extraction capability, and a parallel network architecture integrating the channel attention mechanism and the multilayer perceptron based on the utilization of the feature channels is proposed, which effectively exploits the global correlation information of the image and the feature information between different channels and fully integrates the spatial and channel location correlation. It also allows the lexical features to have a richer expression, so that the pixel-level image classification can be better realized. the level image classification and the comprehensive classification accuracy can both reach 99.00%.
(2)
Adding short wave infrared radiometer (SWIR) bands to the commonly used RGB + NiR bands to form the input seven bands, which can deeply excavate the a priori potentially useful information, and is more conducive to the classification of specific land cover in the precise study area.
(3)
The analysis of land use dynamics changes in the study area focused on the changes in the distribution area of rapeseed and vegetable during the period of 2018–2022. Land use rate and land cover changes can vividly demonstrate the relationship between local economic development and conservation of ecological diversity. These data provide strong support for highly standardized aquaculture and rationalized farmland construction, regularized land improvement projects, and effective decision-making by relevant departments.
The rest of this paper is organized as follows. Section 2 will introduce the materials used for the study and the methods used. Section 3 will present our findings and provide further explanations. Section 4 will explore the potential research directions and future perspectives of these results. Finally, Section 5 will summarize the main findings of this paper and provide suggestions for future research.

2. Materials and Methods

In this section, first, we will briefly introduce the details of the study area and data-processing operations. Subsequently, we will detail the network architecture and the optimization of the framework.

2.1. Study Area Overview

South County, which belongs to Yiyang City of Hunan Province, is located at the border of Xiang and Hubei Provinces; the hinterland of Dongting Lake area; is connected with Shishou, public security, and Songzi of Hubei Province in the north; Anxiang and Hanshou counties of Changde City in the west; Huarong County of Yueyang City in the east; Yuanjiang City of Yiyang City in the south; and several large agricultural (fishing) farms, such as Datong Lake, Beizhouzi, Jinpan, Nanwan Lake, and Qianshanhong in the southeast. It is one of the 36 border counties in Hunan Province. South County exists between 29°3′3″N and 29°31′37″N latitude and between 112°10′53′E and 112°49′6″E longitude, with a total area of 1321 km 2 . It is close to Yueyang in the east, Changsha in the south, and the Yangtze River in the north. It is a subtropical transition to monsoon humid climate type, with cool winters and warm summers, four distinct seasons, sufficient heat, abundant rainfall, long sunshine hours, and short frost periods. It is very suitable for aquaculture and crop cultivation.
Xiangyin County, known as Luocheng in ancient times, is a county under the jurisdiction of Yueyang City, Hunan Province. It is located in the northeastern part of Hunan Province, between the Xiangjiang River and the Zijang River, on the southern shore of Dongting Lake, between 28°30′13″N and 29°3′2″N latitude and between 112°30′20″E and 113°1′50″E longitude. The Xiangjiang River runs from the south to the north, dividing the county into two parts: the eastern part is hilly and the western part is a lakeside plain, with similar types of features as in South County, which are suitable for aquaculture and crop cultivation. In this experiment, the most abundant areas of South County and Xiangyin County were selected for the following experiments. The study area is shown in Figure 1.

2.2. Remote Sensing Image Pre-Processing

To obtain high-quality remote sensing image training samples and accuracy verification samples, field surface cover sample point data are crucial, and the quality of these data directly determines the accuracy of classification. For study area feature types, field data were collected in the field in October 2021, and aquaculture and planting areas with areas larger than 100 m 2 were selected as sample points with priority. This helps to obtain better training and accuracy validation samples.
This paper uses multispectral images as a data source with a resolution of 10 m taken on 6 October 2021, by the Sentinel-2 satellite covering the study area. The satellite carries a multispectral instrument (MSI) containing 13 bands with pixel sizes ranging from 10 to 60 m. The blue (B2), green (B3), red (B4), and NIR (B8) bands have a resolution of 10 m, while the red end (B5), NIR (B6, B7, and B8A), and shortwave infrared SWIR (B11 and B12) are sampled on the ground at 20 m. The blue (B2), green (B3), red (B4), and near-infrared (B8) bands have a resolution of 10 m; the red end (B5), near-infrared NIR (B6, B7, and B8A) and short-wave infrared SWIR (B11 and B12) have a ground-based sampling distance of 20 m; and the pixel sizes of the coastal atmospheric aerosol (B1) and cirrus cloud bands (B10) are 60 m. The pixel size of the coastal atmospheric aerosol (B1) and cirrus cloud bands is 60 m [31] as shown in Table 1.
The Level 1C data were first atmospherically corrected using Sen2Cor-02.10.01-win64 software to remove atmospheric effects on the images, and steps such as image enhancement and mask extraction were performed. Then, the band resolution was resampled to 10 m resolution, and finally, eight commonly used bands (Band2, Band3, Band4, Band8, Band9, Band11, and Band12) were selected for band synthesis using ENVI software. Through a priori knowledge and field data, the region of interest (ROI) was labeled to obtain the sample bank data from South County, Yiyang City, where the sample bank contains 16,510 samples in 7 categories, divided into a training set (70%) and a test set (30%). Xiangyin County in Yueyang City is also mentioned.

2.3. The Architecture Proposed in This Paper CAMP-Net

To address the problems that traditional deep learning algorithms cannot weigh the pixel information of different bands, coupled with the high feature dimensionality of the original remote sensing images as the direct classification of them easily leads to overfitting and high computational complexity, the fused channel attention mechanism and multilayer perceptron parallel network architecture CAMP-Net proposed in this paper is shown in Figure 2, which consists of NDVI, NDWI, and MLP. The structure consists of NDVI, NDWI, MLP and the transformer with channel attention. The band information of NDVI and NDWI is directly fused with the input of the transformer; the features obtained by the transformer encoder are merged with the features obtained by MLP; the merged features are reduced in dimensionality by a fully connected layer; and finally, the classification results are obtained by activation function and MLP-head to achieve pixel-level multispectral image classification.

2.3.1. NDVI and NDWI

In the field of remote sensing, we often use vegetation index (VI) to quantitatively assess the growth of vegetation. One of the most commonly used vegetation indices is the normalized difference vegetation index (NDVI), which determines the health and coverage of vegetation by calculating the reflectance of different wavelength bands. NDVI is commonly used to monitor the growth status of various ecosystems such as large-scale agriculture, forestry, grassland, and wetlands.
The value of NDVI is calculated by Equation (1) to obtain the calculation formula:
N D V I = N I R R E D N I R + R E D
where NIR indicates near-infrared band reflectance and RED indicates red band reflectance. NDVI takes values between −1 and 1, with higher values representing more vegetation coverage and better health. If the NDVI is 0, it means that the vegetation coverage is 0. A negative number means that the area is mainly composed of non-vegetation types such as water bodies, buildings, and roads.
The full name of NDWI is “Normalized Difference Water Index”, which means the normalized water body index. In remote sensing, this index is used to monitor water bodies and wetlands and can help detect floods, droughts, snow and ice, soil moisture content, etc. The value of NDWI is calculated by Equation (2):
N D W I = G r e e n N I R G r e e n + N I R
where NIR represents the reflectance of the NIR band and Green represents the reflectance of the green band. In images formed by remote sensing images of visible (green and blue) and infrared band signals under cloudless skies, NIR is less affected by the atmosphere and easier to obtain; there is a natural correlation between the green band and NIR band reflectance, which can be described by an equation. Higher water content corresponds to index values ranging from −1 to +1, with higher values representing more water bodies in the region and lower values if the region is all land.
The steps for fusion using NDVI and NDWI band information in the VIT framework are as follows. First, image data containing NDVI band information need to be prepared. Calculating NDVI and NDWI depends on the remote sensing data used. After calculating NDVI, it is added to the image as an additional band to use as an input to the VIT model. Usually, all bands can be superimposed into a four-dimensional array (i.e., tensor), where the first dimension represents the number of images, the second and third dimensions represent the height and width of the image, and the fourth dimension represents the number of bands (including NDVI and NDW bands). Then, NDVI and NDW are incorporated into the ViT framework as band information.

2.3.2. Multi-Layer Perceptron Based on Feature Channels

MLP based on feature channels is a common artificial neural network model that can be used for tasks such as classification and regression. It comprehends patterns in input data and uses feature channels for learning. Each feature channel corresponds to one dimension of the input data, such as R, G, and B of a color image.
The feature-channel-based multilayer perceptron is a neural network model consisting of an input layer, multiple hidden layers, and an output layer that can handle nonlinear data and adapt to complex data distributions. The model is highly flexible and performs well in different tasks because it does not restrict the relationship between features. In each hidden layer, the result of the weighted summation of the inputs is converted to the output by an activation function, and the output of the previous layer is used as input for the next layer. The output of the last hidden layer is fed to the output layer for classification or regression. With the ability to handle nonlinear data without restricting the relationship between features, it can adapt to complex data distributions, and the model is highly flexible and performs well in different tasks.
Traditional MLPs use sigmoid activation functions but may lead to gradient disappearance in backpropagation. A better choice is a variant such as GELU, which we use as well as regularization, while L1 and L2 regularization prevents overfitting and improves generalization; applying dropout reduces neuronal correlation and improves model robustness, as shown in Figure 3.

2.3.3. Channel Attention

The transformer is a deep learning architecture that has demonstrated impressive performance in various tasks such as image recognition and semantic segmentation. The transformer was originally designed for sequence-by-sequence problems, and its core consists of an attention mechanism and positional coding. In the field of image processing, it is mainly used for sequence-to-sequence problems. Due to its self-attention and positional encoding, the transformer model outperforms traditional RNN models because it does not require the input and output sequences to be fully aligned, which provides significant benefits for processing sequential data. In addition, the transformer’s structure can understand the interactions between different spatial locations and solve the long-term dependency problem thanks to its parallel computing capability, which significantly reduces computational resource consumption.
Our proposed algorithm uses the transformer encoder as an encoder with N encoder layers, each including multiple self-attentive modules and feedforward layers, as well as inputs including position encoding and feature embedding, with the transformer using a fixed position encoding to represent the absolute position information of the token. The most “core” channel attention structure is used, which is the epitome of parallel attention. Channel attention is not limited to calculating one-time attention since the method fully considers the interrelationship between all spatial locations. Channel attention is used for input F with the maximum pooling layer and the average pooling layer, in the same MLP network after the output, in the sum after the activation function to obtain the final output as shown in Figure 4.
Channel attention is an attention mechanism in neural networks that enhances the interaction and information flow between different channels in a feature graph to improve the performance of the model. By applying channel attention, the relationship between different channels in a feature map can be encoded to eliminate redundant information, focus on important features, and improve the expressiveness of the feature map. The basic idea of channel attention is to calculate the importance weight of each channel in the feature map by a “compression” operation, then to weigh each channel by a “stimulation” operation, and, finally, to output the weighted feature map. This “compress-stimulate” attention mechanism is simple to implement and also has good robustness and scalability.
The benefits of changing the ViT self-attentive mechanism to channel-attentive are as follows:
(1)
Better capture of image features: By using channel attention instead of the self-attention mechanism, the relationship between different channels (i.e., different color channels) can be better captured and the attention on important channels can be improved. This is important in vision tasks where different color channels may affect the image features that need to be learned.
(2)
Higher computational efficiency: The self-attention mechanism needs to calculate the similarity scores between all position pairs, so it is computationally more expensive. However, using the channel attention mechanism can allocate attention to different channels, thus making the model more computationally efficient.
(3)
Better generalization ability: Changing the self-attention mechanism to the channel-attention mechanism can also improve the generalization ability of the model. This is because the channel attention mechanism allows global evaluation of the whole image, thus improving the model’s ability to understand more complex and variable visual scenes.
The channel attention mechanism implementation can be expressed as Equations (3)–(5):
z = 1 c i = 1 c x i
w = σ W 2 R e L U W 1 z
γ = w x
where x is the input tensor with dimensions (N, C, H, and W), and N denotes the batch size. c indicates the number of channels, and H and W indicate the height and width, respectively. z is the vector obtained by global average pooling of x, the average over the channel dimensions. W 1 and W 2 are the learnable weight matrix, and R e L U denotes the modified linear cell operation. σ is the sigmoid function. γ is the output tensor, which is the input tensor using attentional weighting. ⊙ denotes multiplication by elements. The channel attention mechanism uses two fully connected layers and a sigmoid function, which can be understood as the model automatically learns the importance of each channel and enhances the signal of useful channels and cuts the information of useless channels by weighting the average, thus improving the model performance.

2.3.4. Evaluation Indicators

For the evaluation of the classification of multispectral pixels, three hybrid matrix-based evaluation metrics are used, including overall accuracy, average accuracy, and kappa. Pixel-level evaluation metrics such as accuracy, precision, and kappa are used mainly to evaluate the ability of the proposed model to classify accurately.
(1)
Overall accuracy (OA): The number of correct predictions as a percentage of the number of all predictions.
O v e r a l l A c c u r a c y = T 1 + T 2 T 1 + F 1 + T 1 + F 2
where T 1 is true positive, F 1 is false positive, T 2 is true negative, and F 2 is false negative. T 1 + T 2 denotes the total number of correctly classified samples. F 1 + F 2 denotes the total number of misclassified samples. In the vast majority of cases, a higher OA value indicates a better performance of the classifier. However, in some cases, the OA metric may be less appropriate due to the imbalance of data distribution, for example, a very small percentage of samples in a certain category leads to a high OA value when the classifier only needs to predict all the samples in that category as another category. Therefore, in this case, other metrics need to be used to evaluate the model performance comprehensively.
(2)
Average accuracy (AA): The average accuracy, which is a more accurate evaluation metric. Unlike OA, AA values are calculated separately on each category and then averaged as the final score of the model. Its calculation formula is as follows.
A v e r a g e A c c u r a c y = 1 n T 1 + T 2 T 1 + F 1 + T 1 + F 2
where n denotes the number of categories.
(3)
Kappa coefficients: Kappa is a measure of the accuracy of a classifier or evaluation system, also known as Cohen’s kappa coefficient, which evaluates the consistency of a classification model with its predictions in a real test. The kappa coefficient takes values in the range [−1,1]. The formula is as follows.
K a p p a = P l P x 1 P x
P l denotes the exact match of the observed data, i.e., the proportion of samples for which the classifier predicts exactly the same result as the actual situation. P x is the likelihood that the classifier predicts the classification and the actual situation to obtain consistent results.

2.3.5. Loss Function

Cross-entropy is a method used to measure the variability between probability distributions. The output in a multiclassification task is the probability that the target belongs to each category, and the sum of the probabilities of all categories is 1, in which the category with the highest probability is the one to which the target belongs. We choose cross-entropy as the loss function in our algorithm, which is useful in its multiclassification task because it allows the model to better distinguish the differences between categories, and it can be used to minimize the loss function by optimization algorithms such as gradient descent to improve the accuracy of the model. The cross-entropy loss function compares the predicted class of each pixel with the target class and has the following expression:
J = 1 X i = 1 M y i c c = 1 x y l x log p ^ x y
where X indicates the number of sample categories; x denotes the number of categories, and y l x denotes the true label of the lth sample belonging to class x. p ^ x y is the probability value indicating that the model predicts that the sample belongs to class y, between 0 and 1.

2.3.6. Experimental Environment

In the experiments, the coordinates of the region of interest are obtained using ENVI software, and the Adam optimizer [32,33] is applied in the training phase, the batch size is set to 32, the maximum number of iterations is 300, and the initialized learning rate is 5 × 10 4 , decaying by 0.9 at every 30 iterations. The implementation code is written using Python 3.9 in PyTorch 1.12.1 The training model was written on a Win11 + 12th Gen Intel(R) Core(TM) i5-12400F + NVIDA GeForce RTX 3060 Laptop GPU.

3. Results

The details of the sample pool in the study area are shown in Table 2, with 70% used for training and 30% for testing. The following experiments are conducted using the sample size of the dataset in Table 2.
To highlight the good performance effect improvement of our proposed algorithm, CAMP-Net is compared with SVM, RF, KNN, CNN, RNN, ViT, and SF in this paper.

3.1. Ablation Study

To evaluate the performance of the proposed network structure in this paper, ablation experiments were performed on the dataset in the study area. The experimental results are shown in Table 3. The OA of the original ViT is 95.81%, which is a satisfactory result, indicating that ViT is more suitable for handling multispectral image-classification problems. Although the OA of ViT + CA was improved to 96.76, it also proves the effectiveness of the ViT + CA scheme, which can enhance the tight connection of channels by adding channel attention. The OA of the MLP module inserted into ViT is improved by 1.10% compared with the original ViT, and AA and kappa are also improved, indicating that in multispectral data, MLP may improve ViTs classification of a subset of categories. The feature fusion of the MLP module with ViT incorporating NDVI and NDWI band information achieved an AA of 98.15, which is a 2.34% improvement relative to ViT. Kappa also improves by 1.80% over ViT, thus verifying that the feature fusion of the MLP module with ViT incorporating NDVI and NDWI band information compares with ViT for multispectral image pixel classification task is more advantageous. However, the OA is improved when the attention of ViT is changed to channel attention and the band information of NDVI and NDWI is added. The OA improves by 98.63% when the attention of ViT is changed to channel attention with MLP first processed in parallel and then fused features, but the accuracy of CAMP-Net is 99.00 when all modules are tested together, which is an OA improvement compared to the original ViT 3.19%.AA and kappa also have a good improvement, so the achieved results are relatively good.
The classification results of different module pair combinations on the study area dataset are shown in Figure 5, where water, rapeseed, and pond are present. Comparing b and c shows that NDVI + NDWI classifies lotus better than ViT. The network structure of Figure 5c–e alone plus the modules does not classify pond as well as Figure 5f–h. The complete CAMP-Net outside is more distinguishable than the combination of his modules. The difficulty of the classification task for this region is the mixture of pond and vegetable in this region and the performance of the complete CAMP-Net is the best, with a good denoising effect, which can improve the performance of ViT better.
To explore the effect of the number of neighboring spectra on CAMP-Net and ViT, different numbers of neighboring spectra were taken for the experiments, and the values of the number of neighboring spectra were taken as 1, 2, 3, 4, 5, 6, 7, and the experimental results are shown in Table 4. According to the experimental results, it is known that ViT, AA, and OA, and kappa are the highest when the number of neighboring spectra is 5. For CAMP-Net, the overall results of CAMP-Net were the best when the number of neighboring spectral bands was 6. The experimental results show that the performance is not necessarily better if the number of neighboring spectral bands is higher, but the selection of the appropriate number of neighboring bands has some improvement on the model performance. Therefore, we chose the number of neighboring bands as six in the later experiments.
To evaluate the effect of small sample size training samples on the experimental results, we randomly selected 10–90% of samples from South County’s sample pool for training validation, and the remaining samples were used for testing when no separate validation set was set up, and the experimental results are shown in Table 5. It is not the case that the proportion of samples increases for better results, and as the proportion of classification samples increases with the training samples, AA does not necessarily better effect and the noise generated is also randomly assigned. Since random sample selection leads to random sample point selection, randomly selected training sample points may not contain all sub-regions of ROI, leading to obvious misclassification, such as (c) misclassifying water into a pond. so to make the best overall classification effect, we choose a sample ratio of 7:3 for the experiment.
To evaluate the effect of small-sample-size training samples on the experimental results, we randomly selected 10–90% of samples from South County ’s sample pool for training validation, and the remaining samples were used for testing when no separate validation set was set up; the experimental results are shown in Table 5. It is not the case that the proportion of samples increases for better results, and as the proportion of classification samples increases with the training samples, AA does not necessarily have a better effect and the noise generated is also randomly assigned. Since random sample selection leads to random sample point selection, randomly selected training sample points may not contain all sub-regions of ROI, leading to obvious misclassification, such as (c) misclassifying water as a pond. Therefore, to make the best overall classification effect, we choose a sample ratio of 7:3 for the experiment.
The results of training samples using different ratios are shown in Figure 6. One can see from the figure that difference between the ratios of the training samples is very small, so in the case of the results of the graph, there is not much difference. We chose to test the results of the higher accuracy of the sample ratio of 7:3 for the experiment.
The data source for the study area taken by the Sentinel-2 satellite contains a total of 13 spectral bands, but other satellites do not necessarily contain as many. However, they all must contain RGB and NiR bands. In order to increase the versatility of the dataset, researchers usually select four common bands, RGB + NiR, from the 13 bands as the pixel sequence information for experimental studies. To determine whether it is worthwhile to prioritize fast processing and broad applicability of the four common Sentinel-2 bands and discard the a priori potentially useful information recorded by Sentinel-2, we first added the VNIR band for the experiment, and then the short-wave infrared (SWIR) bands for comparison. This also helps to classify specific land cover in the precise study area when more Sentinel-2 bands are used. The experimental results are shown in Table 6. When all SWIR bands are added, a significant improvement is obtained with very good results, except for vegetables.
The results of different methods using different bands are shown in Figure 7, from which it can be seen that the CAMP-Net algorithm is better than the ViT algorithm for classification, while the four bands + SWIR is particularly good.

3.2. Multi-Method Comparison

The quantitative classification results of OA, AA, and kappa for the study area datasets of South County and Xiangyin County and the accuracy of each category are shown in Table 7 and Table 8. The overall effect of CNN is the worst, OA, AA, and kappa are lower than other models, and the accuracy for building and vegetable is only 70.22% and 72.65%, most likely It is likely because the multispectral images have fewer bands and only individual bands are used, which is not conducive to CNN learning features, while hyperspectral images have 200 bands of information, and CNN outperforms SVM as well as KNN on hyperspectral images. Traditional classifiers SVM and KNN achieve relatively good results with OA of 89.41% and 92.19%, respectively, and SVM has a better performance on lotus, and KNN on building has poor classification ability. RNN, ViT, SF, and CAMP-Net are all spectral sequence classification methods based on deep learning, and the performance of the four models is relatively similar, proving the advantages of deep learning for sequence data processing. To better demonstrate the performance of each model, the Sentinel-2 multispectral images of the study area in 2018, 2020, and 2022 are also analyzed in Section 3.3. The classification performance of ViT and SF is well done, and the OA, AA, and Kappa of CAMP-Net is higher than the other comparable models, and the performance is even better in various categories such as building, water, and rapeseed.
The quantitative classification results of OA, AA, and kappa for the study area datasets of South County and Xiangyin County and the accuracy of each category are shown in Table 7 and Table 8. The overall effect of CNN is the worst; OA, AA, and kappa are lower than other models, and the accuracy for building and vegetable is only 70.22% and 72.65%. This is likely because the multispectral images have fewer bands and only individual bands are used, which is not conducive to CNN learning features, while hyperspectral images have 200 bands of information, and CNN outperforms SVM as well as KNN on hyperspectral images. The traditional classifiers SVM and KNN achieve relatively good results with OA of 89.41% and 92.19%, respectively. SVM has a better performance on lotus, and KNN on building has poor classification ability. RNN, ViT, SF, and CAMP-Net are all spectral sequence classification methods based on deep learning, and the performance of the four models is relatively similar, proving the advantages of deep learning for sequence data processing. To better demonstrate the performance of each model, the Sentinel-2 multispectral images of the study area in 2018, 2020, and 2022 are also analyzed in Section 3.3. The classification performance of ViT and SF is well done, and the OA, AA, and kappa of CAMP-Net is higher than the other comparable models, and the performance is even better in various categories such as building, water, and rapeseed.
The classification map of South County obtained by different models is shown in Figure 8, and the study area in the square box is labeled according to the a priori knowledge and outdoor sampling data. CAMP-Net can better distinguish between wetland and water, which are less different within groups and does not misclassify vegetable, and it is obvious that CAMP-Net is clearer than the other algorithms.
As shown in Figure 9, the overall classification results of SVM and CNN are very poor; the remaining algorithms of CAMP-Net have clearer edge extraction and more obvious chunking for pond and vegetable, and the part between adjacent water regions is not misclassified as wetland. In general, RNN, SF, and ViT are significantly better for the classification of multispectral images than CNN and SVM. The proposed CAMP-Net, although the model structure is more complex than ViT, has some improvement in OA, AA, and kappa, and the map-forming effect is also better than ViT, which has great value.

3.3. Land Use Change

This study used 17 July 2017 and 4 November 2019 Sentinel-2 series imagery downloaded from the USGS website (https://earthexplorer.usgs.gov/ (accessed on 1 March 2023)) for the study of feature classification. After pre-processing, the ROI was re-tagged using ENVI5.3 software based on the a priori knowledge accumulated from the 2021 outdoor collection data, and a .txt file containing the ROI coordinates was exported. Based on the coordinates of the .txt file, the sample set was randomly divided into a training set and a test set according to a 7:3 ratio. A variety of different models from Table 4 were also used in this paper for comparative analysis among the models. Among them, the sample pool data for 2017 and 2019 are detailed in Table 9. Finally, the CAMP-Net proposed in this paper was used to carry out a land use change analysis for the study area, focusing on the changes in rapeseed and vegetables.
The classification results of different models for the sample pool data in 2018, 2019, and 2022 are shown in Table 10, Table 11 and Table 12. In the datasets of three different years, the OA of CAMP-Net is 98.25%, 97.06%, and 99.31%, which are higher than the other models. The overall classification is also better than the other models. Overall, 18, 20 and 22 years are better than the original ViT. There is an improvement of 1.70%, 4.27%, and 1.64% compared to the original ViT. The lower level of improvement in 18 and 22 years may be due to the original VIT already being high, while the improvement in 20 years is more demonstrative compared to the 21 year-results we obtained, and the classification performance of CAMP-Net meets the requirements of conducting land use change on the study area.
The spatial distribution of land use in 2018, 2020, and 2022 is shown in Figure 10 and the spatial distribution of each feature can be observed. The spatial distribution of each feature shows that vegetable is mainly distributed in the northwest and central part of the country, and rapeseed is more concentrated in the middle of the country. The pond is mainly located in the west part of the study area, surrounded by buildings, and water is mainly distributed in the southeast part. The spatial distribution of major land uses in the study area has become more balanced and rational in recent years.
The land use and land use change rate of the study area in the last 4 years are shown in Table 13. In the three images, the largest proportion of the coverage area in the study area is vegetable, and the largest change is lotus. Except for lotus, the change in other species in the last four years is relatively small, due to the increase in pond. Aquatic and crops are relatively increased while building and wetland are relatively decreased. The area is also increased, and the aquatic category is naturally valued by the people of Hunan as a specialty food and the natural geographical advantage of Hunan, so it has been growing continuously. Various data during the 4 years indicate that the study area meets the sustainable development strategy.
Since rapeseed and vegetables are the main aquatic species and crops in the study area, this paper focuses on the spatial distribution of rapeseed and vegetables, whose spatial distribution map is shown in Figure 11. From the overall observation, the area where pond is located is mainly scattered near the river in the central region. The rich water resources in the area are conducive to the development of aquaculture as well as cultivation, and the good sales of seafood products have promoted the local people to vigorously develop aquaculture. Rapeseed and vegetable are scattered in various areas of the study area, and these two species play a very important role in making full use of natural and labor resources to increase food production, so it is reasonable for the government to carry out the full aspect of agricultural farming.
The dynamics of rapeseed and vegetables from 2018 to 2022 are shown in Figure 12. Rapeseed’s farming area was 20.69 km 2 in 2018, 18.99 km 2 in 2020, and 19.72 km 2 in 2022. In 2020, 2020–2022, and 2018–2022, respectively, the area decreased by 8%, increased by 4%, and decreased by 5%. The dynamics of total rapeseed and vegetable from 2018 to 2022 are shown in Figure 12. Rapeseed’s farming area was 20.69 km 2 in 2018, 18.99 km 2 in 2020, 19.72 km 2 in 2022, and 19.72 km 2 in 2018. In 2020, 2020-2022, and 2018–2022, respectively, the area decreased by 8%, increased by 4%, and decreased by 5%. The is an overall decreasing trend from 2018 to 2022, but 2020–2022 is a rebound state, probably because rapeseed farmers can obtain more income. At the same time, the climate of the Hunan region is also more suitable for planting, leading farmers to increase planting rapeseed in recent years. The increase in vegetable planting area is larger, especially from 2020 to 2022, which is expected. Because of the fertile soil and water in the region, there is a natural advantage of growing more vegetables to obtain greater returns, and the distribution of the various species planted may be due to some coordination by the national government to achieve joint development in many aspects. Vegetable planting area in 2018 was 41.83 km 2 , in 2020 it was 37.94 km 2 , and 2020 it was 37.94 km 2 . The increase in vegetable planting area, especially from 2020 to 2022, was expected because the area is fertile and has natural advantages for cultivation. There is a natural advantage of growing more vegetables for greater profitability, and the distribution of the various species planted may be due to some coordination by the national government to achieve a multifaceted co-development.

4. Discussion

In this work, we propose a new novel network for land cover classification, which is different and innovative from previous studies because of the following points:
(1)
Our work adds NDVI and NDWI at the beginning neighboring bands and then uses Gse spectral feature module and selects every two sequences for band fusion in the transformer branch, while it only uses half of every two sequences for fusion in the experiment.
(2)
Our paper uses two layers of attention stacked in the transformer branch and a channel MLP branch for feature extraction, while the other papers use a CNN branch and choose to cascade the two convolutions to form a DenseNet structure.
(3)
Our paper merges the features obtained by a transformer encoder with the features obtained by channel MLP, while other papers use the features obtained by the transformer encoder with the features obtained by CNN to merge in order to produce the features in a parallel manner.
(4)
Our paper merges the resulting features to reduce the dimensionality through the fully connected layer and finally classifies them through the activation function and conv1*1 and the MLP head in the vision transformer (ViT), while it uses only the activation function and conv1*1 to obtain the classification results.
Because SWIR bands cover a wider spectral range, more optical information can be captured. Although the bandwidth of SWIR technology is narrower than that of RGB bands, it possesses a strong lateral extension capability, which enables it to cover weaker areas and areas that cannot be sensed by RGB bands, thus improving the accuracy of feature information. SWIR bands have a strong penetration capability, which allows them to penetrate clouds, smoke, and haze, and maintain a good performance even in the presence of thicker cloud layers. Since SWIR wavelengths are closely related to the structure and composition of substances, the unique spectral characteristics of certain substances or objects can be captured. Therefore, instead of adopting the RGB + NiR four bands, which is significantly applicable, the four bands + SWIR method was used to carry out the experiments, which helps to enhance the accuracy of the study and the credibility of the results.
Compared to traditional classification methods that cannot comprehensively consider the interrelationships between pixels in an image, the classification effect may be less than ideal in some complex application scenarios. In contrast, multispectral data provide a wider range of spectral information and take into account the interaction between various bands. This makes land cover mapping techniques based on multispectral data more accurate and reliable. Image classification based on multispectral data can identify land cover types by analyzing the correlation between different bands. In addition, multispectral data can provide information on many other surface characteristics, such as vegetation index, soil moisture content, etc., which are important for land management and environmental monitoring. This experiment compares the advantages and disadvantages of the traditional machine learning algorithms SVM, KNN, and RF and the deep learning algorithms CNN, RNN, transformer (ViT), SF, and CAMP-Net (a total of eight methods), with respect to the multispectral images, and the results show that our proposed algorithm, CAMP-Net, has a great advantage in the classification performance. Among the traditional methods, SVM, KNN, and RF have a high efficiency with a short training time, but their ability to capture band information is not as good as the RNN model, which is good at processing sequence information. Since the information of multispectral images is not as good as hyperspectral images, it means the effect of classification accuracy of CNN is not as good as expected, but unlike RNN, which is good at dealing with sequence data, the classification performance of RNN is better than that of CNN, but it is difficult for it to learn the long-term dependency of the input and output like transformer. This is because the transformer can capture global sequence information through positional encoding. The classification performance of SF on multispectral images is better than all models except CAMP-Net in all datasets.
Interestingly, CNN with the addition of SWIR band information also fails to capture the spatial information well. This may be due to the low dimensionality of the spatial kernel and the band information is just too little. It would be better if a dataset of hyperspectral images is used. Adding the attention mechanism or increasing the network depth can further improve the image-classification performance. The scaled 1D sequence data is transformed from a Cartesian coordinate system to a polar coordinate system, and the temporal correlation of different time points is recognized by considering the angle sum/difference between different points. GASF (corresponding to angular sum) or GADF (corresponding to angular difference) can be utilized to do the angles, and the implementation method depends on the specific requirements. Adding NDVI and NDWI band information for grouped spectral nesting, a multilayer perceptron network based on feature channels, and CAMP-Net with the transformer, which adds channel attention and performs best on five different datasets. The fusion of parallel networks can combine the features learned by transformer and MLP, respectively, to enhance the feature expression ability and thus obtain better classification results. Then, the land use dynamic change analysis can be carried out to obtain a clearer understanding of the land cover classification changes in the study area in recent years.

5. Conclusions

In this paper, a new novel network for land cover classification is proposed, which employs a network architecture that incorporates a fusion channel attention mechanism in parallel with a multilayer perceptron based on feature channel utilization. In order to enhance the band local information interaction, grouped spectral nesting is used to process the features after adding NDVI and NDWI band information. Then, in the ViT branch, the information between spectral features is deeply mined by incorporating the channel attention mechanism; in the other branch, a multilayer perceptron network based on feature channel utilization is designed to improve the network’s ability to extract features from different feature channels. The feature information obtained from both is fused, the dimensionality of the features is reduced by a fully connected layer and the classification results are obtained by activation function and MLP-head. The integrated classification accuracy of Nanxian in 2021 in the study area can reach 99.00%, which is a 3.19% improvement in OA compared to the original ViT. The classification accuracy of Xiangyin County in 2021 also reaches 99.39%, which is a 2.47% improvement compared to the ViT’s 96.92. In the Nanxian dataset, in 18, 20, and 22 years the OA of CAMP-Net is, respectively, 98.25%, 97.06%, and 99.31%, which are 1.70%, 4.27%, and 1.64% improved compared to the original ViT. This huge improvement proves that pixel-level multispectral image classification is achievable. Finally, in future works, we will further improve the pixel data sequences in order to explore more information of the information features to be used. We believe that a useful way to continue this research in future works is to explore an enhanced network architecture that can utilize more information from the pixel data to improve the performance and allow for wider applications to multiple fields and domains to improve efficiency.

Author Contributions

Conceptualization, X.F. and X.L.; methodology, X.L. and X.F.; software, X.L., X.F., C.Y., J.F., N.W. and L.C.; validation, X.L., X.F. and J.F.; formal analysis, X.L., X.F. and C.Y.; investigation, X.L., X.F., C.Y., J.F., N.W. and L.C.; resources, J.F. and X.F.; data curation, X.L., X.F., C.Y., J.F., N.W. and L.C.; writing—original draft preparation, X.L.; writing—review and editing, X.L., X.F., C.Y., J.F., N.W. and L.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is the result of the research project funded by the National Natural Science Foundation of China (62261004 and 62001129) and the Innovation Project of Guangxi University of Science and Technology Graduate Education (GKYC202321).

Data Availability Statement

The data are available at https://github.com/lixuyaaaaa/CAMP-Net accessed on 1 September 2023.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CAMP-NetCombines the channel attention mechanism and the multi-layer perceptron
parallel algorithm.
ViTVision transformer
RSRemote sensing
PolSARPolarized synthetic aperture radar
CNNConvolutional neural network
SFSpectralFormal
SVMSupport vector machine
RFRandom forest
CN.Class number
GSEGrouped spectral embedding
K-MeansK-means clustering
SWIRShort wave infrared radiometer
ROIRegion of interest
NDVINormalized difference vegetation index
NDWINormalized difference water index
MLPMulti-Layer Perceptron
OAOverall Accuracy
AAAverage Accuracy
UAVUnmanned aerial vehicle

References

  1. Huang, C.; Davis, L.; Townshend, J. An assessment of support vector machines for land cover classification. Int. J. Remote Sens. 2002, 23, 725–749. [Google Scholar]
  2. Duro, D.C.; Franklin, S.E.; Dubé, M.G. A comparison of pixel-based and object-based image analysis with selected machine learning algorithms for the classification of agricultural landscapes using SPOT-5 HRG imagery. Remote Sens. Environ. 2012, 118, 259–272. [Google Scholar] [CrossRef]
  3. Wang, J.; Wang, S.; Wang, F.; Zhou, Y.; Wang, Z.; Ji, J.; Xiong, Y.; Zhao, Q. FWENet: A deep convolutional neural network for flood water body extraction based on SAR images. Int. J. Digit. Earth 2022, 15, 345–361. [Google Scholar] [CrossRef]
  4. Rey, S.J.; Anselin, L.; Li, X.; Pahle, R.; Laura, J.; Li, W.; Koschinsky, J. Open geospatial analytics with PySAL. ISPRS Int. J. Geo. Inf. 2015, 4, 815–836. [Google Scholar] [CrossRef] [Green Version]
  5. Jones, H.G. Use of infrared thermometry for estimation of stomatal conductance as a possible aid to irrigation scheduling. Agric. For. Meteorol. 1999, 95, 139–149. [Google Scholar] [CrossRef]
  6. Schofield, R.; Thomas, D.S.; Kirkby, M.J. Causal processes of soil salinization in Tunisia, Spain and Hungary. Land Degrad. Dev. 2001, 12, 163–181. [Google Scholar] [CrossRef]
  7. Peng, J.; Pan, Y.; Liu, Y.; Zhao, H.; Wang, Y. Linking ecological degradation risk to identify ecological security patterns in a rapidly urbanizing landscape. Habitat Int. 2018, 71, 110–124. [Google Scholar] [CrossRef]
  8. Long, H.; Li, Y.; Liu, Y.; Woods, M.; Zou, J. Accelerated restructuring in rural China fueled by ‘increasing vs. decreasing balance’ land-use policy for dealing with hollowed villages. Land Use Policy 2012, 29, 11–22. [Google Scholar] [CrossRef]
  9. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar]
  10. Planinsic, P.; Singh, J.; Gleich, D. SAR image categorization using parametric and nonparametric approaches within a dual tree CWT. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1757–1761. [Google Scholar] [CrossRef] [Green Version]
  11. Papadomanolaki, M.; Vakalopoulou, M.; Karantzalos, K. A novel object-based deep learning framework for semantic segmentation of very high-resolution remote sensing data: Comparison with convolutional and fully convolutional networks. Remote Sens. 2019, 11, 684. [Google Scholar] [CrossRef] [Green Version]
  12. Guo, X.; Chen, Z.; Wang, C. Fully convolutional DenseNet with adversarial training for semantic segmentation of high-resolution remote sensing images. J. Appl. Remote Sens. 2021, 15, 016520. [Google Scholar] [CrossRef]
  13. Poursanidis, D.; Chrysoulakis, N.; Mitraka, Z. Landsat 8 vs. Landsat 5: A comparison based on urban and peri-urban land cover mapping. Int. J. Appl. Earth Obs. Geoinf. 2015, 35, 259–269. [Google Scholar] [CrossRef]
  14. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  15. Ayerdi, B.; Romay, M.G. Hyperspectral image analysis by spectral–spatial processing and anticipative hybrid extreme rotation forest classification. IEEE Trans. Geosci. Remote Sens. 2015, 54, 2627–2639. [Google Scholar] [CrossRef]
  16. Lin, T.H.; Li, H.T.; Tsai, K.C. Implementing the Fisher’s Discriminant Ratio in ak-Means Clustering Algorithm for Feature Selection and Data Set Trimming. J. Chem. Inf. Comput. Sci. 2004, 44, 76–87. [Google Scholar] [CrossRef]
  17. Alimjan, G.; Sun, T.; Liang, Y.; Jumahun, H.; Guan, Y. A new technique for remote sensing image classification based on combinatorial algorithm of SVM and KNN. Int. J. Pattern Recognit. Artif. Intell. 2018, 32, 1859012. [Google Scholar] [CrossRef]
  18. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  19. Lu, H.; Zhang, M.; Xu, X.; Li, Y.; Shen, H.T. Deep fuzzy hashing network for efficient image retrieval. IEEE Trans. Fuzzy Syst. 2020, 29, 166–176. [Google Scholar] [CrossRef]
  20. Wang, J.; Wang, D.; Wang, S.; Li, W.; Song, K. Fault Diagnosis of Bearings Based on Multi-Sensor Information Fusion and 2D Convolutional Neural Network; IEEE: New York, NY, USA, 2021; Volume 9, pp. 23717–23725. [Google Scholar]
  21. Qiu, S.; Zhao, H.; Jiang, N.; Wang, Z.; Liu, L.; An, Y.; Zhao, H.; Miao, X.; Liu, R.; Fortino, G. Multi-Sensor Information Fusion Based on Machine Learning for Real Applications in Human Activity Recognition: State-of-the-Art and Research Challenges; Elsevier: Amsterdam, The Netherlands, 2022; Volume 80, pp. 241–265. [Google Scholar]
  22. Zhang, W.; Tang, P.; Zhao, L.; Huang, Q. A comparative study of U-nets with various convolution components for building extraction. In Proceedings of the 2019 Joint Urban Remote Sensing Event (JURSE), Vannes, France, 22–24 May 2019; IEEE: New York, NY, USA; pp. 1–4. [Google Scholar]
  23. Deng, Z.; Zhu, X.; He, Q.; Tang, L. Land use/land cover classification using time series Landsat 8 images in a heavily urbanized area. Adv. Space Res. 2019, 63, 2144–2154. [Google Scholar] [CrossRef]
  24. Wu, C.; Jia, W.; Yang, J.; Zhang, T.; Dai, A.; Zhou, H. Economic Fruit Forest Classification Based on Improved U-Net Model in UAV Multispectral Imagery. Remote Sens. 2023, 15, 2500. [Google Scholar] [CrossRef]
  25. Schiefer, F.; Kattenborn, T.; Frick, A.; Frey, J.; Schall, P.; Koch, B.; Schmidtlein, S. Mapping forest tree species in high resolution UAV-based RGB-imagery by means of convolutional neural networks. ISPRS J. Photogramm. Remote Sens. 2020, 170, 205–215. [Google Scholar] [CrossRef]
  26. Li, G.; Han, W.; Huang, S.; Ma, W.; Ma, Q.; Cui, X. Extraction of sunflower lodging information based on UAV multi-spectral remote sensing and deep learning. Remote Sens. 2021, 13, 2721. [Google Scholar] [CrossRef]
  27. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  28. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  29. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  30. Fan, X.; Li, X.; Yan, C.; Fan, J.; Yu, L.; Wang, N.; Chen, L. MARC-Net: Terrain Classification in Parallel Network Architectures Containing Multiple Attention Mechanisms and Multi-Scale Residual Cascades. Forests 2023, 14, 1060. [Google Scholar] [CrossRef]
  31. Du, J.; Zhou, H.; Jacinthe, P.A.; Song, K. Retrieval of lake water surface albedo from Sentinel-2 remote sensing imagery. J. Hydrol. 2023, 617, 128904. [Google Scholar] [CrossRef]
  32. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  33. Reddi, S.J.; Kale, S.; Kumar, S. On the convergence of adam and beyond. arXiv 2019, arXiv:1904.09237. [Google Scholar]
Figure 1. Location of the study area and Sentinel-2 remote sensing image (6 October 2021).
Figure 1. Location of the study area and Sentinel-2 remote sensing image (6 October 2021).
Remotesensing 15 03924 g001
Figure 2. Schematic diagram of the CAMP-Net architecture for multispectral image classification tasks. Where * stands for multiplication.
Figure 2. Schematic diagram of the CAMP-Net architecture for multispectral image classification tasks. Where * stands for multiplication.
Remotesensing 15 03924 g002
Figure 3. Schematic diagram of the structure of the multilayer perceptron based on feature channels: (a) MLP block. (b) MLP branch structure.
Figure 3. Schematic diagram of the structure of the multilayer perceptron based on feature channels: (a) MLP block. (b) MLP branch structure.
Remotesensing 15 03924 g003
Figure 4. Channel Attention Module.
Figure 4. Channel Attention Module.
Remotesensing 15 03924 g004
Figure 5. Plot of the classification results of CAMP-Net using different combinations of modules for the study area dataset: (a) Image. (b) ViT. (c) CAMP-Net (NDVI + NDWI). (d) CAMP-Net (MLP). (e) CAMP-Net (CA). (f) CAMP-Net (MLP + NDVI + NDWI). (g) CAMP-Net (CA + NDVI + NDWI). (h) CAMP-Net (CA + MLP). (i) CAMP-Net (CA + MLP + NDVI + NDWI).
Figure 5. Plot of the classification results of CAMP-Net using different combinations of modules for the study area dataset: (a) Image. (b) ViT. (c) CAMP-Net (NDVI + NDWI). (d) CAMP-Net (MLP). (e) CAMP-Net (CA). (f) CAMP-Net (MLP + NDVI + NDWI). (g) CAMP-Net (CA + NDVI + NDWI). (h) CAMP-Net (CA + MLP). (i) CAMP-Net (CA + MLP + NDVI + NDWI).
Remotesensing 15 03924 g005
Figure 6. Plots of the results of different proportions of training samples: (a) 10%. (b) 20%. (c) 30%. (d) 40%. (e) 50%. (f) 60%. (g) 70%. (h) 80%. (i) 90%.
Figure 6. Plots of the results of different proportions of training samples: (a) 10%. (b) 20%. (c) 30%. (d) 40%. (e) 50%. (f) 60%. (g) 70%. (h) 80%. (i) 90%.
Remotesensing 15 03924 g006
Figure 7. ViT results in different bands: (a) 4 bands. (b) 4 bands + VNIR. (c) 4 bands + SWIR. CAMP-Net results in different bands. (d) 4 bands. (e) 4 bands + VNIR. (f) 4 bands + SWIR.
Figure 7. ViT results in different bands: (a) 4 bands. (b) 4 bands + VNIR. (c) 4 bands + SWIR. CAMP-Net results in different bands. (d) 4 bands. (e) 4 bands + VNIR. (f) 4 bands + SWIR.
Remotesensing 15 03924 g007
Figure 8. Classification results obtained by different models on South County data: (a) Image. (b) SVM. (c) KNN. (d) RF. (e) CNN. (f) RNN. (g) Transformer (ViT). (h) SF. (i) CAMP-Net.
Figure 8. Classification results obtained by different models on South County data: (a) Image. (b) SVM. (c) KNN. (d) RF. (e) CNN. (f) RNN. (g) Transformer (ViT). (h) SF. (i) CAMP-Net.
Remotesensing 15 03924 g008
Figure 9. Classification results obtained by different models on Xiangyin County dataset: (a) Image. (b) SVM. (c) KNN. (d) RF. (e) CNN. (f) RNN. (g) Transformer (ViT). (h) SF. (i) CAMP-Net.
Figure 9. Classification results obtained by different models on Xiangyin County dataset: (a) Image. (b) SVM. (c) KNN. (d) RF. (e) CNN. (f) RNN. (g) Transformer (ViT). (h) SF. (i) CAMP-Net.
Remotesensing 15 03924 g009
Figure 10. True color maps of the three images: (a) 2018, (b) 2020, and (c) 2022; spatial distribution of features in the three images: (d) 2018, (e) 2020, and (f) 2022.
Figure 10. True color maps of the three images: (a) 2018, (b) 2020, and (c) 2022; spatial distribution of features in the three images: (d) 2018, (e) 2020, and (f) 2022.
Remotesensing 15 03924 g010
Figure 11. Spatial distribution of pond in three images: (a) 2018, (b) 2020, and (c) 2022; spatial distribution of rapeseed in three images (d) 2018, (e) 2020, and (f) 2022; spatial distribution of vegetable in three images (g) 2018, (h) 2020, and (i) 2022.
Figure 11. Spatial distribution of pond in three images: (a) 2018, (b) 2020, and (c) 2022; spatial distribution of rapeseed in three images (d) 2018, (e) 2020, and (f) 2022; spatial distribution of vegetable in three images (g) 2018, (h) 2020, and (i) 2022.
Remotesensing 15 03924 g011
Figure 12. Dynamics of rapeseed in 2018–2020 (a), 2020–2022 (b), and 2018–2022 (c). Dynamics of vegetable in 2018–2020 (d), 2020–2022 (e), and 2018–2022 (f).
Figure 12. Dynamics of rapeseed in 2018–2020 (a), 2020–2022 (b), and 2018–2022 (c). Dynamics of vegetable in 2018–2020 (d), 2020–2022 (e), and 2018–2022 (f).
Remotesensing 15 03924 g012
Table 1. Sentinel 2 image band information.
Table 1. Sentinel 2 image band information.
Wave BandResolutionCenter WavelengthDescriptive
B160 m443 nmultramarine
B210 m490 nmblue
B310 m560 nmgreener
B410 m665 nmred
B520 m705 nmVNIR
B620 m740 nmVNIR
B720 m783 nmVNIR
B810 m842 nmVNIR
B8A20 m865 nmVNIR
B960 m940 nmSWIR
B1060 m1375 nmSWIR
B1120 m1610 nmSWIR
B1220 m2190 nmSWIR
Table 2. Sample size of the study area 2021 data set.
Table 2. Sample size of the study area 2021 data set.
NO.ClassSouth CountyXiangyin Xian
TrainingTestingTrainingTesting
1Building133057027621185
2Water77533337381603
3Lotus16767191374589
4Pond3079132029201252
5Wetland13045591407603
6Vegetable17597551768759
7Rapeseed16317002107903
Total11,554495616,0766894
16,51022,970
Table 3. Results of ablation experiments on the study area dataset using different combinations of modules of CAMP-Net.
Table 3. Results of ablation experiments on the study area dataset using different combinations of modules of CAMP-Net.
Different MethodsDifferent ModuleMetric
CAMLPNDVI + NDWIOA (%)AA (%)Kappa
ViT×××95.8196.0594.98
CAMP-Net××96.7697.1496.12
CAMP-Net××96.9197.1896.30
CAMP-Net×98.1598.2997.78
CAMP-Net×98.3398.4098.00
CAMP-Net×98.6398.8798.36
CAMP-Net99.0099.0898.81
Where × means that the part is not used, and the check mark indicates that the part is used.
Table 4. Effect of the number of neighboring bands on ViT as well as CAMP-Net.
Table 4. Effect of the number of neighboring bands on ViT as well as CAMP-Net.
The Number of Neighboring Bands  Class No.Metrics
1234567OA (%)AA (%)Kappa
1 (ViT)96.2497.6793.2595.9098.6994.8295.7695.8196.0594.98
1 (CAMP)98.0498.9696.5996.5599.4695.8497.3097.2297.5496.67
2 (ViT)97.5198.5894.2797.8598.4696.5397.7397.2097.2896.64
2 (CAMP)99.3998.7097.6195.1999.6196.0799.1497.4697.9696.96
3 (ViT)98.6498.9696.4296.9799.8494.6598.0997.3597.6696.83
3 (CAMP)97.5199.2294.9297.8599.4697.2798.5297.6797.8397.21
4 (ViT)97.9698.9693.6195.2998.7797.4999.2096.8897.3396.27
4 (CAMP)97.9699.2295.8896.7899.5396.4196.6297.1997.4996.63
5 (ViT)97.5997.5495.8896.9199.1597.2198.5297.4197.5596.90
5 (CAMP)98.9498.8395.5297.5998.6996.7095.6497.2597.4296.70
6 (ViT)98.5797.8096.0096.4298.0897.0495.4096.8497.0596.21
6 (CAMP)99.9299.6198.2199.1599.4697.8999.3599.0099.0898.81
7 (ViT)98.4999.3595.6497.7298.7795.6796.5097.2697.4596.71
7 (CAMP)98.8799.3595.2897.1099.5397.8998.4697.7898.0897.35
Table 5. CAMP-Net test results using different training sample proportions in the 2021 South County dataset.
Table 5. CAMP-Net test results using different training sample proportions in the 2021 South County dataset.
Ratio of TrainingClass No.Metrics
1234567OA (%)AA (%)Kappa
10%98.4293.6397.4897.2698.3893.6299.1497.0396.8596.44
20%98.6898.1997.0797.611.0093.6296.1397.1597.3396.59
30%99.8299.6995.2697.4999.1097.3499.1497.9898.2797.58
40%99.0799.5497.0797.4999.8697.1197.7498.0098.2897.60
50%98.5299.6395.5797.2798.9296.3498.9697.6197.8997.14
60%98.4299.6991.4497.4299.7398.1498.7897.3897.6696.87
70%99.9299.6198.2199.1599.4697.8999.3599.0099.0898.81
80%99.4099.7796.6098.2399.5996.1298.0698.0598.2697.66
90%98.7199.4996.8997.3799.8296.2498.8597.9198.2097.50
Table 6. Test results in the 2021 South County dataset using different methods in different bands.
Table 6. Test results in the 2021 South County dataset using different methods in different bands.
Different Bands (Method) Class No.Metrics
1234567OA (%)AA (%)Kappa
4 bands (ViT)76.6991.6179.8982.2693.3291.5288.7785.4986.3082.59
4 bands (CAMP-Net)79.5492.3883.8386.0692.5691.6488.3587.3287.7784.77
4 bands + VNIR (ViT)91.9594.3292.7889.0597.7796.4193.0792.9593.6391.56
4 bands + VNIR (CAMP-Net)93.0896.6493.4990.6198.8498.4695.4694.5395.2393.45
4 bands + SWIR (ViT)96.2497.6793.2595.9098.6994.8295.7695.8196.0594.98
4 bands + SWIR (CAMP-Net)99.9299.6198.2199.1599.4697.8999.3599.0099.0898.81
Table 7. Comparison of classification results of different classification methods on the South County 2021 data set.
Table 7. Comparison of classification results of different classification methods on the South County 2021 data set.
C N.Different Methods
SVMKNNRFCNNRNNViTSFCAMP-Net
187.8982.4589.4770.2297.4496.2499.3999.92
289.1896.9996.0987.4897.0397.6799.4899.61
385.5390.4091.2377.8086.4593.2595.1098.21
489.1692.8792.1278.2094.8695.9097.2799.15
591.2397.3194.9978.9197.5498.6998.3199.46
692.8493.6492.8472.6595.2894.8297.1097.89
790.0092.7191.4285.1695.7095.7697.6099.35
OA (%)89.4192.1992.2978.0794.5795.8197.4999.00
AA (%)89.4192.3492.6078.6494.9196.0597.7599.08
Kappa87.2990.6290.7673.6793.5094.9896.9998.81
Table 8. Comparison of classification results of different classification methods on Xiangyin County 2021 data set.
Table 8. Comparison of classification results of different classification methods on Xiangyin County 2021 data set.
C N.Different Methods
SVMKNNRFCNNRNNViTSFCAMP-Net
195.1890.1295.5281.1394.6097.2496.7099.56
297.6999.6299.5695.3199.2798.7499.3399.97
396.2696.7797.4586.1798.0399.6399.34100.00
484.6689.9390.3377.3992.9493.1895.5898.69
576.1191.8788.8849.1897.7591.3294.1097.58
694.3398.0296.3178.0596.3898.3597.6299.26
797.3499.5598.0090.3699.4799.1999.71100.00
OA (%)92.4795.1395.5282.2696.2796.9297.6199.39
AA (%)91.6695.1395.1679.6696.0796.8197.4999.30
Kappa91.0294.1994.6678.8395.5596.3397.1599.27
Table 9. Sample sizes for the 2018, 2020, and 2022 datasets for South County.
Table 9. Sample sizes for the 2018, 2020, and 2022 datasets for South County.
No.Class201820202022
TrainingTestingTrainingTestingTrainingTesting
1Building9073899073891187509
2Water71630771630734661486
3Lotus6112636112631484636
4Pond120051512005151451622
5Wetland471203471203993426
6Vegetable11364881136488991426
7Rapeseed7833367833361041447
Total582425015824250110,6134552
8325832515,165
Table 10. Comparison of classification results of different classification methods on the South County 2018 data set.
Table 10. Comparison of classification results of different classification methods on the South County 2018 data set.
C N.Different Methods
SVMKNNRFCNNRNNViTSFCAMP-Net
191.2586.3789.7177.9491.6292.1794.1597.68
296.0997.7196.0988.6898.6098.7497.7699.44
391.6393.1593.9184.6192.1491.9894.5995.09
494.7595.5398.4483.4196.4197.3398.3398.08
598.52100.00100.0091.7199.3699.5799.78100.00
696.1097.7496.9281.8698.5999.2098.1599.47
788.0991.3689.7965.7793.4896.2996.0497.70
OA (%)93.7294.3694.7281.3495.7696.5597.0098.25
AA (%)93.7894.5694.7082.0095.7596.4796.9898.21
Kappa92.5693.3293.7477.8794.9895.9196.4497.93
Table 11. Comparison of classification results of different classification methods on the South County 2020 data set.
Table 11. Comparison of classification results of different classification methods on the South County 2020 data set.
C N.  Different Methods
SVMKNNRFCNNRNNViTSFCAMP-Net
187.9183.8088.9467.4792.8297.2496.4799.00
297.7199.3498.6995.8197.9099.8699.30100.00
358.5585.5182.5044.1881.0186.7485.5994.43
478.4482.1381.9475.5086.5891.3394.6697.25
598.0297.5397.0489.1799.1599.78100.0099.36
673.7773.9777.6666.5484.1589.9688.6495.42
764.2877.3878.5738.5680.4587.9992.4694.89
OA (%)78.9883.8985.0967.8688.0092.7993.5397.06
AA (%)79.8285.6886.4868.1888.7993.2893.8897.20
Kappa75.0180.9482.3461.5785.7891.4692.3396.52
Table 12. Comparison of the classification results of different classification methods on the South County 2022 data set.
Table 12. Comparison of the classification results of different classification methods on the South County 2022 data set.
C N.Different Methods
SVMKNNRFCNNRNNViTSFCAMP-Net
194.4993.5195.6777.5096.9698.7399.4999.91
2100.0099.93100.0099.8299.6599.79100.0099.82
393.5593.3995.2888.2093.7395.9598.4599.12
495.1795.1795.3385.2597.3196.0798.8299.72
592.0194.3694.6090.4398.9998.9999.3998.69
670.1889.9090.8441.9786.7890.1186.1796.36
799.3299.3299.3299.3299.80100.00100.00100.00
OA (%)94.2296.1396.7987.3896.9697.6798.2299.31
AA (%)92.1195.0995.8783.2296.1897.1097.4899.09
Kappa92.9095.2696.0684.4996.2797.1597.8199.16
Table 13. Table of land use type changes in the study area for different periods from 2018 to 2022.
Table 13. Table of land use type changes in the study area for different periods from 2018 to 2022.
ClassArea (km 2 )Area Change Rate (%)
2018202020222018–20202020–20222018–2022
Building26.7925.8626.15−0.030.01−0.02
Water4.024.003.65−0.01−0.09−0.09
Lotus3.6211.8523.312.270.975.43
Pond24.4224.8225.270.020.020.03
Wetland7.575.496.93−0.270.26−0.08
Vegetable41.8337.9448.43−0.090.280.16
Rapeseed20.6918.9919.72−0.080.04−0.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fan, X.; Li, X.; Yan, C.; Fan, J.; Chen, L.; Wang, N. Converging Channel Attention Mechanisms with Multilayer Perceptron Parallel Networks for Land Cover Classification. Remote Sens. 2023, 15, 3924. https://doi.org/10.3390/rs15163924

AMA Style

Fan X, Li X, Yan C, Fan J, Chen L, Wang N. Converging Channel Attention Mechanisms with Multilayer Perceptron Parallel Networks for Land Cover Classification. Remote Sensing. 2023; 15(16):3924. https://doi.org/10.3390/rs15163924

Chicago/Turabian Style

Fan, Xiangsuo, Xuyang Li, Chuan Yan, Jinlong Fan, Lin Chen, and Nayi Wang. 2023. "Converging Channel Attention Mechanisms with Multilayer Perceptron Parallel Networks for Land Cover Classification" Remote Sensing 15, no. 16: 3924. https://doi.org/10.3390/rs15163924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop