Next Article in Journal
SCRM-Net: Self-Supervised Deep Clustering Feature Representation for Urban 3D Mesh Semantic Segmentation
Next Article in Special Issue
Large-Scale Hyperspectral Image-Projected Clustering via Doubly Stochastic Graph Learning
Previous Article in Journal
Far-Field Earthquake-Induced Crustal Deformation and Mud Volcano Activity in Azerbaijan Based on the InSAR Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Feature Lightweight DeeplabV3+ Network for Polarimetric SAR Image Classification with Attention Mechanism

1
School of Computer Science and Technology, Xi’an University of Technology, Xi’an 710048, China
2
Shaanxi Key Laboratory for Network Computing and Security Technology, Xi’an University of Technology, Xi’an 710048, China
3
School of Electronic Engineering, Xidian University, No. 2 South TaiBai Rood, Xi’an 710071, China
4
School of Computer Science and Engineering, Nanyang Technological University, Singapore 639798, Singapore
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(8), 1422; https://doi.org/10.3390/rs17081422
Submission received: 19 February 2025 / Revised: 6 April 2025 / Accepted: 11 April 2025 / Published: 16 April 2025
(This article belongs to the Special Issue Remote Sensing Image Classification: Theory and Application)

Abstract

:
Polarimetric Synthetic Aperture Radar (PolSAR) is an advanced remote sensing technology that provides rich polarimetric information. Deep learning methods have been proved an effective tool for PolSAR image classification. However, relying solely on source data input makes it challenging to effectively classify all land cover targets, especially heterogeneous targets with significant scattering variations, such as urban areas and forests. Besides, multiple features can provide more complementary information, while feature selection is crucial for classification. To address these issues, we propose a novel attention mechanism-based multi-feature lightweight DeeplabV3+ network for PolSAR image classification. The proposed method integrates feature extraction, learning, selection, and classification into an end-to-end network framework. Initially, three kinds of complementary features are extracted to serve as inputs to the network, including polarimetric original data, statistical and scattering features, textural and contour features. Subsequently, a lightweight DeeplabV3+ network is designed to conduct multi-scale feature learning on the extracted multidimensional features. Finally, an attention mechanism-based feature selection module is integrated into the network model, adaptively learning weights for multi-scale features. This enhances discriminative features but suppresses redundant or confusing features. Experiments are conducted on five real PolSAR data sets, and experimental results demonstrate the proposed method can achieve more precise boundaries and smoother regions than the state-of-the-art algorithms. In this paper, we develop a novel multi-feature learning framework, achieving a fast and effective classification network for PolSAR images.

1. Introduction

Polarimetric synthetic aperture radar (PolSAR) is an active microwave remote sensing detection technology, which is capable of obtaining multi-channel back-scattered echoes from the Earth’s surface. It possesses the capability of imaging in all weather and day-night. In comparison to Synthetic Aperture Radar (SAR) system [1], PolSAR system can capture a broader range of polarimetric characteristics and target scattering information, which has found widespread applications in various fields, including land cover classification [2], target identification [3], and ship detection [4] etc. The classification of PolSAR image is a crucial prerequisite in comprehending and interpreting PolSAR data, thus attracting increasing attention from researchers in the field of remote sensing.
Over the decades, numerous classification methods for PolSAR images have been proposed by researchers. The conventional approaches predominantly exploit shallow features to classify PolSAR images, which can be broadly categorized into three types. The first category involves statistical characteristic-based classification methods, notably including Wishart distribution [5], KummerU distribution [6], as well as the Wishart mixture distribution model [7] etc. These methods are effective in modeling the statistical distribution of PolSAR data. The second category comprises classification methods based on target decomposition [8], such as model-based polarimetric decomposition [9], polarimetric covariance eigenvalues [10], dominant scattering mechanism identification [11] and polarimetric roll-invariant features [12] etc. These methods utilize various scattering features to describe land covers from different views. In addition, machine learning approaches have been widely applied to PolSAR image classification. Widely adopted methods include support vector machine [13] and random forest [14], etc. These techniques aim to take advantages of potent polarimetric features for effective classification. However, these methods are limited to learning shallow characteristics. They encounter challenges in accurately classifying heterogeneous land covers due to the absence of high-level semantic features.
Recently, deep learning has witnessed rapid development in the realm of PolSAR image classification. In contrast to traditional methods, deep learning eliminates the need of manual selection of image features by autonomously extracting deeper features. Typical deep learning models include deep boltzmann machines [15], stacked auto-encoder [16], deep belief network [17], convolutional neural network (CNN) [18] and graph convolutional network [19], among others. To be specific, Zhang et al. [20] introduced a convolutional neural network with complex-valued coding (CV-CNN), enhancing feature extraction capability specifically tailored for PolSAR images. Guo et al. [21] presented an adaptive fuzzy super-pixel algorithm based on polarimetric scattering information for PolSAR image classification. In addition, Zhang et al. [22] introduced transfer learning and unsupervised deep learning networks to improve the classification results. Ai et al. [23] proposed a convolutional auto-encoder that incorporated textural feature fusion for land cover classification in PolSAR image.
With the evolution of deep learning models, various neural networks focusing on semantic segmentation have emerged successively. Notable networks in this domain include FCN [24], Unet [25], and DeepLab series (V1, V2, V3, V3+) [26,27,28,29]. Among them, the DeepLab series have achieved remarkable success in the field of semantic segmentation. Concurrently, in recent years, many PolSAR image classification algorithms based on the DeepLab series have been proposed. For instance, Zhang et al. [30] employed wavelet fusion and PCA dimensional reduction to suppress redundant features, combined with DeepLabV3 network for the classification of PolSAR images. In addition, Zhang et al. [31] incorporated the concept of SKNet units into DeepLabV3+, learning multi-scale polarimetric feature effectively, and thereby better applying to classification tasks. The DeepLabV3+ network [29] not only utilizes improved Xception network to integrate multi-dimensional information effectively, but also design spatial pyramid pooling (ASPP) module to learn multi-scale features. Therefore, DeepLabV3+ method can adaptive learn multi-scale features and semantic information, thereby exhibiting better performance for image classification. Here, we opt for the DeepLabv3+ network as the backbone network for PolSAR image classification. However, with too many parameters, DeepLabv3+ presents a complex network structure with a lot of time cost. To alleviate this problem, we design a lightweight DeepLabv3+ network tailored for PolSAR image classification, which can efficiently learn multi-scale semantic features for PolSAR image.
The aforementioned deep learning approaches empower the network to autonomously learn high-level features of targets, contributing to the enhanced accuracy of PolSAR image classification. However, these methods don’t fully exploit multiple scattering characteristics, which makes it challenging to effectively classify all land cover targets, particularly for heterogeneous ones such as urban areas and forests. It can be observed that significant scattering variations exists even within the interior of a heterogeneous ground object, making it difficult for the original data to categorize them into semantically consistent class. Actually, PolSAR data contain rich scattering features, target decomposition features, and texture features, which can provide complementary information for effective image classification [32,33]. For example, Wang et al. [34] introduced a PolSAR image classification method that jointly considered polarimetric features and adjacency information to gain classification advantage. Shi et al. [35] proposed a topic model for multi-feature joint learning, producing enhanced classification accuracy. Shang et al. [36] presented a CNN model based on spatial features, improving classification performance through a dual-branch network learning spatial features. Xu et al. [37] combined deep CNN with the Freeman-Durden decomposition algorithm to extract polarimetric features from PolSAR data, yielding more accurate result of building extraction. Wu et al. [38] employed a hybrid generative/discriminative network to learn the diversity of PolSAR echo statistical features and image spatial features, combined with variational Bayesian theory to integrate multiple features for PolSAR image classification. Collectively, these methods underscore the importance of leveraging multiple polarimetric features rather than relying solely on the original data.
However, existing deep learning classification methods have not fully exploited and utilized the diverse polarimetric information present in PolSAR images. In addition, it is widely acknowledged that an increased number of scattering features may not necessarily lead to better performance. Therefore, there are still two challenges for existing deep learning methods in PolSAR image classification. (1) How to maximize the utilization of the rich features embedded in PolSAR data to enhance the generalization and robustness of deep learning models. (2) How to adaptively select discriminative features from the multitude of available features, while simultaneously suppressing redundant or confusing features. Addressing these challenges is crucial for enhancing the overall classification performance of PolSAR images.
Fortunately, many feature selection methods have been proposed, such as correlation measurement [39], feature ensemble algorithm [40], dimensionality reduction [41], etc. Besides, the attention mechanism [42] has emerged as a widely used method for feature selection. It can automatically select interest regions and assign different weights for diverse regions, which has found extensive applications in target detection [43], image fusion [44] and classification [45]. However, it’s worth noting that these methods primarily operate on shallow feature selection, limiting their effectiveness in classifying complex terrain objects.
To alleviate these challenges, in this paper, we introduces a novel multi-feature lightweight deeplabV3+ network classification model, referred to as MLDnet. The model is an end-to-end deep network that integrates feature learning, feature selection, and classification together. Initially, we extract multiple scattering features from PolSAR images to provide a comprehensively representation of land cover information, encompassing original data, target decomposition features, textural and contour features. Subsequently, we design a lightweight L-DeeplabV3+ to train the network, facilitating the integration of multiple features and learning multi-scale high-level features, which significantly reduce network parameters. Additionally, an attention mechanism module is embedded after the L-DeeplabV3+ network. This module adaptively assigns weights to multidimensional features, enhancing relevant features for classification while suppressing redundant and confusing ones. Experimental results conducted on five real PolSAR data sets indicate that the proposed MFDAnet achieves higher classification accuracy compared to the state-of-the-art algorithms. The main contributions of the proposed MFDAnet method can be summarized into three key aspects.
(1)
A novel multi-feature deep attention network is proposed, which fully exploits diverse types of polarimetric features as the input, and automatically integrate the multi-feature learning, selection and classification into a unified framework. The multiple features include PolSAR original data, scattering features and image features, providing complementary information from different perspectives.
(2)
A lightweight DeeplabV3+ is developed as the backbone network of the MFDAnet. This architecture facilitates the learning of both multi-scale and multi-feature information, designing a lightweight scheme tailed for PolSAR images, ensuring a fast and effective network.
(3)
Attention mechanism-based feature selection module is embedded in the proposed MLDnet, adaptively learning weights of multi-dimensional features. This module enhances valuable features and suppresses redundant ones, thereby improving the classification performance.
The remaining article is organized as follows. Section 2 provides a detailed explanation of the proposed approach. Section 3 presents the experimental results on five real PolSAR datasets. Section 4 illustrates the discussion about parameter analysis and ablation study. Conclusion and further work are given in Section 5.

2. Proposed Method

In this paper, a novel multi-feature deep attention network (MLDnet) is proposed for PolSAR image classification. It can learn multi-scale and multiple scattering features adaptively by designing a lightweight multi-feature DeeplabV3+ network. Additionally, it selects effective features using the attention mechanism, so as to finally realize the the effective classification of different land cover types.
The framework of our proposed multi-feature deep attention network is shown in Figure 1, comprising three key components: multi-feature extraction, lightweight DeeplabV3+ network design and attention-based feature fusion and classification. Firstly, in contrast to traditional network using 9-dimension PolSAR data as the input, we extract multiple scattering features as the network input. They include original data, scattering mechanism-based and image analysis-based features, fostering enhanced information complement; Secondly, we design a lightweight DeeplabV3+ network to automatically learn features of different scales from multiple features; Finally, the multi-scale multi-features are automatically selected by the channel attention module embedded in the network. This process enhances valuable discriminating features with larger weights while disregarding useless features with smaller weights. The classification layer is then employed to determine the probability of the target, yielding the final classification result. Further details are introduced as follows.

2.1. Multi-Feature Extraction Module

Traditional CNN model usually converts the PolSAR covariance matrix into a 9-dimension real vector as the network input, creating a cube with 9 channels for each pixel. However, 9-dimension vector is not adequate to represent the complexity of PolSAR image for each pixel. Additional information, such as scattering characteristics [46,47], has been proved effective for identifying ground objects. Moreover, the contextual features, including textural and contour features, play a significant role in describing image contextual information. Numerous studies highlight the value of multiple features in enhancing classification performance. Therefore, we propose a multi-feature data representation as the network input. Before multi-feature extraction, a refined Lee filter [48] is applied to the original PolSAR image. Compare with other advanced filter methods [4,49], the refined Lee filter is a classical filtering method to reduce speckle noises fastly and effectively, which can be easily conducted upon the PolSARPro software. Specifically, three types of features are extracted to formulate a 57-dimensional feature vector as PolSAR data expression, which is extracted from PolSAR original data, target decomposition features, as well as textural and contour features respectively. The detailed feature extraction procedure is outlined below.
(1) PolSAR original data (16-dimension): It consists of scattering matrix S , coherency matrix T and SPAN image. The original PolSAR data is scattering matrix S , defined as:
S = S h h S h v S v h S v v
After multi-look processing, the most commonly used coherency matrix T is derived by Equation (2), which is shown on the top of the next page.
T = 1 2 S h h + S v v 2 S h h + S v v S h h S v v * 2 S h h + S v v S h v * S h h S v v S h h + S v v * S h h S v v 2 2 S h h S v v S h v * 2 S h v S h h + S v v * 2 S h v S h h S v v * 4 S h v 2
In Equation (2), ∗ represents the complex conjugate operation. Since the complex matrix cannot be fed to the network directly, similar to other networks, we convert the coherency matrix T to a 9-dimension vector, defined as
I T = [ T 11 , T 22 , T 33 , T 12 r e a l , T 13 r e a l , T 23 r e a l , T 21 i m a g , T 31 i m a g , T 32 i m a g , ]
where T 12 r e a l is the real part of T 12 , and T 12 i m a g is the imaginary part of T 12 . In addition, the Polarimetric-Total-Power (SPAN) value is the power image, which is calculated by summarizing the diagonal elements of T matrix, defined by
S P A N = S h h 2 + 2 S h v 2 + S v v 2
(2) Target decomposition-based features (17-dimension): It contains three target decomposition-based parameters and two polarimetric parameters, which can describe target’s scattering mechanism from various perspectives. Three types of target decompositions are Cloude and Pottier decomposition [50], Freeman decomposition [51] and Huynen decomposition [52]. These decomposition methods have been proven effective for PolSAR image classification, serving as valuable features for distinguishing different ground objects. Furthermore, the two polarimetric parameters are the co-polarization ratio [53] and the cross-polarization ratio [54]. The former is used to measure the difference between S h h and S v v , and the latter is highly sensitive to volume scattering. All these scattering features represent independent attributes of each resolution unit, providing distinct descriptions for each pixel.
(3) Image features (24-dimension): In addition to the original data and scattering features, each pixel, acting as a unique scattering unit, also has a close relationship with neighborhoods to formulate a PolSAR image, that is contextual features such as texture and contour. These features play a crucial role in identifying the ground objects. To capture the contextual relationship, we extract Gray-level Co-occurrence Matrix (GLCM) features [55], a widely used technique for textual feature extraction in natural images. It encompasses four types of features, including contrast, energy, entropy and relativity. Additionally, we extract the edge and line energies using filters with 18 directions and 4 scales, providing information on the contour and strength of image edges [56]. All 57-dimension features are summarized into Table 1.

2.2. Lightweight DeepLabV3+ Network

The DeepLab series [26,27,28,29], a collection of semantic segmentation algorithms developed by the Google team, has demonstrated notable advancements in semantic learning. Leveraging depth-wise separable convolution technology, DeeplabV3+ efficiently integrates features in the depth direction, requiring fewer parameters. In addition, the DeepLabV3+ network introduces the Atrous Spatial Pyramid Pooling (ASPP) technique to integrate multi-scale information. This technique utilizes convolution layers with different dilation factors, increasing the receptive field and reducing information loss during down sampling. Therefore, we select the DeeplabV3+ network as the backbone of the proposed method, since it can integrate multi-dimensional features and learn multi-scale semantic information effectively.
However, the original backbone network used by DeeplabV3+ network is an improved Xception network, as depicted in Figure 2, featuring a complex network structure. Furthermore, in semantic segmentation tasks, the input unit is typically an image, while for PolSAR image classification, the input unit is a small image block, with a smaller input feature block and a larger data amount. So, considering these differences between semantic segmentation and pixel-wise classification, a revised DeeplabV3+ network should be tailored for PolSAR image classification.
In this paper, we develop a lightweight DeeplabV3+ network, named L-DeeplabV3+, as an alternative. Taking a PolSAR image with 100 × 100 pixels as an example, Table 2 provides the comparison of network parameters between the L-DeeplabV3+ network used in this article and the original DeeplabV3+network. In addition, Figure 3 also illustrates the structure differences between the proposed L-DeeplabV3+ and original DeeplabV3+ networks, which can be summarized into three key aspects.
(1) Input: First of all, as mentioned earlier, the whole image serves as the network input in the semantic segmentation task in Figure 3a. However, for PolSAR image classification task, the input image block is sampled by a square window centered on each pixel, as shown in Figure 3b. It also can be found in Table 2 that the size of the input feature map are 100 × 100 (image size) and 16 × 16 (window size) for DeeplabV3+ and L-DeeplabV3+ networks respectively.
(2) Backbone network: For the backbone network, the DeeplabV3+ network employs an improved Xception network, as shown in Figure 2, adept at handling larger input features. However, it may lead to overfitting when dealing with small input blocks. As depicted in Table 2, the proposed L-DeeplabV3+ network adopts two depth-wise separable convolutional layers(shorted by “DWConv”) as the backbone network. This design allows for the extraction of shallow features without significantly reducing the feature dimension. In addition, we introduce a max-pooling layer between the two DWConv layers to further reduce the number of model parameters and mitigate the risk of model overfitting.
(3) Multi-scale feature learning: In addition, as is well known, with the increasing dilated rate, the receptive field will also become larger. Therefore, in the original DeeplabV3+, a relatively large dilation rate is set to achieve the segmentation of large objects. However, for a PolSAR image block, an excessively large dilation rate can lead to the loss of local detailed information, thereby affecting the final classification results. In L-DeeplabV3+, we address this issue by significantly reducing the dilation rate in the ASPP module. This adjustment allows the network to capture multi-scale contextual information while preserving local detailed information. The detailed parameter comparison can be seen in Table 2.
In summary, the proposed multi-feature-based L-DeeplabV3+ network achieves multi-scale feature learning for the multi-dimensional features of PolSAR images. This is accomplished while maintaining the original DeeplabV3+ network framework and further reducing network parameters and unnecessary time overhead.

2.3. Attention-Based Multi-Feature Fusion and Classification

After processing through L-DeeplabV3+ network, we attain multi-scale multi-feature information in channel dimension, which can provide rich contextual features for classification. However, recognizing that not all scattering features contribute equally to classification, the selective extraction of valuable features while suppressing less relevant ones remains a challenging issue. In this paper, we ensemble an attention-based multi-feature fusion module with the L-deeplabV3+ network. This module dynamically learns different weights for diverse scattering features in channel direction, highlighting important features and diminishing less useful ones. Ultimately, we achieve multi-scale multi-feature fusion in channel dimension.
The multi-feature fusion module utilizes the attention mechanism [42] to adaptive weight the channel features, illustrated in Figure 4. Channel attention pays attention on exploring “what” is the meaningful features for the input image on the channel dimension. It begins by separately aggregating multi-scale spatial features through max-pooling and average-pooling layers. This process generates two distinct feature maps, each encapsulating spatial contextual information. Subsequently, these two feature maps are input to a shared fully connection layer, producing their respective channel attention feature maps. The next step involves summing these two attention feature maps, followed by the application of an activation function. This results in the aggregation of their information, ultimately yielding channel attention weights.
Assuming that the size of the output feature map X by L-deeplabV3+ network is M × N × C , the obtained channel attention weight is W, with the size of 1 × 1 × C , through the attention mechanism. The procedure of the channel attention module can be described as:
W = s i g m o i d ( W 2 · R e l u W 1 · A v g P o o l X + W 2 · R e l u W 1 M a x P o o l X )
where X is the learned feature map from L-DeeplabV3+, W 1 and W 2 are the parameters of two fully connecting layers and R e L u is the non-linear translation. A v g P o o l · and M a x P o o l · are average and max pooling layers respectively. S i g m o i d is the activation function.
The feature map X H × W × C is then multiplied by the weight W 1 × 1 × C to obtain the channel attention fusion features. W is firstly padded to H × W × C by p a d ( . ) function, and then X is element-wise multiplied by W to obtain weighted features, defined as:
F = X p a d ( W )
where ⊗ represents element-wise multiplication. After obtained the fused feature F, a s o f t m a x classifier is applied to achieve the final classification map.
The whole algorithm procedure of the proposed MLDnet is given in Algorithm 1. The algorithm begins by applying the refined Lee filter to mitigate speckle noise. Subsequently, multiple features are extracted from PolSAR images, including original PolSAR data, target decomposition features, textual and contour features. These multi-features are then input into the L-DeeplabV3+ network to learn the multi-scale high-level features. Furthermore, an attention mechanism is applied to the multi-scale features, enabling the automatic selection and fusion of multiple features by learning different weights. This ensures that valuable features are assigned larger weights while redundant features receive smaller weights. Finally, the output of the MLDnet passes through a s o f t m a x classifier to obtain the final classification result.
Algorithm 1 Procedure of the proposed MLDnet method.
Input: PolSAR original data S , PolSAR original image P , class label map L and class number C .
Step 1: Apply a refined Lee filter to PolSAR original image P to obtain the filtered PolSAR image P L .
Step 2: Extract 57-dimensional features F from PolSAR image P L and PolSAR original data S based on Table 1.
Step 3: Using a fixed-size square window h × h to sample the extracted 57-dimensional features F pixel by pixel, obtaining N data with a shape of h × h × 57 , where N is the total number of image pixels.
Step 4: According to a certain ratio, the N data obtained in step 3 and the label map L are divided into training set R and test set T .
Step 5: Import the training set R into the MLDnet network for training until reaching the iteration number, saving the training model, training loss, and model parameters.
Step 6: Using the trained model to predict and classify the test set T .
Output: class label estimation map Y and various evaluation indicators.

3. Experimental Results

3.1. Experimental Data and Settings

In order to verify classification performance of the proposed MLDnet, extensive experiments are conducted on five real PolSAR data sets. These data sets cover Xi’an, Oberpfaffenhofen, San Francisco and Flevoland areas respectively. Detailed information is given as follows:
(A) Xi’an data set: The first data set is C-band full polarimetric SAR data collected from Xi’an in the western region of China using RADARSAT-2 system. The image size is 512 × 512 pixels, with a resolution of 8 × 8 m. It encompasses three types of land cover classes: building, grass, and water. The Pauli-RGB pseudo-color image and the corresponding ground truth image are presented in Figure 5a.
(B) Oberpfaffenhofen data set: The second data set covers the Oberpfaffenhofen area in German and consists of an 8-look L-band fully PolSAR image acquired by the E-SAR sensor at the German Aerospace Center. This polarimetric radar image has the dimension of 1300 × 1200 pixels with a resolution of 3 × 2.2 m. It includes five distinct land cover types: bare ground, forest, farmland, road and building. Figure 5b illustrates the Pauli RGB image of the Oberpfaffenhofen data set along with the corresponding ground truth map.
(C) San Francisco data set: The third data set is the San Francisco data set, which is a C-band fully PolSAR image collected by the RADARSAT-2 system over the San Francisco area in the United States. The image has a resolution of 10 × 5 m and a size of 1800 × 1380 pixels. Its Pauli RGB image and ground truth map are displayed in Figure 5c, which contains five different land cover types: ocean, vegetation, low density urban, high density urban, and developed.
(D) Flevoland1 data set: The fourth data set is C-band fully polarimetric SAR data collected by the RADARSET-2 sensor in the Flevoland region of the Netherlands. The image size is 1400 × 1200 pixels, with a resolution of 12 × 8 m. The PolSAR image encompasses four types of land cover, including water, urban, woodland, and cropland. We name this data set the Flevoland1 data set. Figure 5d exhibits the Pauli-RGB image along with the corresponding ground truth map of the Flevoland1 data set.
(E) Flevoland2 data set: The last data set comes from the Flevoland region and consists of full-polarization L-band SAR data acquired by the AIRSAR system. The spatial resolution is 12.1 × 6.7 m, and the image size is 750 × 1024 pixels. The Pauli RGB image and its ground truth are shown in Figure 5e. This image includes 15 types of crops: stem beans, peas, forest, lucerne, beets, wheat, potatoes, bare soil, grasses, rapeseed, barley, wheat2, wheat3, water and buildings. We name it the Flevoland2 data set.
The quantitative metrics employed in our experiments include average accuracy (AA), overall accuracy (OA), and Kappa coefficient. Experiments are conducted on Windows 10 operating system, using a 64 GB NVIDIA GeForce GTX 3070 PC. The algorithm is implemented with Python 3.7 and Pytorch GPU 1.9.0. In addition, we utilize the Adam optimizer with a learning rate of 0.0001. The batch size is defined as 16 × 16 , and the number of training iteration is 50. The multi-class cross-entropy loss is employed. During the experimental process, 10% of the training samples are selected for each class of land cover in the image, with the remaining 90% serving as test samples. It should be noted that the white regions in the ground truth maps are unlabeled pixels in Figure 5. During classification, all pixels in PolSAR images can obtain classification result, while unlabeled pixels are not considered into calculating quantitative metrics. Therefore, we just give the classification map with labeled pixels, and unlabeled pixels are colored by white in the following experimental results.
To assess the effectiveness of the proposed MLDnet, we compare it with seven state-of-the-art classification algorithms: Super-RF [14], CNN [18], CV-CNN [20], 3D-CNN [57], PolMPCNN [58], CEGCN [59] and SGCN-CNN [60]. Specifically, the first compared method combines the random forest algorithm with superpixels, utilizing G 0 statistical texture features to mitigate the interference from background targets, such as forests, on farmland classification, referred to as Super-RF. The second method, referred to as CNN, converts PolSAR data into normalized 9-D real feature vectors and learns hierarchical polarimetric spatial features automatically through two cascaded convolutional layers. The third method is a complex-valued convolutional neural network, named by CV-CNN. This approach transforms the PolSAR complex matrix into a complex vector, effectively leveraging both the amplitude and phase information presented in PolSAR images. The fourth method is based on 3-D convolutional neural network, shorted by 3D-CNN. It utilizes 3-D CNN to extract deep channel-spatial combined features, adapting well to the 3D data structure for classification. The fifth method combines polarimetric rotation kernels, employs a multi-path structure and dual-scale sampling to adaptively learn the polarimetric rotation angles of different land cover types, noted by PolMPCNN. The sixth method employs a graph encoder and decoder to facilitate collaboration between CNN and GCN within a single network, shorted by CEGCN. It enables feature learning in small-scale regular and large-scale irregular regions, respectively. The last approach similarly integrates super-pixel level and pixel-level networks into a single framework, named SGCN-CNN for short. It can obtain both global features and local features by defining the correlation matrix for feature transformation between superpixel and pixel.

3.2. Experimental Results on Xi’an Data Set

Classification results on the Xi’an data set are displayed in Figure 6b–i, illustrating the classification maps generated by Super-RF, CNN, CV-CNN, 3D-CNN, PolMPCNN, CEGCN, SGCN-CNN and the proposed MLDnet methods, respectively. Specifically, Figure 6b displays the output of the Super-RF method, revealing numerous misclassified pixels, primarily concentrated around the water class. At the same time, these misclassifications form distributed blocks due to superpixels. In contrast, the CNN method performs exceptionally well in distinguishing the water class, but exhibits misclassifications in the building class due to the lack of global information. It erroneously classifies some building as grass class. Similar to the CNN method, the CV-CNN exhibits noticeable misclassifications in the building and grass classes due to pixel-wise local features. 3D-CNN can adapt to the 3D structure of input data for classification, thus achieving better regional consistency. However, it struggles to adaptively handle distinct individual features, limiting its classification effectiveness at the boundaries of different land covers. In comparison, PolMPCNN excels in adaptively learning the polarimetric characteristics of various land covers, greatly improving overall classification effect. However, the feature learning procedures of different terrains are independent, resulting in a small number of misclassified pixels still existing at the land covers boundaries. Compared to PolMPCNN, the CEGCN method utilizes two branches, CNN and GCN, to generate complementary contextual features at the pixel and super-pixel levels, significantly improving classification accuracy at boundaries of diverse land covers. However, the CEGCN method cannot handle the inherent speckle noise in PolSAR images well, so there are some noise points in the image. The SGCN-CNN method, similar to the CEGCN method, also struggles to effectively handle the inherent speckle noise in images, resulting in the generation of a significant number of noise points. The proposed MLDnet combines the L-DeeplabV3+ network to deeply explore multi-dimensional features of PolSAR image, and select effective features through the incorporation of a channel attention module. Compared to six contrastive methods, the proposed MLDnet method in (h) demonstrates substantial improvements in both regional consistency and edge details.
Table 3 lists the classification accuracy of all methods in the Xi’an data set. It can be seen that the Super-RF method without deep features has lower accuracy than other methods. The main reason is that the classification accuracy of water class is only 70.91%, which is relatively consistent with the classification result in Figure 6b. In comparison, CNN has greatly improved its accuracy in water class, but its accuracy in the building class is relatively lower, resulting in the low overall accuracy. It only improved by 0.27% compared to Super-RF. CV-CNN further improves the classification accuracy of the water class, with the highest accuracy among all methods. However, its classification accuracy in the grass class is the lowest among all methods, which also results in its OA value being slightly lower than CNN. Compared to the previous three methods, 3D-CNN has achieved a significant improvement in OA value, and PolMPCNN has further improve the performance, especially in the building class, with the accuracy of 97.68%, which is significant higher than other compared methods. Analogously, due to better distinguishing grass class, CEGCN has further improved the overall classification accuracy. The SGCN-CNN method experiences a decline in overall classification metrics due to a significant number of misclassifications within grass class. The proposed MLDnet achieves the best classification accuracy in both two land cover types and three overall evaluation indicators. Specifically, the OA values are improved by 7.44%, 7.17%, 7.79%, 4.17%, 3.37%, 1.17% and 6.20% compared to other methods.
Figure 7 shows the confusion matrix of the Xi’an data set. From the figure, it can be seen that the proposed method has the less misclassified pixels in both grass and building classes than other compared methods. The main confusion of the proposed method lies on misclassifying 2030 (about 5.70%) pixels of water class as grass, while misclassifying 1654 (about 1.40%) pixels of grass as water class.

3.3. Experimental Results on Oberpfaffenhofen Data Set

The classification maps produced by compared and proposed methods are illustrated in Figure 8b–i, respectively. As shown in Figure 8b, the Super-RF method effectively distinguishes bare ground and farmland classes due to large number of samples. However, for classes with fewer samples such as building, forest, and road classes, misclassification is more noticeable. The CNN method shown in (c) has enhanced classification performance for building and road classes. However, its effectiveness in differentiating the farmland class remains suboptimal. In comparison, the CV-CNN method in (d) shows a marginal improvement in mitigating misclassification issues related to the farmland class. Nevertheless, it exacerbates misclassification in the bare ground class in the upper left corner, and overall regional consistency remains relatively poor. The 3D-CNN method in (e) exhibits similar to the Super-RF method, displaying a notable number of misclassified pixels in the road and farmland classes. Despite this, there is an enhancement in building class. Compared to the preceding four methods, the PolMPCNN method in (f) demonstrates a substantial improvement in regional consistency. However, this method almost entirely misclassifies the road class in the lower half of the images as the building class. Additionally, the boundaries between different land covers appear blurry. CEGCN has further enhanced the classification performance of the road class, building upon the foundation laid by PolMPCNN. However, the effectiveness is still suboptimal. The SGCN-CNN method in (h) likewise demonstrates subpar performance when dealing with classes with fewer samples, such as road and farmland. The proposed MLDnet method, as exhibited in (i), demonstrates a remarkable improvement in the classification performance of all five land covers compared to other contrastive methods. Furthermore, it effectively preserves the boundaries among different land covers.
Table 4 compares the classification accuracy of different methods in the Oberpfaffenhofen area. From the table, it can be visually observed that the seven compared methods have relatively low OA values on the Oberpfaffenhofen data set, with the highest value being only 88.93% (CEGCN). Moreover, among the five land cover types, bare ground class exhibits relatively high accuracy in discrimination. The four comparative methods (Super-RF, 3D-CNN, CEGCN and SGCN-CNN) all achieve accuracy rates exceeding 90%. Meanwhile, all seven comparative methods have an accuracy rate surpassing 81% for forest class. However, for the remaining three classes, particularly farmland and road classes, none of the seven comparative methods achieve satisfactory performance. The highest discrimination rates are 80.70% (CEGCN) for farmland class and 76.20% (CVCNN) for road class. Notably, for the road class in PolMPCNN, the correct discrimination rate is only 10.38%, consistent with the classification result displayed in Figure 8f. This indicates that for data sets with unbalanced samples, such as Oberpfaffenhofen, the classification performance of the six comparison methods is constrained. However, The proposed MLDnet method excels in achieving the highest classification accuracy across all evaluation metrics, with values consistently exceeding 93%. Furthermore, in comparison to the seven comparative algorithms, the OA value sees an increase of 17.79%, 15.32%, 21.31%, 13.37%, 20.37%, 7.26% and 18.90%, respectively. This highlights the superior capability of the proposed method in effectively handling complex scenes of PolSAR images and enhancing classification accuracy.
Additionally, the confusion matrices for the Oberpfaffenhofen data set are presented in Figure 9. Our method demonstrates the highest number of correctly classified pixels in the dialog elements, indicating superior classification performance across all classes. The CV-CNN shows the most misclassifications in the bare ground category. The Super-RF frequently misclassifies pixels from buildings as bare ground. Furthermore, the SGCN-CNN misclassifies a significant number of pixels from farmland and road into the bare ground class, which aligns with the classification map shown in Figure 8h. It is evident that the primary confusion across all methods occurs between the bare ground and road classes, where our method outperforms the others.

3.4. Experimental Results on San Francisco Data Set

The experimental results of seven contrastive methods and MLDnet algorithms on the San Francisco data set can be seen in Figure 10b–i, respectively. Figure 10a displays the ground truth map of the San Francisco area. Upon comparing the ground truth map, it is evident that super-RF produces many misclassifications in high density urban (shorted by “HD urban”) class with significant surface scattering differences. It also exhibits a portion of confusion between vegetation and developed classes. The CNN method also faces challenges in distinguishing high density urban class from developed class. In comparison to the previous two methods, the classification performance of CV-CNN has been improved in high density urban and developed classes, while there are still a small number of misclassified pixels. Simultaneously, the misclassification phenomenon in low density urban (shorted by “LD urban”) class is also quite evident. The 3D-CNN method demonstrates improved classification performance on high-density urban, low-density urban, and developed classes but exhibits more misclassified pixels in the vegetation class. PolMPCNN improves classification performance in various classes. However, its classification map still contains some visually apparent misclassified pixels. CEGCN and SGCN-CNN almost perfectly distinguish various land covers; however, upon closer observation, it is not difficult to notice that they have some misclassified pixels along land cover boundaries. The proposed MLDnet method also effectively distinguishes various land covers, meanwhile, it preserves the boundaries of different land cover types well. This suggests that the MLDnet method is effective in achieving good classification results even for large-scale PolSAR data sets.
In addition, Table 5 provides the classification accuracy of all methods for various land covers in the San Francisco data set, along with the corresponding OA, AA, and Kappa. Based on the table, it can be observed that, in comparison to the other three land cover types, both the Super-RF and CNN methods exhibit relatively low accuracy in classifying high density urban and developed classes. This is especially evident for the Super-RF method, with precision rates of 77.76% and 81.00%, respectively. CV-CNN exhibits a relatively low accuracy rate in the low density urban class. In comparison, the classification accuracy by 3D-CNN and PolMPCNN methods is higher, with OA values both above 98%. The CEGCN and SGCN-CNN methods both exhibit excellent performance across five different land cover types, with accuracy exceeding 99% for all categories except vegetation class. In contrast, the proposed MLDnet method achieves the highest class accuracy for all five land cover types, demonstrating superior classification performance. In terms of overall metrics, using OA metric as an example, compared to the seven other comparative algorithms, MLDnet improves OA by 5.64%, 3.69%, 2.26%, 1.59%, 1.04%, 0.42%, and 0.50% respectively.
The confusion matrices for the San Francisco data set are illustrated in Figure 11. Our method shows the least misclassification across all categories. The CV-CNN exhibits the highest misclassifications, particularly misclassifying ocean as vegetation. The Super-RF frequently misclassifies pixels from high density urban as low density urban. Additionally, the Super-RF, CNN, and CV-CNN methods show significant misclassifications within the develop class. It is clear that all methods struggle with misclassifications in heterogeneous regions, while our method demonstrates the best performance.

3.5. Experimental Results on Flevoland1 Data Set

The experimental results of various methods on the Flevoland1 data set in the Netherlands are depicted in Figure 12. Specifically, the Super_RF in (b) is a low-level feature learning method that exhibits limitation in effectively classifying urban class. In contrast, the CNN method in (c) demonstrates some enhancement in the classification of land cover boundaries; however, the overall classification performance remains inadequate, marked by numerous noisy points in the image. The CV-CNN method in (d) and the 3D-CNN method in (e) have enhanced regional consistency to some extent, but the improvement effect is not significant. There are still relatively noticeable misclassification pixels in the classification maps. PolMPCNN in (f) further enhances the regional consistency, effectively distinguishing between water and woodland classes, with only a small amount of misclassified pixels in cropland and urban classes. CEGCN in (g) and SGCN-CNN in (h) further enhance the ability to distinguish various land cover types, with only a few misclassifications occurring at the boundary between cropland and urban classes. The proposed MLDnet in (i), leveraging a 57-dimensional multi-feature and the L-DeeplabV3+ deep learning network, exhibits a notable improvement compared to previous methods. There are almost no obvious misclassification phenomena in the image. The consistency of regions and the preservation of boundaries have been significantly enhanced.
In addition, to quantitatively assess the classification performance, the classification accuracies of all methods on the Flevoland1 data set are presented in Table 6. It is evident that Super-RF attains the lowest correct classification rate of 81.84% in the urban class. CNN method achieves the lowest classification accuracy in the cropland class, at 93.16%. CV-CNN demonstrates great improvements, especially for the water class, with a correct discrimination rate second only to the proposed method MLDnet, reaching 99.85%. 3D-CNN method improves the classification accuracy in the cropland class compared to the previous three methods, with increases of 2.59%, 3.59%, and 2.81% respectively. PolMPCNN method further improves the classification metrics for various classes in the Flevoland1 data set, achieving an OA value of 98.49%. The CEGCN and SGCN-CNN further improve the classification accuracy of various types of land cover, with numerical values consistently above 98%. Additionally, it is noticeable that multiple comparison algorithms exhibit relatively lower correct classification rates for the urban class. This suggests that the classification of the urban class is more difficult. However, it is obvious that the proposed MLDnet excels in classifying all types of land covers and in overall classification performance. Especially for the urban class, the accuracy has also reached 99.98%. This demonstrates the effectiveness of the proposed method.
The confusion matrices for the Flevoland1 region are presented in Figure 13. It is clear that the proposed method exhibits the least misclassification across all four classes. Notably, the highest misclassification for our method occurs with 236 pixels misclassified from woodland to cropland, yet this is still the lowest count compared to other methods. Specifically, for the urban and water classes, our method shows only around 30 misclassified pixels, while other methods report misclassification numbers in the hundreds or even thousands.

3.6. Experimental Results on Flevoland2 Data Set

The experimental results using the proposed MLDnet method and seven comparison methods on the Flevoland2 data set are shown in Figure 14. By comparing these results with the label map in (a), it is evident that the superpixel-level classification method, Super-RF, and the pixel-level classification method, CNN, exhibit numerous patchy misclassified areas and scattered misclassification noise across multiple classes. In contrast, the CV-CNN method achieves better classification results, though it still shows noticeable misclassified pixels in the rapeseed class. Furthermore, due to the lack of global information, the 3D-CNN method generates spotty noise in multiple classes such as stem beans, wheat, and rapeseed. PolMPCNN effectively learns various land cover features through a multi-path architecture but still has limitations in distinguishing similar land covers such as barley, wheat2, wheat3, and rapeseed. CEGCN combines pixel-level and superpixel-level features effectively, resulting in good differentiation of various land covers, with only a few misclassifications in the rapeseed and wheat classes. The SGCN-CNN method exhibits a small number of misclassifications in the wheat2, rapeseed, and grasses classes. Additionally, the analysis reveals that all seven comparison methods exhibit varying degrees of misclassified pixels in the rapeseed class. However, the proposed MLDnet method achieves excellent classification performance across all land covers, including the rapeseed class. This success is contributed to the proposed MLDnet method, which comprehensively considers different land cover characteristics and extracts various features such as texture and contour features, thereby enhancing the network’s ability to distinguish similar crops.
Meanwhile, Table 7 provides the per-class accuracy, overall accuracy (OA), average accuracy (AA), and Kappa coefficient for the above methods and the proposed model. The data show that the Super-RF method achieves excellent classification performance in the barley and water classes, with an accuracy of 100%, but completely misclassified the buildings class, thus lowering the average accuracy. The CNN method shows a slight improvement in OA, but not reach 90% accuracy in several classes, such as grasses and barley. In contrast, CV-CNN significantly improves both class differentiation and overall evaluation metrics, achieving an OA of 98.96%. Similar to Super-RF, 3D-CNN performs well in the water class but fails in buildings class. PolMPCNN, on the other hand, differentiates the buildings class well but has the lowest class accuracy in the barley class, with only 40.25%. CEGCN achieves over 99% class accuracy in multiple classes, with main misclassifications occurring in the peas, grasses, and buildings classes. The SGCN-CNN method shows a slight improvement over CEGCN, with all metrics exceeding 90%. However, the proposed MLDnet achieves the highest class accuracy in all 13 land cover types, with an overall accuracy of 99.95%. These results fully validate the effectiveness and reliability of the proposed method.
The confusion matrix for the Flevoland2 region is shown in Figure 15. From this figure, it is evident that the proposed method misclassifies fewer pixels than the other methods across all classes, with the highest number of misclassified pixels being only 86. Notably, even in the buildings class, which contains fewer pixels, the proposed method still outperforms the others. This highlights the method’s exceptional ability to accurately classify complex PolSAR datasets with diverse land cover types.

4. Discussion

4.1. Effect of Each Submodule

In this paper, we formulate 57-dimensional multi-features as the input, with L-DeeplabV3+ as the main network, combined with the channel attention module to form the proposed MLDnet. Therefore, in order to verify the effectiveness of the three modules mentioned above, we conducts the ablation experiments. Firstly, to assess the effectiveness of the proposed 57-dimension multi-features, we replace the 57-dimension multi-features by the original 9-dimensional PolSAR data as the input without changing other modules (denoted by “wihtout MF”). Secondly, to validate the effectiveness of proposed L-DeepLabV3+ network, we replace it by a traditional CNN model, while keeping other components the same as the proposed method, noted by “without L-deeplabV3+”. Finally, to validate the channel attention module, we remove the channel attention module from the proposed MLDnet, denoted by “without Attention”.
Table 8 presents the classification performance of three ablation experiments and the proposed MLDnet on five experimental data sets, evaluated using OA value and Kappa coefficient. From the Table, it is evident that for the three sets of ablation experiments, each group is able to achieve satisfactory classification performance. The proposed MLDnet integrates the advantages of all three methods, thus achieving higher statistical classification accuracy. For example, compared with “without MF” method, MLDnet improves the OA value by 0.52%, 2.94%, 0.75%, 1.03% and 0.17% on five data sets respectively, demonstrating the effectiveness of the proposed multi-feature module. The same conclusion can be found in other modules, which prove the effectiveness of the proposed method.

4.2. Validity Analysis of L-DeeplabV3+ Network

In order to verify the performance of L-deeplabV3+, the multi-dimensional features of the extracted Flevoland2 data set are put into Deeplabv3+ and L-deeplabV3+ networks respectively for comparison, as shown in Table 9. From the table data, it can be seen that the proposed L-deeplabV3+ network achieves better classification performance by fully considering the characteristics of PolSAR data. Among the fifteen different land cover types, fourteen achieve the highest accuracy. Additionally, in terms of overall evaluation metrics, the L-deeplabV3+ network improved by 1.96% and 6.63% respectively compared to the original DeeplabV3+ network.

4.3. Validity Analysis of Multi-Features

To verify the effectiveness of the proposed 57-dimensional multi-feature input in the MLDnet method, we conduct ablation experiments on the Xi’an data set. Specifically, experiments are performed using the original data features (16 dimensions), the target decomposition features (17 dimensions), and the texture and contour features (24 dimensions) within the proposed network framework. For clarity, we name these three features as feature1, feature2, and feature3, respectively. The experimental results are shown in Figure 16. It can be seen that classification with only original data (feature1) cannot classify the heterogenous regions well in Figure 16b (such as the region in yellow circle), sine there are great scattering variations within heterogenous objects. In addition, compared with feature1, the method with feature2 can improve classification performance with multiple target decomposition-based features, while edges of urban areas, such as the region in the black circle, still cannot be well identified. The feature3 is contour and textual features, which can identify the urban and grass areas well, but confuse some pixels in water area. After utilizing multi-features, the proposed method can obtain superior performance in both heterogenous and water areas.
In addition, to quantitatively evaluate the effectiveness of multi-features, the classification accuracy of different features are given in Table 10. Clearly, as mentioned earlier, using only the original data features is insufficient for effective classification of heterogeneous targets. For instance, in distinguishing building class, feature1 has the lowest accuracy at only 85.95%, whereas feature2 and feature3 both achieve accuracy above 97%. Additionally, for the relatively homogeneous grass class, feature3, which uses texture and contour features, demonstrates a clear advantage. For water class with intricate boundaries, feature2, which uses target decomposition features, achieves higher class accuracy. The 57-dimensional multi-feature input used in this paper comprehensively considers various terrain characteristics and provides multiple complementary features, thereby achieving the best classification performance. In terms of overall accuracy (OA), compared to the three individual feature sets, the proposed method shows improvements of 5.99%, 4.13%, and 1.4%, respectively.

4.4. Validity Analysis of Channel Attention Module

Feature selection is one of the key factors influencing classification performance. To verify the effectiveness of the channel attention module in our network framework, we conduct experimental comparative analyses of the position and number of attention modules on the Xi’an data set, as shown in Table 11. Specifically, considering that different types of features may be cross-fused under convolution operations, we introduce varying numbers of channel attention modules before and after the L-deeplabV3+ network in three sets of experiments. For the experiments with attention modules placed before the L-deeplabV3+ network, we name them front1, front2, and front3, with the numbers indicating the number of added channel attention modules. Similarly, for those placed after the L-deeplabV3+ network, we name them behind1, behind2, and behind3. From the data in the table, it is evident that adding multiple channel attention modules, whether before or after the convolution operations, can lead to a decrease of varying degrees in classification performance. Additionally, it increases the model’s parameter, causing unnecessary computational burden. What’s more, comparing front1 and behind1, it is clear that behind1 significantly outperforms the former in both class accuracy and overall classification metrics, with class accuracy improvements of 7.15%, 5.45%, and 3.38%, respectively. Therefore, in the proposed MLDnet method, the optimal strategy is to introduce one channel attention module after the L-deeplabV3+ network (behind1), thereby significantly enhancing classification performance and maintaining a low computational burden.
In summary, the experimental results show that introducing an appropriate number of channel attention modules in the network, especially when placed after the network, can effectively improve the classification performance of PolSAR images. At the same time, avoiding the excessive addition of attention modules prevents a significant increase in model complexity and computational cost, thus achieving a balance between performance and efficiency.

4.5. Effect of Sampling Window Size

Figure 17 further summarizes the change curve of the overall classification accuracy on sampling window size. The size of the sampling square window changes from 8 × 8 to 24 × 24 . From the figure, it can be observed that overall tendency of the change curve is gradually rising with increased sampling window size for most of data sets. However, there is a slightly decrease for Xi’an data set when the window size is larger than 16. The reason is analysed as that too small window is difficult to effectively capture specific contextual information, while too large window may involve a large amount of irrelevant neighborhood information, leading to misclassification. Moreover, it is evident that when the window size reaches 16 × 16 , the OA values of the five data sets do not significantly increase. Especially for the Flevoland1 and San Francisco data sets, as the window size varies from 16 × 16 to 24 × 24 , the OA value only increases by 0.04% and 0.01%. However, during this process, the network’s complexity is significantly increased. Therefore, considering the classification performance and time complexity comprehensively, we select the sampling window size as 16 × 16 .

4.6. Effect of Training Sample Ratio

The proportion of training samples is a crucial factor in influencing the classification performance. To analyze its impact on classification accuracy by the proposed MLDnet, we set the training sample proportions from 1%, 5%, 10%, 15% to 20%, and experiments are conducted on the five data sets above. The effect of the training sample ratio on classification accuracy is given in Figure 18. It is obvious that for the Oberpfaffenhofen, Flevoland1, Flevoland2 and San Francisco data sets, when the training sample ratio reaches 10%, the model becomes stable, and the OA value no longer experiences a significant increase. However, for the Xi’an data set, it reaches stable when the training sample ratio is in the range of 10–15%. Nevertheless, when the training sample ratio is in the range of 15–20%, the OA value shows a downward trend. This may be attributed to the continuous increase in the training sample ratio exacerbating sample imbalance in the Xi’an data set, leading to the model overfitting. Therefore, considering the four experimental data sets overall, we chooses 10% as the training sample ratio for experiments.

4.7. Analysis of Running Time

Table 12 summarizes the running time of each method on the Xi’an data set. From the table, it can be visually seen that, compared to first five deep learning methods, the traditional method Super-RF consumes the least time in both training and testing. Although it requires slightly longer time than CNN, CEGCN, SGCN-CNN and 3D-CNN methods, MLDnet can significantly reduce training and testing time compared with CV-CNN and PolMPCNN methods. In addition, compared with DeeplabV3+ method, the proposed MLDnet cost less time in both training and testing stages, which proves the effectiveness of the proposed lightweight deeplabV3+ network. Besides, the proposed MLDnet achieves the highest classification accuracy. This indicates that the proposed method can significantly improve the classification accuracy of PolSAR images without increasing too much time, thereby achieving dual improvements in time and performance.

5. Conclusions

In this paper, we propose a novel multi-feature deep attention network for PolSAR image classification. Unlike conventional networks using the original data as input, our proposed MLDnet extracts a variety of complementary features, including the original data, scattering features, and image features. Subsequently, a lightweight DeeplabV3+ network is designed to learn multi-scale high-level features, significantly reducing parameters and improving computation efficiency. Finally, a channel attention fusion module is developed for feature selection, enhancing valuable features and suppressing useless ones to improve classification performance. Experiments were conducted on five real PolSAR data sets with different bands and sensors. The results demonstrate that the proposed method achieves superior classification performance in both region homogeneity and edge preservation compared to state-of-the-art methods.
Furthermore, while the attention mechanism serves as an effective feature fusion method for high-dimensional multiple features, it remains a relatively straightforward feature selection mechanism. In our future work, we aim to incorporate mutual information to thoroughly exploit relationships among various features, thereby enhancing feature selection. Moreover, some few-shot learning methods for unbalanced samples should be exploited in the further work.

Author Contributions

Conceptualization, J.S. and H.J.; Methodology, S.J.; Software, Y.Z.; Formal analyses, M.G.; writing—review and editing, W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 62471387, 62272383, the Youth Innovation Team Research Program Project of Shaanxi Provincial Education Department under Grant 23JP111.

Data Availability Statement

The original data presented in the study are openly available in https://github.com/fly-shi (accessed on 6 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dong, Y.; Li, F.; Hong, W.; Zhou, X.; Ren, H. Land cover semantic segmentation of Port Area with High Resolution SAR Images Based on SegNet. In Proceedings of the 2021 SAR in Big Data Era (BIGSARDATA), Nanjing, China, 22–24 September 2021; pp. 1–4. [Google Scholar]
  2. Liu, H.; Zhu, T.; Shang, F.; Liu, Y.; Lv, D.; Yang, S. Deep Fuzzy Graph Convolutional Networks for PolSAR Imagery Pixelwise Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 504–514. [Google Scholar] [CrossRef]
  3. Li, F.; Yi, M.; Zhang, C.; Yao, W.; Hu, X.; Liu, F. POLSAR Target Recognition Using a Feature Fusion Framework Based on Monogenic Signal and Complex-Valued Nonlocal Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7859–7872. [Google Scholar] [CrossRef]
  4. Chen, S.; Cui, X.; Wang, X.; Xiao, S. Speckle-Free SAR Image Ship Detection. IEEE Trans. Image Process. 2021, 30, 5969–5983. [Google Scholar] [CrossRef]
  5. Sanchez, S.; Marpu, P.R.; Plaza, A.J.; Paz, A. Parallel Implementation of Polarimetric Synthetic Aperture Radar Data Processing for Unsupervised Classification Using the Complex Wishart Classifier. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 5376–5387. [Google Scholar] [CrossRef]
  6. Bombrun, L.; Vasile, G.; Gay, M.; Totir, F. Hierarchical Segmentation of Polarimetric SAR Images Using Heterogeneous Clutter Models. IEEE Trans. Geosci. Remote Sens. 2011, 49, 726–737. [Google Scholar] [CrossRef]
  7. Liu, M.; Deng, Y.; Han, C.; Hou, W.; Gao, Y.; Wang, C.; Liu, X. An Innovative Supervised Classification Algorithm for PolSAR Image Based on Mixture Model and MRF. Remote. Sens. 2022, 14, 5506. [Google Scholar] [CrossRef]
  8. Pallotta, L.; Tesauro, M. Screening Polarimetric SAR Data via Geometric Barycenters for Covariance Symmetry Classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 4002905. [Google Scholar] [CrossRef]
  9. Chen, S.W.; Wang, X.S.; Li, Y.Z.; Sato, M. Adaptive Model-Based Polarimetric Decomposition Using PolInSAR Coherence. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1705–1718. [Google Scholar] [CrossRef]
  10. Pallotta, L.; Orlando, D. Polarimetric Covariance Eigenvalues Classification in SAR Images. IEEE Geosci. Remote Sens. Lett. 2019, 16, 746–750. [Google Scholar] [CrossRef]
  11. Hanis, D.; Hadj-Rabah, K.; Belhadj-Aissa, A.; Pallotta, L. Dominant Scattering Mechanism Identification from Quad-Pol-SAR Data Analysis. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 14408–14420. [Google Scholar] [CrossRef]
  12. Chen, S.W.; Li, M.D.; Cui, X.; Li, H.L. Polarimetric Roll-Invariant Features and Applications for Polarimetric Synthetic Aperture Radar Ship Detection: A comprehensive summary and investigation. IEEE Geosci. Remote Sens. Mag. 2024, 12, 36–66. [Google Scholar] [CrossRef]
  13. Luo, S.; Sarabandi, K.; Tong, L.; Pierce, L.E. A SAR Image Classification Algorithm Based on Multi-Feature Polarimetric Parameters Using FOA and LS-SVM. IEEE Access 2019, 7, 175259–175276. [Google Scholar] [CrossRef]
  14. Chen, Q.; Cao, W.; Shang, J.; Liu, J.; Liu, X. Superpixel-Based Cropland Classification of SAR Image with Statistical Texture and Polarization Features. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4503005. [Google Scholar] [CrossRef]
  15. Salakhutdinov, R.; Hinton, G.E. Deep Boltzmann Machines. In Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics, Clearwater Beach, FL, USA, 16–18 April 2009. [Google Scholar]
  16. Luo, J.; Lv, Y.; Guo, J. Multi-temporal PolSAR Image Classification Using F-SAE-CNN. In Proceedings of the 2022 3rd China International SAR Symposium (CISS), Shanghai, China, 2–4 November 2022; pp. 1–5. [Google Scholar]
  17. Hinton, G.E.; Osindero, S.; Teh, Y.W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  18. Chen, S.; Tao, C. PolSAR Image Classification Using Polarimetric-Feature-Driven Deep Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2018, 15, 627–631. [Google Scholar] [CrossRef]
  19. Shi, J.; He, T.; Ji, S.; Nie, M.; Jin, H. CNN-Improved Superpixel-to-Pixel Fuzzy Graph Convolution Network for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4410118. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Wang, H.; Xu, F.; Jin, Y. Complex-Valued Convolutional Neural Network and Its Application in Polarimetric SAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
  21. Guo, Y.; Jiao, L.; Qu, R.; Sun, Z.; Wang, S.; Wang, S.; Liu, F. Adaptive Fuzzy Learning Superpixels Representation for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5217818. [Google Scholar] [CrossRef]
  22. Zhang, L.; Zhang, S.; Zou, B.; Dong, H. Unsupervised Deep Representation Learning and Few-Shot Classification of PolSAR Images. IEEE Trans. Geosci. Remote Sens. 2020, 60, 5100316. [Google Scholar] [CrossRef]
  23. Ai, J.; Wang, F.; Mao, Y.; Luo, Q.; Yao, B.; Yan, H.; Xing, M.d.; Wu, Y. A Fine PolSAR Terrain Classification Algorithm Using the Texture Feature Fusion Based Improved Convolutional Autoencoder. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5218714. [Google Scholar] [CrossRef]
  24. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  26. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.P.; Yuille, A.L. Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  27. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.P.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 40, 834–848. [Google Scholar] [CrossRef]
  28. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  29. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  30. Zhang, Q.; Yang, Z.; Zhao, W.; Yu, X.; Yin, Z. Polarimetric SAR Landcover Classification Based on CNN with Dimension Reduction of Feature. In Proceedings of the 2021 IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; pp. 331–335. [Google Scholar]
  31. Zhang, F.; Li, P.; Zhang, Y.; Liu, X.; Ma, X.; Yin, Z. A Enhanced DeepLabv3+ for PolSAR image classification. In Proceedings of the 2023 4th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 7–9 April 2023; pp. 743–746. [Google Scholar]
  32. Chen, S.W.; Wang, X.S.; Sato, M. Uniform Polarimetric Matrix Rotation Theory and Its Applications. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4756–4770. [Google Scholar] [CrossRef]
  33. Li, M.D.; Xiao, S.P.; Chen, S.W. Three-Dimension Polarimetric Correlation Pattern Interpretation Tool and its Application. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5238716. [Google Scholar] [CrossRef]
  34. Wang, X.; Zhang, L.; Wang, N.; Zou, B. Joint Polarimetric-Adjacent Features Based on LCSR for PolSAR Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6230–6243. [Google Scholar] [CrossRef]
  35. Shi, J.; Jin, H.; Li, X. A Novel Multi-Feature Joint Learning Method for Fast Polarimetric SAR Terrain Classification. IEEE Access 2020, 8, 30491–30503. [Google Scholar] [CrossRef]
  36. Shang, R.; Wang, J.; Jiao, L.; Yang, X.H.; Li, Y. Spatial feature-based convolutional neural network for PolSAR image classification. Appl. Soft Comput. 2022, 123, 108922. [Google Scholar] [CrossRef]
  37. Xu, X.; Lu, Y.; Zou, B. Building Extraction From PolSAR Image Based on Deep CNN with Polarimetric Features. In Proceedings of the 2020 21st International Radar Symposium (IRS), Warsaw, Poland, 5–8 October 2020; pp. 117–120. [Google Scholar]
  38. Wu, Q.; Wen, Z.; Wang, Y.; Luo, Y.; Li, H.; Chen, Q.Y. A Statistical-Spatial Feature Learning Network for PolSAR Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4016305. [Google Scholar] [CrossRef]
  39. Singh, S.R.; Murthy, H.A.; Gonsalves, T.A. Feature Selection for Text Classification Based on Gini Coefficient of Inequality. In Proceedings of the Fourth International Workshop on Feature Selection in Data Mining, Hyderabad, India, 21 June 2010. [Google Scholar]
  40. Lv, Z.; Zhang, P.; SUN, W.; Benediktsson, J.A.; Li, J.; Wang, W. Novel Adaptive Region Spectral–Spatial Features for Land Cover Classification With High Spatial Resolution Remotely Sensed Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5609412. [Google Scholar] [CrossRef]
  41. Yao, C.; Zheng, L.; Feng, L.; Yang, F.; Guo, Z.; Ma, M. A Collaborative Superpixelwise Autoencoder for Unsupervised Dimension Reduction in Hyperspectral Images. Remote Sens. 2023, 15, 4211. [Google Scholar] [CrossRef]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  43. Wang, W.; Xiao, C.; Dou, H.; Liang, R.; Yuan, H.; Zhao, G.; Chen, Z.; Huang, Y. CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection. Remote Sens. 2023, 15, 5539. [Google Scholar] [CrossRef]
  44. Sparse Mix-Attention Transformer for Multispectral Image and Hyperspectral Image Fusion. Remote Sens. 2023, 16, 144. [CrossRef]
  45. You, H.; Gu, J.; Jing, W. Multi-Label Remote Sensing Image Land Cover Classification Based on a Multi-Dimensional Attention Mechanism. Remote. Sens. 2023, 15, 4979. [Google Scholar] [CrossRef]
  46. Chen, S.; Li, Y.; Wang, X.; Xiao, S.; Sato, M. Modeling and Interpretation of Scattering Mechanisms in Polarimetric Synthetic Aperture Radar: Advances and perspectives. IEEE Signal Process. Mag. 2014, 31, 79–89. [Google Scholar] [CrossRef]
  47. Chen, S. Polarimetric Coherence Pattern: A Visualization and Characterization Tool for PolSAR Data Investigation. IEEE Trans. Geosci. Remote Sens. 2018, 56, 286–297. [Google Scholar] [CrossRef]
  48. Lee, J.S.; Grunes, M.; de Grandi, G. Polarimetric SAR speckle filtering and its implication for classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2363–2373. [Google Scholar]
  49. Chen, S.W. SAR Image Speckle Filtering with Context Covariance Matrix Formulation and Similarity Test. IEEE Trans. Image Process. 2020, 29, 6641–6654. [Google Scholar] [CrossRef]
  50. Zhang, Y.; Wang, W.; Guo, Z.; Li, N. Enhanced PGA for Dual-Polarized ISAR Imaging by Exploiting Cloude-Pottier Decomposition. IEEE Geosci. Remote Sens. Lett. 2024, 21, 4000505. [Google Scholar] [CrossRef]
  51. Zhao, L.; Zhou, X.; Jiang, Y.; Kuang, G. Iterative classification of polarimetric SAR image based on the freeman decomposition and scattering entropy. In Proceedings of the 2007 1st Asian and Pacific Conference on Synthetic Aperture Radar, Huangshan, China, 5–9 November 2007; pp. 473–476. [Google Scholar]
  52. Li, D.; Zhang, Y. Unified Huynen Phenomenological Decomposition of Radar Targets and Its Classification Applications. IEEE Trans. Geosci. Remote Sens. 2016, 54, 723–743. [Google Scholar] [CrossRef]
  53. Wang, S.; Pei, J.; Liu, K.; Zhang, S.; Chen, B. Unsupervised classification of POLSAR data based on the polarimetric decomposition and the co-polarization ratio. In Proceedings of the 2011 IEEE International Geoscience and Remote Sensing Symposium, Vancouver, BC, Canada, 24–29 July 2011; pp. 424–427. [Google Scholar]
  54. Hong, S.H.; Wdowinski, S. Double-Bounce Component in Cross-Polarimetric SAR From a New Scattering Target Decomposition. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3039–3051. [Google Scholar] [CrossRef]
  55. Benco, M.; Kamencay, P.; Radilova, M.; Hudec, R.; Šinko, M. The Comparison of Color Texture Features Extraction based on 1D GLCM with Deep Learning Methods. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Niteroi, Brazil, 1–3 July 2020; pp. 285–289. [Google Scholar]
  56. Shi, J.; Jin, H.; Xiao, Z. A Novel Hybrid Edge Detection Method for Polarimetric SAR Images. IEEE Access 2020, 8, 8974–8991. [Google Scholar] [CrossRef]
  57. Li, Y.; Zhang, H.; Shen, Q. Spectral-Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote. Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  58. Cui, Y.; Liu, F.; Jiao, L.; Guo, Y.; Liang, X.; Li, L.; Yang, S.; Qian, X. Polarimetric Multipath Convolutional Neural Network for PolSAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5207118. [Google Scholar] [CrossRef]
  59. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. CNN-Enhanced Graph Convolutional Network With Pixel- and Superpixel-Level Feature Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 8657–8671. [Google Scholar] [CrossRef]
  60. Jin, H.; He, T.; Shi, J.; Ji, S. Combine Superpixel-Wise GCN and Pixel-Wise CNN for PolSAR Image Classification. In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 8014–8017. [Google Scholar]
Figure 1. Framework of the proposed multi-feature lightweight deeplabV3+ network.
Figure 1. Framework of the proposed multi-feature lightweight deeplabV3+ network.
Remotesensing 17 01422 g001
Figure 2. Improved Xception network.
Figure 2. Improved Xception network.
Remotesensing 17 01422 g002
Figure 3. Structure comparison between DeeplabV3+ and L-DeeplabV3+ networks, the yellow module represents the backbone network. (a) DeeplabV3+, (b) L-DeeplabV3+.
Figure 3. Structure comparison between DeeplabV3+ and L-DeeplabV3+ networks, the yellow module represents the backbone network. (a) DeeplabV3+, (b) L-DeeplabV3+.
Remotesensing 17 01422 g003
Figure 4. The attention-based multi-feature fusion module.
Figure 4. The attention-based multi-feature fusion module.
Remotesensing 17 01422 g004
Figure 5. PauliRGB images and the group-truth maps on the first three PolSAR data sets. (a) Xi’an data set; (b) Oberpfaffenhofen data set; (c) San Francisco data set; (d) Flevoland1 data set; (e) Flevoland1 data set.
Figure 5. PauliRGB images and the group-truth maps on the first three PolSAR data sets. (a) Xi’an data set; (b) Oberpfaffenhofen data set; (c) San Francisco data set; (d) Flevoland1 data set; (e) Flevoland1 data set.
Remotesensing 17 01422 g005
Figure 6. Classification results of different methods on the Xi’an data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Figure 6. Classification results of different methods on the Xi’an data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Remotesensing 17 01422 g006
Figure 7. Confusion matrices on Xi’an data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Figure 7. Confusion matrices on Xi’an data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Remotesensing 17 01422 g007
Figure 8. Classification results of different methods on the Oberpfaffenhofen data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Figure 8. Classification results of different methods on the Oberpfaffenhofen data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Remotesensing 17 01422 g008
Figure 9. Confusion matrices on Oberpfaffenhofen data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Figure 9. Confusion matrices on Oberpfaffenhofen data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Remotesensing 17 01422 g009
Figure 10. Classification results of different methods on the San Francisco data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Figure 10. Classification results of different methods on the San Francisco data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Remotesensing 17 01422 g010
Figure 11. Confusion matrices on San Francisco data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Figure 11. Confusion matrices on San Francisco data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Remotesensing 17 01422 g011
Figure 12. Classification results of different methods on the Flevoland1 data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Figure 12. Classification results of different methods on the Flevoland1 data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Remotesensing 17 01422 g012
Figure 13. Confusion matrices on Flevoland1 data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Figure 13. Confusion matrices on Flevoland1 data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Remotesensing 17 01422 g013
Figure 14. Classification results of different methods on the Flevoland2 data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Figure 14. Classification results of different methods on the Flevoland2 data set. (a) The label map; (b) Super-RF; (c) CNN; (d) CV-CNN; (e) 3D-CNN; (f) PolMPCNN; (g) CEGCN; (h) SGCN-CNN; (i) MLDnet.
Remotesensing 17 01422 g014
Figure 15. Confusion matrices on Flevoland2 data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Figure 15. Confusion matrices on Flevoland2 data set. (a) Super-RF; (b) CNN; (c) CV-CNN; (d) 3D-CNN; (e) PolMPCNN; (f) CEGCN; (g) SGCN-CNN; (h) MLDnet.
Remotesensing 17 01422 g015
Figure 16. Classification results of different features on the Xi’an data set. (a) The label map of Xi’an data set; (b) Feature1; (c) Feature2; (d) Feature3; (e) Multi-features.
Figure 16. Classification results of different features on the Xi’an data set. (a) The label map of Xi’an data set; (b) Feature1; (c) Feature2; (d) Feature3; (e) Multi-features.
Remotesensing 17 01422 g016
Figure 17. Overall classification accuracy of five data sets under different sampling window sizes.
Figure 17. Overall classification accuracy of five data sets under different sampling window sizes.
Remotesensing 17 01422 g017
Figure 18. Overall classification accuracy of training samples of five data sets at different sampling ratios.
Figure 18. Overall classification accuracy of training samples of five data sets at different sampling ratios.
Remotesensing 17 01422 g018
Table 1. 57-dimension features list.
Table 1. 57-dimension features list.
FeatureFeature NameFeature ParameterNumber
Original data featuresScattering matrix
elements
r e a l ( S h h ) , i m g ( S h h ) , r e a l ( S h v ) , i m g ( S h v ) , r e a l ( S v v ) , i m g ( S h v ) 6
Coherency matrix
elements
T 11 , T 22 , T 33 , r e a l ( T 12 ) , i m g ( T 12 ) , r e a l ( T 13 ) , i m g ( T 13 ) , r e a l ( T 23 ) , i m g ( T 23 ) 9
SPAN s p a n = S h h 2 + S h v 2 + S v v 2 1
Target decomposition
features
Cloud and Pottier
decomposition
H, A, α 3
Freeman decompositionthe surface, double-bounce and volume scattering power3
Huynen decomposition A 0 , B 0 , B , C , D , E , F , G , H 9
Co-polarization ratio r o = S v v S v v * S h h S h h * 1
Cross-polarization ratio r x = S h v S h v * S h h S h h * 1
Textural and contour
features
GLCM features C o n t r a s t : c o n = i j ( i j ) 2 P i , j 4
E n e r g y : A s m = i j P i , j 2 4
E n t r o p y : E n t = i j P i , j log P i , j 4
R e l a t i v i t y : C o r r = i j ( ( i , j ) p ( i , j ) ) μ x μ y σ x σ y 4
Edge-line energy
features
E e d g e = 1 n i = 1 n w i x i ( 1 m j = 1 m w j x j ) 4
E l i n e = min { E edge ij , E edge jk } 4
Total57
Table 2. Parameter comparison between DeeplabV3+ and L-DeeplabV3+ networks.
Table 2. Parameter comparison between DeeplabV3+ and L-DeeplabV3+ networks.
NetworkDeeplabV3+L-DeeplabV3+
Input feature map 100 × 100 16 × 16
Backbone networkXception networkDWConv 5 × 5 + Maxpooling + DWConv 3 × 3
ASPPConv 3 × 3 , r = 6, 12, 18Conv 3 × 3 , r = 2, 4, 6
Table 3. Classification accuracy of different methods on Xi’an Data Set (%).
Table 3. Classification accuracy of different methods on Xi’an Data Set (%).
ClassSuper-RF [14]CNN [18]CV-CNN [20]3D-CNN [57]PolMPCNN [58]CEGCN [59]SGCN-CNN [60]MLDnet
Water70.9193.1297.2690.2795.5294.4786.5794.10
Grass94.9792.8285.9793.6090.9596.5093.3397.72
Building90.9485.3191.4393.9197.6896.5490.1298.28
OA89.9490.2189.5993.2194.0196.2191.1897.38
AA85.6190.4291.5592.6094.7195.8390.0196.70
Kappa83.0283.8383.1888.7790.2593.7485.3395.67
The highest accuracy in each line is highlighted in bold.
Table 4. Classification accuracy of different methods on Oberpfaffenhofen Data Set (%).
Table 4. Classification accuracy of different methods on Oberpfaffenhofen Data Set (%).
ClassSuper-RF [14]CNN [18]CV-CNN [20]3D-CNN [57]PolMPCNN [58]CEGCN [59]SGCN-CNN [60]MLDnet
Bare ground91.7685.9568.8691.7686.9293.1492.9496.97
Forest85.2683.2281.1684.5982.5190.1381.9696.17
Building59.5583.5087.5283.9185.8294.4687.6095.77
Farmland57.2853.4470.4965.3365.6580.7032.0295.50
Road49.1967.4176.2050.3410.3866.1414.6793.78
OA78.4080.8774.8882.8275.8288.9377.2996.19
AA68.5974.7076.8575.1966.2684.9161.8495.64
Kappa67.4272.0365.7374.2863.4683.6664.7694.44
The highest accuracy in each line is highlighted in bold.
Table 5. Classification accuracy of different methods on San Francisco Data Set (%).
Table 5. Classification accuracy of different methods on San Francisco Data Set (%).
ClassSuper-RF [14]CNN [18]CV-CNN [20]3D-CNN [57]PolMPCNN [58]CEGCN [59]SGCN-CNN [60]MLDnet
Ocean99.9899.5999.9910099.5899.99100100
Vegetation93.8993.5196.4294.8095.6397.8797.4099.89
LD urban97.3194.4794.5197.9699.3999.5899.4899.97
HD urban77.7692.5796.3797.6198.5799.4999.4899.98
Developed81.0090.3195.9296.2596.8199.8899.95100
OA94.3396.2897.7198.3898.9399.5599.4799.97
AA89.9994.0996.6497.3298.0899.3699.2699.97
Kappa91.8194.6596.7097.6698.4699.3599.2499.96
The highest accuracy in each line is highlighted in bold.
Table 6. Classification accuracy of different methods on Flevoland1 Data Set (%).
Table 6. Classification accuracy of different methods on Flevoland1 Data Set (%).
ClassSuper-RF [14]CNN [18]CV-CNN [20]3D-CNN [57]PolMPCNN [58]CEGCN [59]SGCN-CNN [60]MLDnet
Urban81.8488.8796.2694.7496.2999.4599.7399.98
Water98.6999.6899.8598.8799.1499.6599.7799.99
Woodland94.9295.5996.4896.0398.7798.9699.5899.89
Cropland94.1693.1693.9496.7598.6699.4799.6399.92
OA93.8895.0196.5796.8498.4999.3799.6799.94
AA92.4094.3396.6396.6098.2199.3899.6899.95
Kappa91.6193.1895.3395.6997.9499.1499.5599.92
The highest accuracy in each line is highlighted in bold.
Table 7. Classification accuracy of different methods on Flevoland2 Data Set (%).
Table 7. Classification accuracy of different methods on Flevoland2 Data Set (%).
ClassSuper-RF [14]CNN [18]CV-CNN [20]3D-CNN [57]PolMPCNN [58]CEGCN [59]SGCN-CNN [60]MLDnet
Stem beans96.7799.5899.7288.4599.7998.5495.4199.96
Peas98.6497.9599.9985.9099.0085.7099.63100
Forest95.8897.3699.8299.8599.9799.9599.8999.98
Lucerne96.6388.9598.1796.2698.4510099.8499.98
Beets99.0593.8698.6699.2195.5299.4998.9199.97
Wheat95.7097.4399.2298.5798.0098.5497.6099.69
Potatoes96.0494.9899.1498.1498.7699.4099.5699.96
Bare soil94.5799.0910099.7199.8410010099.89
Grasses84.0376.5199.9476.5095.7188.3290.9699.68
Rapeseed53.1383.9094.1894.8393.5599.3097.7999.98
Barley10081.7599.5875.3640.2510091.74100
Wheat279.9395.2699.4197.0495.9799.9599.85100
Wheat399.3998.4799.3799.7793.8899.9799.7899.97
Water10099.8199.9910091.2710099.96100
Buildings086.0110050.6396.0180.8896.22100
OA92.1893.9698.9695.2893.8298.3298.5099.95
AA85.9892.7399.1590.6893.0696.6797.8199.94
Kappa91.4493.4098.8794.8493.2798.1698.3699.95
The highest accuracy in each line is highlighted in bold.
Table 8. Classification accuracy of different submodule networks on five experimental data sets (%).
Table 8. Classification accuracy of different submodule networks on five experimental data sets (%).
Data SetXi’anOberpfaffenhofenFlevoland1Flevoland2San Francisco
AccuracyOAKappaOAKappaOAKappaOAKappaOAKappa
wihtout MF96.8694.8393.8291.0499.1998.8998.9298.8299.8099.72
without L-deeplabV3+96.7394.6295.5093.4799.8399.7899.7799.7599.9699.95
without Attention96.2993.9396.5294.9399.8699.8199.5599.5099.7599.64
MLDnet97.3895.6796.7695.2999.9499.9299.9599.9599.9799.96
The highest accuracy in each line is highlighted in bold.
Table 9. Classification accuracy of Deeplabv3+ and L-deeplabV3+ networks on Flevoland2 data set (%).
Table 9. Classification accuracy of Deeplabv3+ and L-deeplabV3+ networks on Flevoland2 data set (%).
ClassDeeplabV3+ [31]L-DeeplabV3+
Stem beans95.6199.96
Peas70.36100
Forest89.4999.97
Lucerne97.5199.87
Beets94.72100
Wheat84.7499.25
Potatoes92.9399.40
Bare soil97.23100
Grasses96.0499.96
Rapeseed95.8499.38
Barley96.71100
Wheat292.5896.52
Wheat384.4199.84
Water90.43100
Buildings98.4090.90
OA97.5999.55
Kappa92.8799.50
The highest accuracy in each line is highlighted in bold.
Table 10. Classification accuracy of different types of feature input on Xi’an data set (%).
Table 10. Classification accuracy of different types of feature input on Xi’an data set (%).
ClassFeature 1Feature 2Feature 3Multi-Features
Water91.2593.7989.2494.10
Grass95.3189.5797.1597.72
Building95.3189.5797.1598.28
OA91.3993.2595.9897.38
AA90.8393.8594.5396.70
Kappa85.6989.0093.3395.67
The highest accuracy in each line is highlighted in bold.
Table 11. Classification accuracy of different positions and numbers of channel attention modules on Xi’an data set (%).
Table 11. Classification accuracy of different positions and numbers of channel attention modules on Xi’an data set (%).
ClassFront1Front2Front3Behind1Behind2Behind3
Water86.9577.9179.6594.1080.4879.99
Grass92.2796.0094.0697.7297.5792.03
Building94.9092.3493.9098.2888.7894.43
OA92.4091.9991.8597.3891.9191.07
AA91.3788.7589.2096.7088.9488.82
Kappa87.5086.5886.4095.6786.3885.20
The highest accuracy in each line is highlighted in bold.
Table 12. Running time of different methods on Xi’an data set (s).
Table 12. Running time of different methods on Xi’an data set (s).
Super-RFCNNCV-CNN3D-CNNPolMPCNNCEGCNSGCN-CNNDeeplabV3+MLDnet
Training59.22150.593463.20121.8421,600.35129.4330.591543.21511.10
Testing1.8510.5638.4322.80327.534.453.0053.5023.48
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shi, J.; Ji, S.; Jin, H.; Zhang, Y.; Gong, M.; Lin, W. Multi-Feature Lightweight DeeplabV3+ Network for Polarimetric SAR Image Classification with Attention Mechanism. Remote Sens. 2025, 17, 1422. https://doi.org/10.3390/rs17081422

AMA Style

Shi J, Ji S, Jin H, Zhang Y, Gong M, Lin W. Multi-Feature Lightweight DeeplabV3+ Network for Polarimetric SAR Image Classification with Attention Mechanism. Remote Sensing. 2025; 17(8):1422. https://doi.org/10.3390/rs17081422

Chicago/Turabian Style

Shi, Junfei, Shanshan Ji, Haiyan Jin, Yuanlin Zhang, Maoguo Gong, and Weisi Lin. 2025. "Multi-Feature Lightweight DeeplabV3+ Network for Polarimetric SAR Image Classification with Attention Mechanism" Remote Sensing 17, no. 8: 1422. https://doi.org/10.3390/rs17081422

APA Style

Shi, J., Ji, S., Jin, H., Zhang, Y., Gong, M., & Lin, W. (2025). Multi-Feature Lightweight DeeplabV3+ Network for Polarimetric SAR Image Classification with Attention Mechanism. Remote Sensing, 17(8), 1422. https://doi.org/10.3390/rs17081422

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop