Next Article in Journal
A Multi-Active and Multi-Passive Sensor Fusion Algorithm for Multi-Target Tracking in Dense Group Clutter Environments
Previous Article in Journal
Evaluation of the Predictive Capability of CMA Climate Prediction System Model for Summer Surface Heat Source on the Tibetan Plateau
Previous Article in Special Issue
CPINet: Towards A Novel Cross-Polarimetric Interaction Network for Dual-Polarized SAR Ship Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network

1
Hubei Key Laboratory of Optical Information and Pattern Recognition, School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
2
School of Geosciences and Infophysics, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(21), 4119; https://doi.org/10.3390/rs16214119
Submission received: 20 August 2024 / Revised: 30 October 2024 / Accepted: 31 October 2024 / Published: 4 November 2024
(This article belongs to the Special Issue SAR in Big Data Era III)

Abstract

:
Supervised polarimetric synthetic aperture radar (PolSAR) image classification demands a large amount of precisely labeled data. However, such data are difficult to obtain. Therefore, many unsupervised methods have been proposed for unsupervised PolSAR image classification. The classification maps of unsupervised methods contain many high-confidence samples. These samples, which are often ignored, can be used as supervisory information to improve classification performance on PolSAR images. This study proposes a new unsupervised PolSAR image classification framework. The framework combines high-confidence superpixel pseudo-labeled samples and semi-supervised classification methods. The experiments indicated that this framework could achieve higher-level effectiveness in unsupervised PolSAR image classification. First, superpixel segmentation was performed on PolSAR images, and the geometric centers of the superpixels were generated. Second, the classification maps of rotation-domain deep mutual information (RDDMI), an unsupervised PolSAR image classification method, were used as the pseudo-labels of the central points of the superpixels. Finally, the unlabeled samples and the high-confidence pseudo-labeled samples were used to train an excellent semi-supervised method, similarity matching (SimMatch). Experiments on three real PolSAR datasets illustrated that, compared with the excellent RDDMI, the accuracy of the proposed method was increased by 1.70%, 0.99%, and 0.8%. The proposed framework provides significant performance improvements and is an efficient method for improving unsupervised PolSAR image classification.

1. Introduction

Polarimetric synthetic aperture radar (PolSAR) has drawn significant attention because it can operate all day and night. A polarization scattering matrix can provide rich feature information that is widely used in land cover classification. In the field of PolSAR image classification, various unsupervised, semi-supervised, and supervised methods have been proposed. A large number of accurately labeled data are necessary in supervised classification methods. Unsupervised learning only uses the similarity relationships between samples to train models without labeled data. PolSAR images often do not have sufficient accurately labeled samples. The accurately labeled samples require significant manpower and time. Therefore, people are increasingly concerned about unsupervised methods.
Recently, many unsupervised methods have been proposed.The maximum likelihood classification [1] has been widely used and developed. This method proposed a new distance metric derived based on the complex Wishart distribution [2]. Pottier et al. [3] proposed a method using H / α to initialize the iterative Wishart classifier. Cao et al. [4] referred to this idea and used H / α / A S p a n to initialize the Wishart classifier. Lee et al. [5] also proposed a robust classification method to maintain the uniform scattering mechanisms of various classes. Freeman decomposition [6] and the Wishart classifier are combined in this method. In [7], a new spectral clustering method was proposed based on the Wishart-derived distance metric to construct a similarity matrix. Additionally, spectral clustering was popular due to its excellent performance and its well-defined framework. Song et al. [8] proposed a spectral clustering affinity matrix that was easy to process and saved memory, reducing the enormous computational resources required for clustering on large-sized PolSAR images. Yang et al. [9] proposed that kernel fuzzy C-means achieved better clustering results for PolSAR images. In [10], Yang et al. utilized the divergence of analytic information theory to construct an affinity matrix and proposed a kernel fuzzy similarity measurement method based on the membership distribution using the fuzzy C-means method. This method performed well in the spectral clustering of PolSAR images. However, traditional unsupervised clustering methods still suffer from the problem of category boundaries not matching the true distribution of the data.
Deep learning has been further developed. More and more unsupervised deep learning methods have been proposed in the field of PolSAR image classification. The deep cluster [11] used k-means to calculate pseudo-labels and trained deep neural networks for supervision. Haeusser et al. [12] proposed an associated deep clustering framework that did not use the clustering model with the feature maps extracted by the network. This framework directly utilized deep neural networks for clustering. Ji et al. [13] proposed deep subspace clustering networks (DSCs) that combined deep learning with subspace clustering. Compared with traditional autoencoders, DSCs added a sub-representation layer for low-dimensional feature space. Song et al. [14] introduced DSCs to generative adversarial networks (GANs) and proposed deep adversarial subspace clustering (DASC), which performed impressively on real-world data with complex subspaces. However, these deep learning methods have not yet effectively integrated traditional clustering techniques, and they have not implemented robust semantic filtering mechanisms to enhance classification accuracy. Feature post-processing and clustering mechanisms outside the network make these clustering methods more cumbersome [15]. In [15], invasive information clustering (IIC) was proposed, which involved slight modifications to the convolutional neural network (CNN) by maximizing the objective function with mutual information and building a two-headed CNN architecture. Unsupervised learning has great research value. A deep belief network (DBN) [16] was applied to PolSAR image classification [17]. The DBN is a classic deep learning probabilistic generative network composed of multiple stacked restricted Boltzmann machines (RBMs) [18]. The superpixel segmentation algorithm was introduced to the autoencoder (AE) network to extract the neighborhood information of PolSAR images through the unsupervised learning of polarization features [19]. Based on the AE, overcomplete sparse features were extracted from hidden layers, and a stacked sparse autoencoder (SSAE) network was proposed to improve the accuracy of unsupervised classification further [20]. Then, the SSAE added adaptive non-local spatial information to obtain robust polarization features [21]. These methods can effectively learn the deep features of polarimetric data. Bi et al. combined polarimetric image decomposition with deep convolutional networks within a principled framework, introducing a new unsupervised learning classification method [22]. Wang et al. [23] proposed a model known as rotation-domain deep mutual information (RDDMI) that operates on a convolutional long short-term memory (ConvLSTM) network. RDDMI combined the deep mutual information of IIC with deep comprehensive correlation mining (DCCM) [24]. The performance of unsupervised PolSAR image classification has been elevated to a better level. Zuo et al. proposed deep similarity clustering (DSC). This newest model combined an unsupervised feature extraction pipeline with a Wishart distance metric and a deep clustering pipeline with a feature similarity metric. Moreover, the regularization combined two parts to maintain the clarity of edges and the semantic continuity of image content [25]. The DSC has achieved very good unsupervised classification results. Deep learning has made significant progress in the field of unsupervised PolSAR image classification. The potential applications of the results generated through unsupervised learning deserve further exploration. Unsupervised learning, as a powerful tool, can help us identify potential structures and patterns in a dataset during the initial analysis stage. At present, few methods further analyze the results of unsupervised classification of PolSAR images. In the maps of the unsupervised classification of PolSAR images, there are a large number of high-confidence samples that have not been fully utilized. These high-confidence samples can be transformed into high-confidence pseudo-labels and effectively applied in supervised or semi-supervised learning frameworks to make the result of PolSAR image classification more accurate.
Semi-supervised learning can still demonstrate good performance with only a small number of labeled samples. Therefore, the pseudo-labels generated through unsupervised learning are particularly suitable for semi-supervised learning methods. The classic two-stage training paradigm is simple yet efficient in semi-supervised learning methods. For example, the residual network (ResNet) [26], visual transformer (ViT) [27], and Swin transformer (Swin-T) [28] are typically pre-trained on large-scale datasets in a supervised manner, and then a small number of labeled samples are used to fine-tune the pre-trained model. Currently, pseudo-labeling [29] and consistency regularization [30] are used in the joint learning paradigm to label samples directly. MixMatch [31] made full use of this idea to generate pseudo-labels by sharpening the average of multiple strongly augmented views. The MixUp trick was used in MixMatch to augment pseudo-labels [32]. ReMixMatch [33] introduced weakly augmented views to generate pseudo-labels and used an alignment strategy to encourage pseudo-labels to align with the edge distribution of labeled samples. Fixmatch [34] adopted a simpler approach, as it retained the model but used only high-confidence pseudo-labels. It greatly simplified the idea of augmenting the anchoring in consistency regularization and produced better results. These methods used labeled samples to train semantic classifiers and treated the predicted results as pseudo-labels of unlabeled samples. However, models often fit overconfident but incorrect pseudo-labels, leading to inaccurate results when simply using pseudo-labels. To solve this problem, similarity matching (SimMatch) [35] optimizes the process of generating pseudo-labels in Fixmatch, taking both semantic and instance similarity into account. SimMatch guided the better propagation of the two similarities by a memory buffer. SimMatch achieved outstanding performance in semi-supervised classification. The smooth integration of the high-confidence samples contained in unsupervised learning classifications into semi-supervised learning frameworks deserves further study. Benefiting from the excellent performance of semi-supervised learning methods, the performance of unsupervised classification can be further improved. The high-confidence samples from unsupervised classification maps of PolSAR images, naturally, can be used as pseudo-labels. By utilizing the advantages of both high-confidence pseudo-labels and advanced semi-supervised classification algorithms, unsupervised PolSAR image classification performance can be further improved.
This study proposes a high-confidence unsupervised pseudo-label learning classification framework for PolSAR images based on representative points of superpixels. The framework segments a PolSAR image through simple linear iterative clustering (SLIC) [36] and extracts the geometric centers of the superpixels as representative points. Then, the generated maps of an advanced unsupervised classification method, RDDMI, are combined with the representative points as high-confidence pseudo-labels. Finally, the pseudo-labels with high confidence are used as supervision for the excellent semi-supervised algorithm SimMatch to train and obtain the final classification results.
The following are the main contributions of this study: (1) This study proposes a novel unsupervised PolSAR image classification framework called superpixel pseudo-label similarity matching (SP-SIM) based on the extraction of high-confidence pseudo-labels through superpixels. (2) This study utilizes classification maps from an advanced unsupervised classification algorithm, RDDMI, as pseudo-labels. Superpixels are used as the basic processing unit to obtain pseudo-labels with high confidence quickly. (3) This study introduces the semi-supervised algorithm SimMatch, which has excellent performance. In this framework, semantic and instance pseudo-labels can propagate to each other and achieve more intrinsic feature matching. It also uses high-confidence pseudo-labels as supervision for learning, making full use of the feature information of pseudo-labels. The high-confidence pseudo-labels with SimMatch obtain excellent classification results.
The results were extensively evaluated through experiments on three real PolSAR images and compared with those of other advanced unsupervised classification algorithms for PolSAR images. The accuracy reached the optimal level, verifying the effectiveness and superiority of SP-SIM.

2. Methods

Figure 1 illustrates the five parts of the proposed framework. The first part shows the steps in the generation of representative points of superpixels. The second part shows the preliminary generation of maps by the unsupervised learning algorithm RDDMI. The third part shows how to use representative points of superpixels to improve the confidence of the maps. The three parts mentioned above constitute the method of how high-confidence pseudo-labels are generated. The fourth part is the semi-supervised algorithm SimMatch. SimMatch uses high-confidence pseudo-labels as supervision to generate pseudo-labels from unlabeled samples. Two types of losses are calculated. The fifth part is a typical use of high-confidence pseudo-labels as supervision to generate supervised predictions. In this part, the pseudo-labels and features are also stored in the memory buffer of the fourth part.
The first part presents the selection of representative points of superpixels using SLIC. The pixels in a PolSAR Pauli pseudo-color image are grouped according to the similarity of their color, brightness, and other characteristics. The generated superpixels are compact and neat like cells, and the neighborhood features are relatively easy to express. SLIC outperforms other superpixel segmentation methods in terms of contour preservation, running speed, and the compactness of the generated superpixels.
In the second part, the unsupervised learning algorithm RDDMI is used to generate preliminary pseudo-label maps. For each pixel in the PolSAR image, paired data x and x 0 are used as inputs for deep mutual information learning. The RDDMI generates preliminary unsupervised classification maps for the creation of pseudo-labels.
In the third part, this study refers to the contour of each superpixel to find its approximate maximum inscribed rectangle and geometric center. The shape of a single superpixel contour may be irregular, resulting in the inscribed rectangle being too small and the central representative point being inaccurate. Therefore, a certain number of contour pixels are allowed to exist in the inscribed rectangle in this process. The points corresponding to the central representative points in the maps generated by RDDMI are selected as pseudo-labels with high confidence.
In the fourth part, high-confidence pseudo-labels are used as supervision through the excellent semi-supervised algorithm SimMatch. SimMatch first uses weakly augmented views to generate semantic and instance pseudo-labels and calculates their similarity through the class center and label embedding. Then, these two similarities are fused through expansion and aggregation operations. The strongly augmented view is used to generate predictions and feature similarities. The semantic pseudo-labels and instance pseudo-labels have instance pseudo-labels and semantic pseudo-labels, respectively. Finally, the unsupervised loss L u and input loss L i n are calculated by the predictions and similarities.
The fifth part is the process of calculating the supervision loss. The high-confidence pseudo-labels mentioned above are used as supervision. “Labels” and feature embeddings of high-confidence pseudo-label samples are stored in a labeled memory buffer.
Figure 1 shows the entire process of the algorithm proposed in this paper. First, RDDMI unsupervised learning classification is used to obtain primary pseudo-labels that contain pseudo-supervision information. Second, the PolSAR image is segmented by SLIC, and the geometric centers of the approximate maximum inscribed rectangles of superpixels are calculated. The centers are combined with the primary pseudo-label to extract high-confidence pseudo-labels. Third, the pseudo-labels are used for classification tasks through the semi-supervised algorithm SimMatch. In SimMatch, the memory buffer stores the features obtained from high-confidence pseudo-labels and performs similarity matching with the prediction results of unlabeled data to obtain more accurate results. Finally, SimMatch uses supervised loss L s , unsupervised loss L u , and consistency regularization loss L i n to jointly form the overall loss function. The generalization ability and the performance of the model are guaranteed.

2.1. Method of Obtaining Pseudo-Labels

The method of obtaining pseudo-labels mainly includes three parts: (1) Using the RDDMI algorithm to generate pseudo-labels with low confidence; (2) using the SLIC algorithm to segment PolSAR images and calculate the representative points of each superpixel; (3) combining representative points of superpixels with low-confidence maps to extract high-confidence pseudo-labels.

2.1.1. RDDMI

The RDDMI algorithm represents a significant advancement in the realm of unsupervised learning for PolSAR image classification. In RDDMI, an end-to-end convolutional long short-term memory (ConvLSTM) network was used. The data processing workflow is simplified. The deep mutual information associated with various polarization directions (POAs) in the rotation domain of the polarization coherence matrix is effectively learned. RDDMI has good classification accuracy and rapid processing capabilities, allowing for the swift generation of accurate pseudo-labels for subsequent analysis. ConvLSTM has taken the place of LSTM’s [38] fully connected gate layers with convolutional layers, enabling it to process sequence data with spatial information.
The architecture of the RDDMI network initiates with two ConvLSTM layers to extract rotational-domain features from the inputs. Next, three convolutional layers, three max-pooling layers, and two fully connected layers are applied to further refine and learn deep features. The network culminates in a softmax layer to output a one-hot vector, with the argmax function being employed to calculate the class of the sample [23].
As shown in Equation (1), the loss function includes pseudo-label loss, pseudo-graph loss, and two mutual losses based on the predicted features of x and x 0 information losses. RDDMI first introduces pseudo-label supervised loss, pseudo-graph supervised loss, and triple mutual information loss into DCCM [23]. The similarity matrix between samples is calculated using the predictive properties of the network. The shallow features output by the first max-pooling layer and the deep features output by the second fully connected layer are used to calculate the mutual loss of triples.
min Φ L = L I I C _ M I + α L P G ^ + β L P L ^ + γ L T R I _ M I

2.1.2. Representative Points of Superpixels

Superpixels are a group of pixels that are combined based on similar colors or low-level features. In the unsupervised classification of PolSAR images, there are common phenomena of blurred boundaries and inaccurate boundary classifications. This will affect the overall classification accuracy and lead to low overall confidence in pseudo-labels. Reference superpixel segmentation overcomes speckle noise by extracting color features. The PolSAR image is divided into several superpixels with similar color features using superpixel segmentation. The geometric center point of each superpixel is used as an accurate and representative pseudo-label point. This approach abandons a large amount of boundary information that is highly likely to be erroneous, increasing the confidence of pseudo-labels. Therefore, there is an urgent need for an accurate and fast superpixel segmentation method. In existing methods, the superpixels generated by SLIC have precise boundary adhesion and uniform regularity. SLIC requires very few parameters to be set, and by default, only the number of pre-segmented superpixels needs to be set. Therefore, this study uses SLIC to quickly segment PolSAR images into many superpixels. Then, the approximate maximum inscribed rectangle of each superpixel is calculated. This rectangle allows an appropriate number of pixels on the superpixel outline inside the rectangle to ensure that it is centered, and the largest rectangle is calculated. At this point, the center point of the rectangle can also be considered as the center point of the superpixel, which is representative. Finally, the representative points are combined with the maps generated by the RDDMI mentioned above to form a high-confidence pseudo-label.

2.2. Framework of SimMatch Semi-Supervised Learning

The extracted pseudo-labels have high confidence, but the quantity is limited. The situation at this point is similar to semi-supervised learning. A small amount of labels are used in semi-supervised learning, so a semi-supervised learning algorithm is needed to complete the remaining processing work. In [39], SimMatch was used to detect changes in PolSAR images, and the effect was significant. SimMatch is a suitable semi-supervised image classification algorithm with good performance under limited label conditions. This is a way to propagate pseudo-labels’ information to each other, and it uses the label memory buffer to achieve the isomorphic transformation of two types of similarity. SimMatch solves the problems of model overfitting and improves the reliability of the model. SimMatch effectively utilizes the idea of combining consistency regularization with pseudo-labels in FixMatch [34]. This section will first review consistency regularization and pseudo-labels and then introduce the innovation of SimMatch, which allows pseudo-label information to mutually propagate.
Firstly, the image classification problem can be defined as there being a set of labeled samples X = ( x b , p b ) : b ( 1 , B ) and unlabeled samples u = u b : b ( 1 , μ B ) , where p b is the label of sample x b . Data augmentation is often used to improve the generalization ability of models. The weakly augmented function and strongly augmented function are defined as T w · and T s · .

2.2.1. Two Methods in Semi-Supervision

Consistency regularization is frequently used in semi-supervised learning. Consistency regularization was first proposed in [40], and it requires ensuring that the same image still has similar prediction results under different perturbations. In the models in [41], the loss function is used for unlabeled samples as follows:
b = 1 μ B p m ( y T w ( u b ) ) p m ( y T w ( u b ) ) 2 2
Here, p m ( y x ) is the predicted classification result of input x, and both T w and p m are random. SimMatch uses cross-entropy loss instead of square loss.
Pseudo-labeling is also commonly used in semi-supervised learning. Pseudo-labels are obtained from the model itself. Specifically, during supervised training, the prediction with a maximum class probability higher than the threshold τ is retained. This prediction is used as the pseudo-label of the unlabeled sample. It can be expressed by the following loss function:
1 μ B b = 1 μ B 1 ( max ( p m ( y u b ) ) τ ) H ( q ^ b , p m ( y u b ) )
where q ^ b = a r g max ( q b ) . FixMatch applies arg max to the prediction to generate an effective one-hot prediction. Differently, in SimMatch, a distribution alignment strategy D A ( · ) [33] is adopted to balance the distribution of pseudo-labels, and D A ( p w ) is directly used as the pseudo-label to calculate the cross-entropy between the strongly augmented sample predictions p s :
1 μ B 1 ( max ( D A ( p w ) ) > τ ) H ( D A ( p w ) , p s )

2.2.2. Instance Similarity

SimMatch encourages the similarity distributions of strongly and weakly augmented views to align closely. A nonlinear head g ( · ) representing h maps to a feature embedding z b = g h b is assumed. The augmented anchoring and the embeddings can be denoted by z b W (the weakly augmented views) and z b s (the strongly augmented views), respectively. For the K embeddings of various weakly augmented samples z k : k ( 1 , K ) , the similarity with z W and the i -th instance are calculated by a similarity function s i m · . This is represented by the dot product between L 2 and the normalized vector sim ( u , v ) = u T v / / / u / / / / v / / . Then, the generated similarity is processed through a softmax layer. In the generated similarity distribution, t is used as a temperature parameter to adjust the sharpness of the distribution:
q i W = exp sim z b W , z i / t k = 1 K exp sim z b w , z k / t
The sim z b s , z k is used to calculate the similarity between strongly augmented views z S and z i to obtain the similarity distribution:
q i s = e x p ( s i m ( z b s , z i ) / t ) k = 1 K e x p ( s i m ( z b s , z k ) / t )

2.2.3. Label Information Dissemination

The SimMatch framework also considers instance-level consistency regularization. The use of entirely unsupervised instance pseudo-labels, denoted as q W , leads to a considerable waste of label information. To address this issue, SimMatch utilizes labeled information at the instance level. The interaction between semantic and instance similarity is enhanced, thereby improving the result of the generated pseudo-labels. The labeled memory buffer saves all labeled samples, as shown in Figure 2. Each z k in Equations (5) and (6) can be associated with a specific class. The vectors in ϕ are regarded as center class references when the labeled samples stored in the memory buffer can be viewed as the same class. p w R 1 × L is semantic similarity and q w R 1 × K is instance similarity. Both of them are computed by using weakly augmented samples. It should be noted that the value of L is typically much smaller than that of K because each class requires at least one sample. Subsequently, q W is calibrated by expanding p w (referred to as p u n f o l d ). Aligning the semantic similarities embedded within labels leads to the following results:
p i u n f o l d = p j w , w h e r e c l a s s ( q j w ) = c l a s s ( p i w )
In Equation (7), c l a s s ( · ) can return the basic truth class. The c l a s s ( q j w ) indicates the label of the jth element in the memory buffer and represents the ith class. Then, the calibrated instance pseudo-labels are regenerated by scaling q W through p u n f o l d and can be expressed as
q ^ i = q i w p i u n f o l d k = 1 K q k w p k u n f o l d
The old labels q W will be updated to the pseudo-labels q ^ of the instance. At the same time, q will be aggregated into an L-dimensional space. The instance similarity is combined with the semantic similarity, represented as q a g g . Then, the instance similarity is summed to achieve the sharing of the same ground-truth labels:
q i m a g g = j = 0 K 1 ( c l a s s ( p i w ) = c l a s s ( q j w ) ) q j w
The semantic pseudo-labels will be updated by using q a g g to smooth p W :
p ^ i = α p i w + ( 1 a ) q i a g g
In Equation (10), α is the hyperparameter that controls the instance information and semantic weight. The pseudo-label p will contain semantic-level information and q will contain instance-level information. Similarly, the old value, p i W , will be replaced with the adjusted semantic pseudo-label. If the two are not close, the histogram will be flatter. As shown in Figure 3, if these two similarities are very close, it means that the predictions of the two similarities’ distributions are consistent, and the histogram will be sharp.

2.2.4. Memory Buffer

SimMatch stores ground-truth labels and feature embeddings of labeled samples in a memory buffer. Specifically, it delineates a label memory buffer, denoted as P l R N × 1 , and a feature memory buffer, referred to as P f R N × D , where D represents the embedding size and N represents the amounts of labeled samples. For the label memory buffer, each label only stores one scalar. Aggregation and unfolding operations are implemented through functions provided by a deep learning library [35]. Given the variance in buffer sizes, SimMatch employs two distinct implementations. When N is large, the memory buffer is represented as M s and M t using the student–teacher framework [42]. The labeled samples and strongly augmented samples are directed to M t , while weakly augmented samples are input into M t to generate pseudo-labels. The updating process for M t is as follows:
M t ( 1 m ) M s + m M t
When N is small, a time integration strategy [43] is used to smooth the feature embeddings, and this can be defined as
Y t ( 1 m ) Y t + m Y t 1
In this case, the same encoder will receive all samples.

2.3. Loss

In SimMatch, the feature information h = f ( T ( x ) ) of weakly augmented labeled samples x = { x b : b ( 1 , , B ) } is extracted by encoder F ( · ) of the convolutional neural network. The fully connected prediction head ϕ ( · ) is used to calculate the semantic similarity p = ϕ ( h ) and the supervised classification loss as follows:
L s = 1 B Σ H ( y , p )
Applying the weakly augmented function T w · and the strongly augmented function T s · to unlabeled samples to obtain weakly augmented samples p w and strongly augmented samples p s , D A ( · ) is a distribution alignment method used to balance the distribution of pseudo-labels [33]. Then, it is necessary to keep the moving average of p a v g w constant and use N o r m a l i z e ( p w / p a v g w ) to adjust the current p w [44]. D A ( p w ) is used directly as pseudo-labels. The cross-entropy between p w (pseudo-label) and p s (semantic-similarity) is used to define unsupervised loss:
L u = 1 μ B 1 ( m a x ( D A ( p w ) ) > τ ) H ( D A ( p w ) , p s )
By minimizing the difference between q w and q s , consistency regularization can be achieved, and this is represented by the cross-entropy as
L i n = 1 μ B H ( q w , q s )
The overall loss function of SimMatch is:
L o v e r a l l = L s + λ u L u + λ i n L i n

3. Experiment

To validate the feasibility of the proposed method, multiple comparative experiments were conducted on three PolSAR datasets. Four PolSAR image classification methods, which were the classic unsupervised method, the Wishart cluster, the deep learning-based method RDDMI, and the traditional supervised random forest (RF) method, were used to compare with the proposed method. The supervised training of Random Forest (RF) uses 25,000 randomly selected labeled samples for each PolSAR datum. During this process, the input data are 15 × 15 neighborhood window data, which is the same size as the neighborhood window used in the proposed method. At the same time, the same normalization method is used for data processing to ensure data consistency.
The experimental results were evaluated using the overall accuracy (OA), kappa, precision, recall, and F1 score.
The proportion of overall correct predictions is calculated by OA. The kappa was used for unbalanced samples. The precision reflects the proportion of samples belonging to a certain class among the samples predicted as that class by the model. Recall is defined as the proportion of samples that are correctly identified by the model as belonging to a certain class. The F1 score is used to comprehensively measure the precision and recall of the model.

3.1. Datasets

This section introduces three real PolSAR image data. The data details are shown in the Table 1.

3.1.1. RADARSAT-2 Flevoland Dataset

In the field of PolSAR image classification, the RADARSAT-2 (RS-2) Flevoland dataset is widely used. This dataset has an image size of 1400 × 1200 pixels and a spatial resolution of 12 × 8 m. It covers the Flevoland region of the Netherlands and has four land cover categories: farmland, water, forests, and buildings. The Pauli pseudo-color image and the corresponding ground-truth map of the RS-2 Flevoland dataset are shown in Figure 4.

3.1.2. RADARSAT-2 Wuhan Dataset

The RS-2 Wuhan dataset covers the scene of Wuhan and was obtained using the RADARSAT-2 C-band PolSAR system in the fine quad-pol mode. The image size is 5500 × 2400 pixels. The spatial resolution is 12 × 8 m. It contains three land cover categories: water, forest, and buildings. This dataset has highly dense buildings with different orientations, which increases the difficulty of clustering. The Pauli pseudo-color image and the corresponding ground-truth map of the RS-2 Wuhan dataset are shown in Figure 5.

3.1.3. AIRSAR Flevoland Dataset

The AIRSAR Flevoland dataset contains four-look fully polarimetric images and was acquired by the NASA/JPL AIRSAR L-band system. This dataset also covers the scene of Flevoland, the Netherlands. The image size is 750 × 1024 pixels. The spatial resolution is 6 × 12 m. It contains 11 types of land cover, including forests, wheat, and 9 other types. The Pauli pseudo-color image and the corresponding ground-truth map of the AIRSAR Flevoland dataset are shown in Figure 6.

3.2. Results

3.2.1. Results on the RS-2 Flevoland Dataset

The classification results on the RS-2 Flevoland dataset are shown in Figure 7 and Table 2. The RS-2 Flevoland dataset contains forest and farmland types embedded in building types, and farmland types have different backscattering properties, which poses difficulties for accurate classification. Wishart performed poorly on this dataset, and a large number of buildings and farmlands were identified as forests. RDDMI showed great classification performance and improved the OA to 87.86%. However, the pixels on the boundaries were not precisely classified. The proposed SP-SIM was superior to the above methods. In particular, the boundary pixels were finely classified, as shown in the black boxes in Figure 7c,d. The building and water types were all well classified. The boundaries of farmlands were also much better than those of RDDMI. Compared with RDDMI, the accuracies of the proposed method for water, forest, and farmland were increased by 0.61%, 4.17%, and 2.62%, respectively. Finally, the OA of the SP-SIM increased by 1.70%. Meanwhile, the OA of the SP-SIM was 1.46% higher than that of supervised RF. Moreover, the precision, recall, and F1 score are all higher than those of other methods, including RF. The above analysis demonstrates the obvious performance improvement and proves that the proposed framework is efficient for unsupervised PolSAR image classification.

3.2.2. Results on the RS-2 Wuhan Dataset

In the RS-2 Wuhan dataset, there is a large number of buildings with different orientations, which exhibit different features in PolSAR images. It is difficult to accurately distinguish building types through unsupervised methods. Figure 8 shows the classification maps, and Table 3 displays the quantitative results. RDDMI achieved very good results on this dataset, with impressive accuracy in identifying buildings. The results of the traditional Wishart cluster were not good. The OA of SP-SIM was 0.99% higher than that of RDDMI and close to that of the supervised RF method. The accuracies for all land cover types were better than those of RDDMI. The precision, recall and F1 score are all higher than those of the other methods. The experiment demonstrated the effectiveness of the proposed method on PolSAR images.

3.2.3. Results on the AIRSAR Flevoland Dataset

The AIRSAR Flevoland dataset has up to 11 categories, and the polarization properties of each category are also very complicated, which is undoubtedly a huge challenge for unsupervised classification. The backscattering properties of some land cover types have similarities in this dataset. In addition, there may be large differences in the polarization matrices observed for the same land cover type. As shown in Figure 9, the properties of water are similar to those of other categories, which will greatly affect the classification performance. The RF method can overcome these interpretation ambiguity problems to a certain extent by supervising the information. However, the unsupervised method makes it difficult to overcome this problem of interpretation ambiguity without supervised information, so the classification performance is poor. Moreover, this dataset contains many unlabeled areas; we only used labeled areas to evaluate the results.
In Figure 10 and Table 4, it can be seen that Wishart could not make accurate judgments on categories with similar polarization characteristics, and RDDMI achieved better results. All methods cannot accurately classify water. This is because the backscattering properties of water are similar to those of other categories. The proposed method had a certain improvement based on RDDMI, and the OA was increased by 0.80%. Other indicators also reached an excellent level. The experiment showed that although the proposed framework was able to improve the unsupervised PolSAR classification performance, it also had its limitations. The maps of RDDMI will affect the performance of the proposed method. If the results of RDDMI are poor, the performance of the proposed method will be limited.

4. Discussion

4.1. Effect of the Unsupervised Algorithm

To study the importance of the excellent unsupervised learning results of RDDMI to the proposed method, the pseudo-labels extracted from the Wishart and DCCM classification results are used for semi-supervised classification by SimMatch. These two methods are called SLIC-Wishart and SLIC-DCCM in the table, indicating the pseudo-labels extracted by the SLIC superpixel segmentation method. The results are shown in the Table 5, Table 6 and Table 7.
In the RS2-Flevoland and RS2-Wuhan datasets, high-confidence pseudo-label extraction was performed on the Wishart and DCCM results. The classification results of SLIC-Wishart and SLIC-DCCM were not as good as those obtained by using RDDMI for pseudo-label extraction as proposed in this paper. However, compared to the results of Wishart and DCCM, there was a significant improvement. All indicators in the results were improved to an extent, proving that the method of extracting high-confidence pseudo-labels through superpixels is feasible.
In the AIRSAR Flevoland dataset, the combination of RDDMI high-confidence pseudo-labels and SimMatch adopted in this paper has results far superior to those of extracting pseudo-labels with Wishart and DCCM, proving the viability of the scheme proposed above. The high-confidence pseudo-labels extracted by using the results of RDDMI still have the best performance, as mentioned above. The result of SLIC-Wishart is also better than that of Wishart. It can be inferred that the quality of high-confidence pseudo-labels extracted from unsupervised results using superpixels depends on the initial unsupervised classification results. The high-performance unsupervised classification results of RDDMI are very important to the results of this framework.
This raises a limitation: the method proposed requires high-quality initial unsupervised classification maps, which limits the quality of unsupervised classification results. The extraction of pseudo-labels through superpixel segmentation can only provide limited improvement for unsupervised algorithms with poor results. Further research is needed to address this limitation.

4.2. Effect of the Semi-Supervised Algorithm

The method proposed in this paper can extract large amounts of high-confidence pseudo-labels, which are directly used as supervision information. The supervision information can be used in the semi-supervised algorithm in the proposed method and can also be used in a supervised classification method. In this study, 25,000 superpixels were used to extract high-confidence pseudo-labels and conduct classification experiments on three datasets using the supervised learning classification method VGG-16 as a comparison. The 20,287 pseudo-label samples of the RS-2 Flevoland dataset were used. As shown in Table 8, the proposed semi-supervised method SP-SIM had an accuracy that was 1.83% higher than that of the supervised method, VGG-16, and the accuracies of each class were higher than those of VGG-16. Moreover, other indicators of SP-SIM are all superior to VGG-16. Although the number of pseudo-labeled samples was large, the effectiveness of using pseudo-labels as supervision information to directly train supervised models was not good. The semi-supervised classification method can effectively improve PolSAR image classification performance by combining pseudo-labeled samples and unlabeled samples.
In the other two datasets, it was also reflected that even with a large number of pseudo-labeled samples, the performance of semi-supervised methods was better than that of supervised methods; 20,127 pseudo-labeled samples were used to train classification models on the RS-2 Wuhan dataset. As shown in Table 9, the OA of SP-SIM was 0.44% higher than that of the supervised method, VGG-16, and the accuracies of all land cover types were all higher. In the AIRSAR Flevoland dataset, due to the large number of unrecognized regions, the number of pseudo-label samples was 5244, which was lower than that in the RS-2 Flevoland and RS-2 Wuhan datasets. However, under the same conditions, the OA of the proposed method was still 0.61% higher than that of the supervised method, as shown in Table 10.
The performance of the supervised classification method was not as good as that in semi-supervised learning methods with the same number of high-confidence pseudo-labels. Even with a large number of pseudo-labeled samples, the semi-supervised learning method outperformed the supervised learning method in terms of classification accuracy on the three datasets. This experiment’s results indicated that semi-supervised learning combined with unlabeled samples allows the more effective utilization of high-confidence pseudo-labeled samples to improve classification performance.

4.3. Effect of the Number of Superpixels

The proposed method utilizes superpixel segmentation. As the number of over-segmented superpixels increases, the number of high-confidence pseudo-label samples for unsupervised classification also increases. In this section, the impact of the number of superpixels in SLIC on classification performance is discussed. For each dataset, 6000 superpixels and 25000 superpixels were used to extract high-confidence pseudo-labels, and then 3730 and 20,287 high-confidence pseudo-labeled samples were obtained in the RS-2 Flevoland dataset. As shown in Table 11, the accuracy was increased by 0.91%. The increased number of pseudo-labeled samples further improved the classification performance. However, when the number of pseudo-labels reached a certain level, the accuracy improvement tended to stabilize.
In the AIRSAR Flevoland dataset, 953 and 5244 high-confidence pseudo-labeled samples were extracted from the superpixels. The classification results are shown in Table 12. The OA was improved by 0.51% with more pseudo-labeled samples. Other indicators are also better.
The impact of this number of pseudo-labels was even more limited in the RS-2 Wuhan dataset, with an accuracy improvement of only 0.29% for 20,127 pseudo-labeled samples compared with 5358 samples. The classification results are shown in Table 13.
Through the experiments on the three different datasets, it was shown that an increase in the number of superpixels leads to an increase in the number of high-confidence pseudo-labels, thereby improving classification performance to a certain extent. However, when the number of pseudo-labels reaches a certain level, the improvement in accuracy gradually stabilizes.

4.4. Hardware Environment and Runtime

The detailed information on the hardware platform used for this paper is shown in Table 14.
Table 15 shows the training time and inference time of five different methods on the RS-2 Flevoland dataset. Different datasets have different image sizes, so the time consumption is a little different. Since the method proposed is based on the results of preliminary unsupervised classification, it will inevitably consume more time. The additionally consumed time is the training and inference time required by the semi-supervised method SimMatch. The running speed is limited by the performance of the hardware platform. Improving the hardware platform with better performance can effectively reduce the time of training and inference. Therefore, the proposed method is acceptable in terms of time.

5. Conclusions

This study proposes a framework for the unsupervised classification of PolSAR images based on superpixel segmentation and a semi-supervised algorithm. The SLIC superpixel segmentation method and unsupervised classification maps from RDDMI are used to extract high-confidence pseudo-labeled samples. The semi-supervised method SimMatch is used to train with the high-confidence pseudo-labeled samples and unlabeled data to generate the final classification results. The proposed method was tested on three PolSAR image datasets and achieved great performance. The four evaluation indicators showed that the proposed method outperformed the state-of-the-art unsupervised classification method, RDDMI. SP-SIM has significant effectiveness for unsupervised PolSAR image classification.

Author Contributions

Conceptualization, L.W.; methodology, L.W.; software, L.P.; validation, L.P.; formal analysis, R.G.; investigation, R.G.; resources, L.W.; data curation, L.P. and S.Z.; writing—original draft preparation, L.P.; writing—review and editing, L.W., L.P., H.H. and S.Z.; supervision, H.H.; funding acquisition, L.W. and H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 42201432 and Grant 62171329, and in part by the Wuhan Knowledge Innovation Special Project under Grant 2022010801010351.

Data Availability Statement

Due to privacy, the data used in this study is unavailable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, J.S.; Grunes, M.R.; Kwok, R. Classification of multi-look polarimetric SAR imagery based on complex Wishart distribution. Int. J. Remote Sens. 1994, 15, 2299–2311. [Google Scholar] [CrossRef]
  2. Goodman, N.R. Statistical analysis based on a certain multivariate complex Gaussian distribution (an introduction). Ann. Math. Stat. 1963, 34, 152–177. [Google Scholar] [CrossRef]
  3. Pottier, E. Unsupervised classification scheme of polsar images based on the complex wishart distribution and the H/A/α polarimetric decomposition theorem. In Proceedings of the 3th European Conference on Synthetic Aperture Radar, Munich, Germany, 23–25 May 2000. [Google Scholar]
  4. Cao, F.; Hong, W.; Wu, Y.; Pottier, E. An Unsupervised Segmentation With an Adaptive Number of Clusters Using the SPAN/H/α/A Space And the Complex Wishart Clustering for Fully Polarimetric SAR Data Analysis. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3454–3467. [Google Scholar] [CrossRef]
  5. Lee, J.; Grunes, M.R.; Pottier, E.; Ferro-Famil, L. Unsupervised terrain classification preserving polarimetric scattering characteristics. IEEE Trans. Geosci. Remote Sens. 2004, 42, 722–731. [Google Scholar]
  6. Freeman, A.; Durden, S.L. A three-component scattering model for polarimetric sar data. IEEE Trans. Geosci. Remote Sens. 1998, 36, 963–973. [Google Scholar] [CrossRef]
  7. Anfinsen, S.N.; Jenssen, R.; Eltoft, T. Spectral Clustering of Polarimetric SAR Data With Wishart-Derived Distance Measures. In Proceedings of the 3rd International Workshop on Science and Applications of SAR Polarimetry and Polarimetric Interferometry, Frascati, Italy, 22–26 January 2007; Volume 7, pp. 1–9. [Google Scholar]
  8. Song, H.; Yang, W.; Bai, Y.; Xu, X. Unsupervised classification of polarimetric SAR imagery using large-scale spectral clustering with spatial constraints. Int. J. Remote Sens. 2015, 36, 2816–2830. [Google Scholar] [CrossRef]
  9. Yang, Y.; Yang, Y.; Wang, Y.; Xue, X. A novel spectral clustering method with superpixels for image segmentation. Optik 2015, 127, 161–167. [Google Scholar] [CrossRef]
  10. Yang, X.; Yang, W.; Song, H.; Huang, P. Superpixel-Based Unsupervised Classification of PolSAR Imagery Using Wishart Mixture Models and Spectral Clustering. In Proceedings of the EUSAR 2016: 11th European Conference on Synthetic Aperture Radar, Hamburg, Germany, 6–9 June 2016. [Google Scholar]
  11. Caron, M.; Bojanowski, P.; Joulin, A. Deep Clustering for Unsupervised Learning of Visual Features. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 132–149. [Google Scholar]
  12. Haeusser, P.; Plapp, J.; Golkov, V.; Aljalbout, E.; Cremers, D. Associative Deep Clustering—Training a Classification Network with no Labels. In Proceedings of the Pattern Recognition: 40th German Conference, GCPR 2018, Stuttgart, Germany, 9–12 October 2018; pp. 18–23. [Google Scholar]
  13. Ji, P.; Zhang, T.; Li, H.; Salzmann, M.; Reid, I. Deep Subspace Clustering Networks. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 23–32. [Google Scholar]
  14. Zhou, P.; Hou, Y.; Feng, J. Deep adversarial subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1596–1604. [Google Scholar]
  15. Ji, X.; Vedaldi, A.; Henriques, J.F. Invariant Information Clustering for Unsupervised Image Classification and Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9864–9873. [Google Scholar]
  16. Hinton, G.; Osindero, S.; Teh, Y. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  17. Lv, Q.; Dou, Y.; Niu, X.; Xu, J.; Xia, F. Urban land use and land cover classification using remotely sensed SAR data through deep belief networks. J. Sens. 2015, 2015, 538063. [Google Scholar] [CrossRef]
  18. Lopes, N.; Ribeiro, B. Towards adaptive learning with improved convergence of deep belief networks on graphics processing units. Pattern Recognit. 2014, 47, 114–127. [Google Scholar] [CrossRef]
  19. Hou, B.; Kou, H.; Jiao, L. Classification of polarimetric SAR images using multilayer autoencoders and superpixels. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3072–3081. [Google Scholar] [CrossRef]
  20. Zhang, L.; Ma, W.; Zhang, D. Stacked Sparse Autoencoder in PolSAR Data Classification Using Local Spatial Information. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1359–1363. [Google Scholar] [CrossRef]
  21. Hu, Y.; Fan, J.; Wang, J. Classification of PolSAR images based on adaptive nonlocal stacked sparse autoencoder. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1050–1054. [Google Scholar] [CrossRef]
  22. Bi, H.; Xu, F.; Wei, Z.; Han, Y.; Cui, Y.; Xue, Y.; Xu, Z. Unsupervised PolSAR Image Factorization with Deep Convolutional Networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019. [Google Scholar]
  23. Wang, L.; Xu, X.; Gui, R.; Yang, R.; Pu, F. Learning Rotation Domain Deep Mutual Information Using Convolutional LSTM for Unsupervised PolSAR Image Classification. Remote Sens. 2020, 12, 4075. [Google Scholar] [CrossRef]
  24. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.; Wong, W.; Woo, W. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015; pp. 7–12. [Google Scholar]
  25. Zuo, Y.X.; Li, G.Z.; Ren, W.J.; Hu, Y.X. A Deep Similarity Clustering Network With Compound Regularization for Unsupervised PolSAR Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11451–11466. [Google Scholar] [CrossRef]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  28. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 9992–10002. [Google Scholar]
  29. Lee, D.H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Proceedings of the Workshop on Challenges in Representation Learning, ICML, Atlanta, GA, USA, 16–21 June 2013; Volume 3, p. 896. [Google Scholar]
  30. Sajjadi, M.; Javanmardi, M.; Tasdizen, T. Regularization with stochastic transformations and perturbations for deep semi-supervised learning. arXiv 2016, arXiv:1606.04586. [Google Scholar]
  31. Berthelot, D.; Carlini, N.; Goodfellow, I.; Papernot, N.; Oliver, A.; Raffel, C. Mixmatch: A holistic approach to semi-supervised learning. arXiv 2019, arXiv:1905.02249. [Google Scholar]
  32. Zhang, H.Y.; Cisse, M.; Dauphin, Y.N. Mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  33. Berthelot, D.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Sohn, K.; Zhang, H.; Raffel, C. Remixmatch: Semi-supervised learning with distribution alignment and augmentation anchoring. arXiv 2019, arXiv:1911.09785. [Google Scholar]
  34. Sohn, K.; Berthelot, D.; Carlini, N.; Zhang, Z.; Zhang, H.; Raffel, C.A.; Cubuk, E.D.; Kurakin, A.; Li, C.L. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Adv. Neural Inf. Process. Syst. 2020, 33, 596–608. [Google Scholar]
  35. Zheng, M.; You, S.; Huang, L.; Wang, F.; Qian, C.; Xu, C. SimMatch: Semi-supervised Learning with Similarity Matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14471–14481. [Google Scholar]
  36. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed]
  37. Zagoruyko, S.; Komodakis, N. Wide Residual Networks. arXiv 2016, arXiv:1605.07146. [Google Scholar]
  38. Wang, L.; Xu, X.; Dong, H.; Gui, R.; Yang, R.; Pu, F. Exploring Convolutional Lstm for Polsar Image Classification. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8452–8455. [Google Scholar]
  39. Wang, L.; Peng, L.M.; Hong, H.; Zhao, S.; Lv, Q.; Gui, R. Semi-supervised PolSAR Image Change Detection using Similarity Matching. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, 48, 655–661. [Google Scholar] [CrossRef]
  40. Bachman, P.; Alsharif, O.; Precup, D. Learning with pseudo-ensembles. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; pp. 3365–3373. [Google Scholar]
  41. Laine, S.; Aila, T. Fixmatch: Temporal ensembling for semi-supervised learning. arXiv 2017, arXiv:1610.02242. [Google Scholar]
  42. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9726–9735. [Google Scholar]
  43. French, G.; Mackiewicz, M.; Fisher, M. Self-ensembling for visual domain adaptation. arXiv 2017, arXiv:1706.05208. [Google Scholar]
  44. Li, J.; Xiong, C.; Hoi, S.C. Comatch: Semi-supervised learning with contrastive graph regularization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9475–9484. [Google Scholar]
Figure 1. The five parts of the framework. The Wide ResNet model adopts the classic wide residual networks (WRNs) [37]. The useful features from the input data are extracted by the backbone to obtain an embedding vector. L s , L u , and L i n represent the supervised loss, unsupervised loss, and similarity distribution, respectively.
Figure 1. The five parts of the framework. The Wide ResNet model adopts the classic wide residual networks (WRNs) [37]. The useful features from the input data are extracted by the backbone to obtain an embedding vector. L s , L u , and L i n represent the supervised loss, unsupervised loss, and similarity distribution, respectively.
Remotesensing 16 04119 g001
Figure 2. The pseudo-label generation structure of SimMatch. SimMatch generates semantic and instance pseudo-labels using weakly augmented views and calculates semantic and instance similarities through class centers. These two similarities are then propagated to each other using expansion and aggregation to obtain better pseudo-labels.
Figure 2. The pseudo-label generation structure of SimMatch. SimMatch generates semantic and instance pseudo-labels using weakly augmented views and calculates semantic and instance similarities through class centers. These two similarities are then propagated to each other using expansion and aggregation to obtain better pseudo-labels.
Remotesensing 16 04119 g002
Figure 3. The propagation of pseudo-labels’ information. As the example in the red box shows. If the similarity between semantics and instances are different, the histogram will be flatter, and if the semantics similarities and the instances similarities are similar, the resulting histogram will be sharper.
Figure 3. The propagation of pseudo-labels’ information. As the example in the red box shows. If the similarity between semantics and instances are different, the histogram will be flatter, and if the semantics similarities and the instances similarities are similar, the resulting histogram will be sharper.
Remotesensing 16 04119 g003
Figure 4. RS-2 Flevoland dataset. (a) Pauli pseudo-color image. (b) Ground-truth map.
Figure 4. RS-2 Flevoland dataset. (a) Pauli pseudo-color image. (b) Ground-truth map.
Remotesensing 16 04119 g004
Figure 5. RS-2 Wuhan dataset. (a) Pauli pseudo-color image. (b) Ground-truth map. (c) An optical image of ROI_1. (d) An optical image of ROI_2.
Figure 5. RS-2 Wuhan dataset. (a) Pauli pseudo-color image. (b) Ground-truth map. (c) An optical image of ROI_1. (d) An optical image of ROI_2.
Remotesensing 16 04119 g005
Figure 6. AIRSAR Flevoland dataset. (a) Pauli pseudo-color image. (b) Ground-truth map.
Figure 6. AIRSAR Flevoland dataset. (a) Pauli pseudo-color image. (b) Ground-truth map.
Remotesensing 16 04119 g006
Figure 7. Classification results on the RS-2 Flevoland dataset. The black boxes show that SP-SIM has more fine classification results than RDDMI. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Figure 7. Classification results on the RS-2 Flevoland dataset. The black boxes show that SP-SIM has more fine classification results than RDDMI. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Remotesensing 16 04119 g007
Figure 8. Classification results on the RS-2 Wuhan dataset. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Figure 8. Classification results on the RS-2 Wuhan dataset. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Remotesensing 16 04119 g008
Figure 9. Similar backscattering properties on the AIRSAR Flevoland dataset. (a) Four similar backscattering properties. (b) Water. (c) Bare soil. (d) Lucerne. (e) Rape seed.
Figure 9. Similar backscattering properties on the AIRSAR Flevoland dataset. (a) Four similar backscattering properties. (b) Water. (c) Bare soil. (d) Lucerne. (e) Rape seed.
Remotesensing 16 04119 g009
Figure 10. Classification results on the AIRSAR Flevoland dataset. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Figure 10. Classification results on the AIRSAR Flevoland dataset. (a) Ground-truth map. (b) Wishart. (c) RDDMI. (d) SP-SIM.
Remotesensing 16 04119 g010
Table 1. Information of the three datasets.
Table 1. Information of the three datasets.
ParameterRS-2 FlevolandAIRSAR FlevolandRS-2 Wuhan
SensorRADARSAT-2NASA/JPL AIRSARRADARSAT-2
BandCLC
Imaging AreaFlevolandFlevolandWuhan
Imaging ModeQuad-polQuad-pol
Imaging time200816 August 1989December 2011
Spatial resolution [Range Azimuth] (m) 12 × 8 6 × 12 12 × 8
Image size [Range Azimuth] (pixel) 1400 × 1200 750 × 1024 5500 × 2400
Table 2. The results of RS-2 Flevoland. The best performance is bolded.
Table 2. The results of RS-2 Flevoland. The best performance is bolded.
Land CoverWishartRDDMISP-SIMRF
Water90.6392.4993.1095.85
Forest77.7782.1486.3187.77
Farmland65.6586.4289.0485.60
Building55.3194.1691.8882.39
OA71.3587.8689.5688.10
Kappa0.610.830.860.83
precision0.760.870.890.88
recall0.720.890.900.88
F10.730.880.900.88
Table 3. The results of the RS-2 Wuhan dataset. The best performance is bolded.
Table 3. The results of the RS-2 Wuhan dataset. The best performance is bolded.
Land CoverWishartRDDMISP-SIMRF
Water88.1092.2692.5494.01
Forest67.8988.4989.2088.34
Building74.0275.8377.7181.44
OA74.5585.3186.3087.45
Kappa0.600.770.790.80
precision0.780.860.870.88
recall0.770.860.860.88
F10.810.860.870.88
Table 4. The results of the AIRSAR Flevoland dataset. The best performance is bolded.
Table 4. The results of the AIRSAR Flevoland dataset. The best performance is bolded.
Land CoverWishartRDDMISP-SIMRF
Stem bean65.1594.8894.6896.69
Forest31.1397.1997.4690.18
Potatoes87.7065.4868.8489.62
Lucerne67.4975.7876.8589.17
Wheat57.7182.1883.3077.94
Bare soil82.3882.8884.4995.57
Beet71.5953.3250.7389.00
Rape seed79.9035.0735.8890.38
Peas57.9578.8078.8073.50
Grass50.3691.5892.4293.56
Water00099.97
OA59.3472.8473.6486.81
Kappa0.50.690.700.85
precision0.560.670.670.87
recall0.560.690.690.90
F10.540.660.660.88
Table 5. The results of the RS-2 Flevoland dataset. The best performance is bolded.
Table 5. The results of the RS-2 Flevoland dataset. The best performance is bolded.
Land CoverWishartSLIC-WishartDCCMSLIC-DCCMSP-SIM
Water90.6392.5593.9894.5293.10
Forest77.7780.7876.8183.8786.31
Farmland65.6567.0083.7486.7889.04
Building55.3159.6394.8392.1191.88
OA71.3573.8985.8788.4089.56
Kappa0.610.640.810.840.86
precision0.760.780.860.880.89
recall0.720.750.870.890.90
F10.730.760.860.880.90
Table 6. The results of the RS-2 Wuhan dataset. The best performance is bolded.
Table 6. The results of the RS-2 Wuhan dataset. The best performance is bolded.
Land CoverWishartSLIC-WishartDCCMSLIC-DCCMSP-SIM
Water88.1089.3797.1997.1692.54
Forest67.8976.4858.1579.5289.20
Building74.0276.2675.0163.2377.71
OA74.5579.3872.6376.3386.30
Kappa0.600.680.590.640.79
precision0.780.810.740.780.87
recall0.770.810.770.800.86
F10.770.810.730.770.87
Table 7. The results of the AIRSAR Flevoland dataset. The best performance is bolded.
Table 7. The results of the AIRSAR Flevoland dataset. The best performance is bolded.
Land CoverWishartSLIC-WishartDCCMSLIC-DCCMSP-SIM
Stem bean65.1576.6969.5169.6294.68
Forest31.1325.0398.3098.3797.46
Potatoes87.7092.0243.1242.6768.84
Lucerne67.4973.6178.4578.7376.85
Wheat57.7160.5534.6734.7383.30
Bare soil82.3891.8292.4492.8284.49
Beet71.5979.812.352.1950.73
Rape seed79.9082.4356.7856.8135.88
Peas57.9562.9674.3574.3578.80
Grass50.3618.6217.5572.9992.42
Water00000
OA59.3463.2560.6256.4373.64
Kappa0.540.580.550.520.70
precision0.570.600.580.580.67
recall0.560.600.570.570.69
F10.540.570.530.530.66
Table 8. Performance evaluation of VGG-16 and SP-SIM on the RS-2 Flevoland dataset. The best performance is bolded.
Table 8. Performance evaluation of VGG-16 and SP-SIM on the RS-2 Flevoland dataset. The best performance is bolded.
Land CoverVGG-16, 20,287 SamplesSP-SIM, 20,287 Sample
Water92.0693.10
Forest83.5686.31
Farmland86.5489.04
Building91.7591.88
OA87.7389.56
Kappa0.830.86
precision0.880.89
recall0.880.90
F10.880.90
Table 9. Performance evaluation of VGG-16 and SP-SIM on the RS-2 Wuhan dataset. The best performance is bolded.
Table 9. Performance evaluation of VGG-16 and SP-SIM on the RS-2 Wuhan dataset. The best performance is bolded.
Land CoverVGG-16, 20,127 SamplesSP-SIM, 20,127 Sample
Water91.3892.54
Farmland89.0789.20
Building77.3877.71
OA85.8686.30
Kappa0.780.79
precision0.870.87
recall0.860.86
F10.860.87
Table 10. Performance evaluation of VGG-16 and SP-SIM on the AIRSAR Flevoland dataset. The best performance is bolded.
Table 10. Performance evaluation of VGG-16 and SP-SIM on the AIRSAR Flevoland dataset. The best performance is bolded.
Land CoverVGG-16, 5244 SamplesSP-SIM, 5244 Samples
Stem bean94.7594.68
Forest96.8297.46
Potatoes70.9668.84
Lucerne76.4976.85
Wheat81.7483.30
Bare soil82.7584.49
Beet49.5050.73
Rape seed35.5635.88
Peas78.4978.80
Grass89.2592.42
Water00
OA73.0373.64
Kappa0.690.70
precision0.670.67
recall0.690.69
F10.660.66
Table 11. The results of 3730 and 20,287 pseudo-labeled samples on the RS-2 Flevoland dataset. The best performance is bolded.
Table 11. The results of 3730 and 20,287 pseudo-labeled samples on the RS-2 Flevoland dataset. The best performance is bolded.
Land Cover3730 Samples20,287 Samples
Water92.5893.10
Forest83.2986.31
Farmland88.6889.04
Building92.6991.89
OA88.6589.56
Kappa0.850.86
precise0.880.89
recall0.890.90
F10.890.90
Table 12. The results of 935 and 5244 pseudo-labeled samples on the AIRSAR Flevoland dataset. The best performance is bolded.
Table 12. The results of 935 and 5244 pseudo-labeled samples on the AIRSAR Flevoland dataset. The best performance is bolded.
Land Cover935 Samples5244 Samples
Stem bean94.7594.68
Forest98.1897.46
Potatoes65.0468.84
Lucerne77.5776.85
Wheat81.5183.30
Bare soil85.9584.49
Beet55.5750.73
Rape seed34.2835.88
Peas78.8678.80
Grass92.6692.42
Water00
OA73.1373.64
Kappa0.700.70
precision0.670.67
recall0.690.69
F10.660.66
Table 13. The results of 5358 and 20,127 pseudo-labeled samples on the RS-2 Wuhan dataset. The best performance is bolded.
Table 13. The results of 5358 and 20,127 pseudo-labeled samples on the RS-2 Wuhan dataset. The best performance is bolded.
Land Cover5358 Samples20,127 Samples
Water82.6192.54
Farmland89.3189.20
Building76.6177.71
OA86.0186.30
Kappa0.780.79
precision0.870.87
recall0.860.86
F10.860.87
Table 14. Hardware environment.
Table 14. Hardware environment.
NameConfiguration Information
SystemWindows11
Development LanguagePython3.8
FrameworkPytorch 1.13.0 + cuda 11.6
GPUNIVIDIA RTX A4000
CPUIntel Core i7-11700
Memory32 G
Table 15. The training and inference time of RS-2 Flevoland. The size of RS-2 Flevoland is 1200 × 1400 pixels, and 20,287 labeled samples are used.
Table 15. The training and inference time of RS-2 Flevoland. The size of RS-2 Flevoland is 1200 × 1400 pixels, and 20,287 labeled samples are used.
MethodTrain TimeInference TimeTotal Time
Wishart10 s10 s20 s
DCCM2 days5 min2 days + 5 min
RDDMI2 days5 min2 days + 5 min
SimMatch1.5 days15 min1.5 days + 15 min
SP-SIM3.5 days20 min3.5 days + 20 min
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Peng, L.; Gui, R.; Hong, H.; Zhu, S. Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network. Remote Sens. 2024, 16, 4119. https://doi.org/10.3390/rs16214119

AMA Style

Wang L, Peng L, Gui R, Hong H, Zhu S. Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network. Remote Sensing. 2024; 16(21):4119. https://doi.org/10.3390/rs16214119

Chicago/Turabian Style

Wang, Lei, Lingmu Peng, Rong Gui, Hanyu Hong, and Shenghui Zhu. 2024. "Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network" Remote Sensing 16, no. 21: 4119. https://doi.org/10.3390/rs16214119

APA Style

Wang, L., Peng, L., Gui, R., Hong, H., & Zhu, S. (2024). Unsupervised PolSAR Image Classification Based on Superpixel Pseudo-Labels and a Similarity-Matching Network. Remote Sensing, 16(21), 4119. https://doi.org/10.3390/rs16214119

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop