Next Article in Journal
Dual Field-of-View Off-Axis Spatially Multiplexed Digital Holography Using Fresnel’s Bi-Mirror
Previous Article in Journal
Optrode-Assisted Multiparametric Near-Infrared Spectroscopy for the Analysis of Liquids
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Semi-Supervised Semantic Segmentation of Remote Sensing Images via Feature Perturbation-Based Consistency Regularization Methods

1
Key Laboratory of Target Cognition and Application Technology, The Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
2
The Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(3), 730; https://doi.org/10.3390/s24030730
Submission received: 22 November 2023 / Revised: 28 December 2023 / Accepted: 21 January 2024 / Published: 23 January 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
In the field of remote sensing technology, the semantic segmentation of remote sensing images carries substantial importance. The creation of high-quality models for this task calls for an extensive collection of image data. However, the manual annotation of these images can be both time-consuming and labor-intensive. This has catalyzed the advent of semi-supervised semantic segmentation methodologies. Yet, the complexities inherent within the foreground categories of these remote sensing images present challenges in preserving prediction consistency. Moreover, remote sensing images possess more complex features, and different categories are confused within the feature space, making optimization based on the feature space challenging. To enhance model consistency and to optimize feature-based class categorization, this paper introduces a novel semi-supervised semantic segmentation framework based on Mean Teacher (MT). Unlike the conventional Mean Teacher that only introduces perturbations at the image level, we incorporate perturbations at the feature level. Simultaneously, to maintain consistency after feature perturbation, we employ contrastive learning for feature-level learning. In response to the complex feature space of remote sensing images, we utilize entropy threshold to assist contrastive learning, selecting feature key-values more precisely, thereby enhancing the accuracy of segmentation. Extensive experimental results on the ISPRS Potsdam dataset and the challenging iSAID dataset substantiate the superior performance of our proposed methodology.

1. Introduction

In recent years, remote sensing image processing has become an important research area, among which the semantic segmentation of remote sensing images is a key task [1,2,3,4]. The goal of semantic segmentation is to assign each pixel in the image to an interpretable category. At present, the semantic segmentation of remote sensing images has played a significant role in fields such as military reconnaissance [5], urban planning [6], and environmental monitoring [7], greatly facilitating the automation and decision optimization in these areas. However, due to the characteristics of remote sensing images, conducting pixel-level manual annotation is extremely laborious and time-consuming. Therefore, semi-supervised semantic segmentation has become a major research direction to reduce the cost of manual annotation.
The goal of semi-supervised semantic segmentation is to achieve high-precision segmentation results with a smaller proportion of labelled images. The main types of semi-supervised semantic segmentation methods currently include methods based on consistency regularization [8,9,10,11,12], methods based on pseudo-labels [13], and the recently popular methods based on contrastive learning [14,15,16]. The core of methods based on consistency regularization is to add perturbations to the input, and to and train the model with the aim of maintaining output consistency. Research on consistency regularization models in natural scenes is relatively comprehensive. However, remote sensing scenarios differ significantly from natural scenes. Particularly, the foreground types in remote sensing images are more complex and densely distributed, and the differences between the categories are minimal, leading to a relatively chaotic feature distribution. Currently, the consistency regularization methods applied in the field of remote sensing are mostly extensions of methods from natural scenes. For instance, Zhang [7] applies various random transformations and perturbations to images and predicted labels; ICNet [17] switches between student and teacher networks based on training rounds, adding network perturbations. Although these methods have improved the results of the remote sensing dataset, they focus more on the image level and do not consider the feature level of the remote sensing images. We believe that optimizing the feature level of remote sensing images is equally important.
For remote sensing images, by analyzing their features, we find that the feature distribution is more scattered and chaotic compared to natural scenes. However, existing consistency regularization methods for remote sensing images do not pay much attention to the feature level. To address this, we aim to add feature-level perturbations to force the model to optimize the feature space. In the Mean Teacher (MT) [8] model, the teacher network assists the student network, requiring consistency in the prediction maps generated by different image enhancements. Considering the feature level of remote sensing images, adding different feature perturbations in the student and teacher networks to implement a consistency regularization strategy is an effective solution for feature optimization. At the feature level, after perturbing image features, a loss function must be chosen to evaluate the consistency of the prediction results of the two networks, thus aiding in with updates. The commonly used cross-entropy loss clearly does not meet our needs. Contrastive learning, essentially targeting the feature space, is obviously a good method. There are already corresponding remote sensing semantic segmentation applications using this method [6], achieving decent results. However, traditional contrastive learning still faces the problem of unreasonable negative sample selection. Therefore, we propose a new strategy, setting an entropy threshold to filter key values, to address the problem of more dispersed feature distribution caused by the unique characteristics of remote sensing images.
In this paper, considering the existing problems of current methods, we propose a novel semi-supervised semantic segmentation framework that combines consistency regularization and contrastive learning. For consistency regularization, based on MT, in addition to image-level disturbances, we introduce feature-level disturbances, allowing the model to pay more attention to the feature level of the image during the learning process, compared to previous methods. To achieve this goal, we choose to use contrastive learning to calculate the contrastive loss, assisting in the optimization at the feature level, and completing the final network parameter updates. For the key selection used in contrastive learning, we use entropy thresholds to select positive keys and negative keys, making the results of contrastive learning more precise.
The main contributions of this paper can be summarized as follows:
  • We have introduced a new semi-supervised segmentation framework for remote sensing images. The framework integrates consistency regularization and contrastive learning, enhancing the disturbances at the data and feature levels, and improves feature classification performance through contrastive learning. In addition, this method achieves state-of-the-art performance in popular segmentation benchmarks.
  • We proposed a new consistency regularization method based on MT [8]. By enhancing perturbations at the feature level, the difficulty of maintaining the consistency of image features increases, thus adding to the training difficulty and improving the generalization ability of complex images. Feature perturbations plays a key role in this process and help the model to learn from more challenging features.
  • We utilize contrastive learning at the feature level to achieve a better divide and category selection for the features. A threshold for entropy is established to aid in feature selection, sifting out more accurate negative samples.

2. Related Works

2.1. Supervised Semantic Segmentation

Semantic segmentation serves as a cornerstone operation in the realm of computer vision. FCN [18] is groundbreaking in the field of semantic segmentation. By replacing fully connected layers with convolutional layers, FCNs extend convolutional neural networks from image classification tasks to pixel-level prediction tasks. FCNs also introduce an upsampling operation to obtain fine segmentation results. Despite some limitations, FCN lays the foundation for subsequent research, as mentioned earlier. U-Net, primarily used for medical image segmentation, adopts a symmetric encoder-decoder structure and achieves feature fusion through skip connections. This design allows U-Net [19] to fully utilize multi-scale features, hence yielding excellent performance in segmentation tasks. DeepLabV3+ [20] is the latest in the DeepLab series that is focused on studying dilated convolution and Atrous Spatial Pyramid Pooling (ASPP). Building on the foundation of DeepLabV3, DeepLabV3+ introduces a decoder module to achieve finer segmentation results. PSPNet [21], through the introduction of a pyramid pooling module, effectively captures the context information of different scales. This capability enables PSPNet to achieve accurate semantic segmentation in multi-scale scenes. Fast-FCN [22] is an efficient semantic segmentation method. By introducing a global average pooling layer, Fast-FCN achieves the rapid aggregation of features. Compared with traditional FCN, Fast-FCN significantly improves running speed while maintaining accuracy. For different scenarios, some models have improved their specific structures to adapt to specific tasks. For example, PSNet [23], addressing the issue of low detection rate and a high false alarm rate in forest fire smoke detection, introduces a detail-difference-aware module to distinguish between smoke and smoke-like objects, and an attention-based feature separation modules to suppress interference features.
In addition, methods based on the Transformer architecture have also achieved very good results. For instance, Segmenter [24], which is based on the Transformer architecture, allows the model to capture global context information in both the first layer and throughout the entire network. TransUNet [25] also employs a hybrid visual Transformer as an encoder for stronger feature extraction, and it has achieved state-of-the-art results in medical image segmentation. SegFormer [26] is also a simple, efficient, and powerful semantic segmentation framework that combines Transformers with a lightweight multi-layer perceptron decoder.

2.2. Semi-Supervised Semantic Segmentation

Common strategies in semi-supervised semantic segmentation involve leveraging the principle of consistency regularization [9,27,28,29], which is also used in semi-supervised classification tasks. In this regard, CutMix [10], MixUp [30] extend unlabelled data and enforce consistency under these deformations. Another strand of consistency regularization philosophy involves transferring knowledge from labelled to unlabelled data. The Mean Teacher [8] is utilized to enforce consistency for the predictions of unlabelled images across different training epochs. Similarly, the work in [31] introduced a technique that maintains the exponential moving average of previous models using self-ensembling to achieve stable outputs for unlabelled data. To increase network perturbations, PSMT [32] employs a dual-teacher network to assist in the training and optimization of the student network.
Pseudo-label-based methods constitute another mature research direction in semi-supervised semantic segmentation. These methods rely more heavily on prior predictions made on unlabelled data and extend the dataset using these predictions. A typical representative example is self-training methods, where ST++ [33] applies data augmentation techniques to unlabelled images during the self-training process.
Another method based on adversarial learning [34,35] has also shown promising results, such as in [36], where the discriminator’s task is to distinguish unlabelled image segmentation outputs originating from the labelled or unlabelled pools, forcing the generator to create indistinguishable segmentation between the two.
Contrastive learning, which emphasizes high-level features, enables the network to distinguish classes well without real labels. There have been some works on semantic segmentation using contrastive learning [37,38,39,40]. Reco [15] was the first to apply pixel-level contrastive learning to the field of semantic segmentation. Moreover, Chen et al. [41] only samples positive examples when using contrastive learning. Lai et al. [42] adopts a self-supervised learning paradigm, selecting reliable pixel points from differently augmented images as positive and negative samples for contrastive learning.

2.3. Semi-Supervised Semantic Segmentation of Aerial Imagery

Remote sensing images have high resolution, with the number of pixels in a single image far exceeding those in other fields. At the same time, the categories in remote sensing images are more complex, making manual pixel-level annotation work more time-consuming and laborious. To address this labor-intensive annotation issue, in recent years, some researchers have started to conduct research on semi-supervised semantic segmentation. Zhang et al. [7] performed Transformation Consistency Regularization on the prediction labels of the teacher network and compared the results with the student network, extending randomness to the label level. ICNet [17], during the training process, alternates between transforming the student network and the teacher network. This approach allows the two networks to supervise each other, thereby increasing perturbations at the network level. PICS [5] adopts a selective self-training strategy. By using labelled images, it selects generated samples that are closer to the true values, thereby reducing the accumulation of potential errors. UniMatch [43] improved the structure of FixMatch and performed Dropout operations on the features after the decoder, introducing additional disturbances at the feature level to achieve better segmentation results. Kerdegari et al. [44] used a GAN network to enlarge the annotated dataset. Yang et al. [6] improved the final segmentation accuracy using contrastive learning. Zhang et al. [45] combined the CPS structure and integrated prediction results to adapt to SAR image semantic segmentation. Fang [46] introduced a clustering algorithm into the co-learning algorithm, generating high-quality pseudo-labels by integrating features. Hong [47] utilized multiple types of data, SAR images, and multispectral images, extracted more features to assist in pixel classification; and redesigned a multimodal classification framework for this purpose.
Compared to the direct transfer of methods from natural scenes, these works improve the segmentation effect of remote sensing images significantly. However, the attention received by the semi-supervised semantic segmentation of high-resolution remote sensing images is far from matching its practical importance. Inspired by the consistency regularization method and the contrastive learning method, and considering the characteristics of remote sensing images, we have rethought how to truly specialize semi-supervised semantic segmentation for remote sensing images.

3. Materials and Methods

3.1. Methods

In this section, we will elaborate on our research methodology in detail. Our approach is designed and optimized based on the Mean Teacher architecture. Mean Teacher, as an effective semi-supervised learning method, primarily aims to reduce the model’s sensitivity to minor variations in input data through consistency regularization, thereby enhancing the model’s generalization capability. Building on this, we have made a series of improvements to the network structure to cater to our specific tasks and data. To perturb the features, we have added different feature perturbation modules after the encoder in both the student and teacher networks. Simultaneously, for the progression of contrastive learning, we have added a feature representation head parallel to the decoder to adapt to our specific task and data. The specific network structure and model flow are shown in Figure 1. First, we set up a semi-supervised dataset.
Labelled dataset: D l = ( x i l , y i l ) i = 1 | D l |
Unlabelled dataset: D u = ( x i l ) i = 1 | D u |
According to the principle of semi-supervised learning, labelled images represent only a small fraction of all training images, | D l | | D u | . The proposed framework consists of a student network and a teacher networks. The student and teacher networks have the same architecture but different parameters, where the parameters θ t of the teacher network is the exponential moving average (EMA) of the student network parameters θ s . The update formula for θ t is as follows:
θ t e = γ θ t e 1 + ( 1 γ ) θ s e
γ is the smoothing factor, γ ( 0 , 1 ) . e represents the training epoch.
Our method optimizes the prediction classification at both the image level and feature level by targeting the features of remote sensing images. Through learning at different levels, effective image segmentation is achieved. The overall framework is divided into two parts, each with the following structure:
Feature Disturbed Mean Teacher Model (FDMT): FDMT is a new paradigm for semi-supervised semantic segmentation. Building upon the traditional Mean Teacher module, it additionally incorporates a feature disturbance component. This enables the multi-level optimization of pixel classification in images, leading to improved classification results.
Contrastive Learning with Entropy Threshold Assisted Feature Sampling: This paper utilizes contrastive learning to aid with the optimization of the feature space. In conducting contrastive learning, we employ entropy as an auxiliary means for sampling queries, positive keys, and negative keys. By setting an entropy threshold, we aim to filter out more accurate key values, thereby facilitating more efficient contrastive learning.

3.1.1. Feature Disturbed Mean Teacher Model (FDMT)

In the method proposed in this study, the model comprises a teacher network and a student network. Unlike the network structure of traditional Mean Teacher, each network includes an encoder head h, a decoder head f, and a representation head r. After the encoder, a feature perturbation module is added to introduce feature perturbations. The reason for adopting an additional representation head here is as follows: The features for contrastive learning need to be more general, and the features directly obtained from the encoder and decoder may not yield the best results. Therefore, an additional representation head is designed to extract and contrast features, enhancing their discriminability. For this, we use the representation head r designed in [15]. After each round of training, the parameters of the student network are updated based on the loss.
During each training iteration, we carry out a random sampling to obtain an identical number of labelled images N l and unlabelled images N u , with | N l | = | N u | . For each labelled image, we input it into the student network to make predictions, and then we compare the predicted labels with the ground truth labels to calculate the supervised loss L s u p .
L s u p = 1 | N l | ( x i l , y i l ) N l l c e ( y s i l , y i l )
y s i l = O ( S ( f h ( x i l ; θ s ) ) )
l c e ( ) represents the cross-entropy loss, and y i l denotes the manually annotated labels. S ( ) represents the softmax function, and O ( ) represents the one-hot encoding form. S ( f h ) represents the segmentation probability map generated after the image passes through the encoder h and decoder f successively. y s i l is the final one-hot encoded form. θ s denotes the parameters of the student network.
For unlabelled images, our process differs significantly from the Mean Teacher model. First, let’s discuss the student network part. The same unlabelled image follows two different processing streams: one with strong data augmentation, where we use Cutmix, and one without data augmentation. Here, the use of strong image augmentation is designed for perturbations at the image level, while the process without image augmentation aims for a more rational and controllable addition of feature perturbations. By doing so, we simultaneously introduce disturbances at both the image and feature levels, thereby optimizing the model on multiple dimensions. The images, having gone through different augmentation processes, enter the student network, generating two prediction labels, namely y s 1 u and y s 2 u . For images that have not undergone image augmentation, after entering the encoder, a VAT perturbation δ is added to them, which is defined as follows:
F s 2 u = h ( x u ) + δ
δ = arg   max | | δ | | ϵ   D K L [ S ( f ( h ( x u ) ; θ s ) ) | | S ( f ( h ( x u ) + δ ; θ s ) ) ]
y s 2 u = O ( S ( f ( F s 2 u ) )
In the above formula, δ represents the Virtual Adversarial Training (VAT) [48] perturbation generated by the student network. S ( f ( h ( x u ) ; θ s ) ) represents the the softmax probability of predicted label generated by the image without image perturbation and feature perturbation through the student network. S ( f ( h ( x u ) + δ ; θ s ) ) is the softmax probability of a predicted label generated after the image without image augmentation is augmented with VAT perturbation. D K L [ ] represents the Kullback-Leibler divergence, which is a measure of the difference between two probability distributions.
As for the teacher network, the same unlabelled image is also divided into two processing steps: weak data augmentation and no augmentation. After the two images pass through the encoder h, it generates two sets of image features, F t 1 u = h ( x u ; θ t ) and F t 2 u . We pass F t 1 u directly through the decoder f, i.e.,: y t 1 u = O ( S ( f ( F t 1 u ) ) . For F t 2 u , we simply jitter its features; specifically, we add extremely weak uniform distribution noise to the features. Here, we do not use the same processing method as the student network, because the VAT perturbation is undoubtedly a strong perturbation at the feature level. Here, we have transferred the idea of image-level perturbation, where the prediction of weak perturbation assists in optimizing the prediction of strong perturbation. Therefore, we chose a feature jittering method in the teacher network that has less impact on the features. After adding the feature perturbations and passing it through the decoder, the prediction map y t 2 u = O ( S ( F t 2 u ) ) is generated.
After the teacher and student networks generate their respective prediction maps, different types of losses are computed. For the unsupervised loss part, we use the prediction maps y s 1 u and y t 1 u , with the unsupervised loss being defined as follows:
L u n s u p = 1 | N u | ( x i u ) N u l c e ( y s 1 i u , y t 1 i u )
For y s 2 u and y t 2 u , we will perform contrastive learning calculations. The specific calculation steps will be given in the next section.

3.1.2. Contrastive Learning with Entropy Threshold Assisted Feature Sampling

Contrastive learning, initially utilized in image classification tasks, aims to define the query, the positive key, and the negative key. It seeks to learn the similarity between the query and the positive key, while simultaneously reinforcing the difference between the query and the negative key. As this method is incorporated into semantic segmentation, the sample scope transitions from image-level to pixel-level. In contrastive learning, a key task is to select appropriate samples as the query, positive keys, and negative keys. The selection of query and positive key is relatively straightforward, but the crucial part lies in choosing suitable negative keys so that the final contrastive learning result can be as accurate as possible. Herein, we introduce the concept of entropy to tackle this sampling problem.
Entropy is a metric of the uncertainty or randomness of data or a probability distribution. The entropy of the probability distribution of each pixel can be used as a measure to determine the uncertainty or confusion of the prediction. If a pixel’s Softmax probability entropy value is low, it means that the model is very certain about regarding its prediction that the pixel belongs to a specific category. Conversely, if a pixel’s Softmax probability entropy value is high, it means that the model is uncertain about regarding its class prediction for that pixel. Based on this, we establish an entropy threshold and select pixel samples with an entropy value lower than this threshold as the negative key. In this way, we can ensure to some extent that the selected negative key is less likely to be a misclassified positive key, thus helping to enhance the precision of contrastive learning. The process of contrastive learning is shown in Figure 2.
In this paper, the contrastive learning loss used is as follows:
L c o n = 1 C × M c C i M l o g e ( z c i , z c i + / τ ) e ( z c i , z c i + / τ ) + j J e ( z c i , z c i j / τ )
C represents the aggregate count of segmentation categories, M represents the number of queries and positive keys, and J represents the number of negative keys; z = r h ( x ) ; τ represents the temperature coefficient.
Contrastive learning is performed on the pixel features of the pseudo-label image, and the overall strategy can be summarized as follows: (a) selecting a query; (b) selecting a positive key; (c) selecting a negative key. At the same time, we define the entropy value of pixels.
E i j = c C S i j ( c ) l o g S i j ( c )
where S i j represents the softmax probability of the j-th pixel in the i-th image being of class c. To assist in selecting the negative key, we set a threshold α and also define the key-value.
When perturbing the features, the teacher network undergoes weak perturbation, while the student network undergoes strong perturbation. Therefore, we mostly select the features generated by the teacher network as our queries. The definition of query is as follows.
Q c s = 1 ( y ^ i j = c ) ( r ( h ( x i j ) + δ ) )
Q c t = 1 ( y ^ i j = c ) ( r ( h ( x i j ) ) )
Subject to E i j < α . We generate 80% of all queries from samples in the teacher network, and the remaining 20% are sampled from the student network. For the i-th labelled image, x i j is the j-th pixel, where y ^ i j is the predicted label generated by the network. 1 ( ) is the indicator function.
The selection rule for the positive key P c is consistent with the query. After selecting the query and positive key, the negative keys N c are randomly sampled from the remaining features, and also need to satisfy the entropy threshold condition, defined as follows:
N c Uniform ( z Q c , P c ) a n d E i j < α
After selecting the final key values, the contrastive loss L c o n can be calculated.
The final loss update of the model in this paper is:
L = L s u p + L u n s u p + L c o n

3.2. Datasets

3.2.1. iSAID

In this paper, we employ the iSAID dataset [49] to evaluate the efficacy and performance of our proposed method for semantic segmentation. The iSAID dataset is a large-scale and challenging benchmark in the field of remote sensing, containing a diverse collection of 15 object categories and a total of 2806 high-resolution aerial images. To facilitate the experimental process, the dataset is divided into a training set with 1411 images and a test set comprising 458 images. During the training phase, we adopt a data augmentation strategy that involves randomly cropping the images to a fixed size of 512 × 512 pixels.

3.2.2. Potsdam

We also employ the Potsdam dataset [50]. The Potsdam dataset, a benchmark dataset in the realm of remote sensing semantic segmentation, is derived from the Potsdam region in Germany. It is a collaborative product of the International Society for Photogrammetry and Remote Sensing (ISPRS) and the German Society for Photogrammetry, Remote Sensing, and Geoinformation (DGPF). The dataset comprises high-resolution aerial images with a spatial resolution of 5 cm per pixel, captured using an UltraCamXp large-format digital aerial camera. Each image in the dataset has a larger dimension compared to the Vaihingen dataset, covering an area of approximately 1000 × 1000 meters, providing more detailed ground information. In the Potsdam dataset, there are a total of 24 annotated images used as the training set and 14 test images used as the test set. During the training phase, we randomly cropped the images to a size of 512 × 512 pixels. Additionally, the image augmentation method used for the Potsdam dataset is the same as that used for the iSAID dataset.

3.3. Evaluation Metrics

The assessment measure employed here is the Mean Intersection over Union ( m I o U ). I o U , a commonly adopted metric for tasks of semantic segmentation, gauges the overlap extent between predicted segments and their respective ground truths. This metric provides a reliable means of assessing the performances of segmentation algorithms, as it takes into account both the false positives and false negatives, thereby offering a comprehensive view of the model’s accuracy. In the context of semantic segmentation, the I o U is often reported as the mean I o U ( m I o U ), which is the average I o U across all classes present in the dataset.
I o U = T P T P + F P + F N
where T P stands for the count of true positives, T N is the number of true negatives, F P symbolizes the quantity of false positives, and F N is the number of false negatives.
The F 1 score is the harmonic mean of precision and recall, serving as a comprehensive performance evaluation metric. The F 1 score simultaneously considers the model’s Precision and Recall, being sensitive to both false positives and false negatives. Especially in cases of class imbalance, the F 1 score can provide a more accurate performance measurement. The calculation formula for the F 1 score is as follows:
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
By using both the F 1 score and I o U , we can simultaneously evaluate the model’s accuracy in classification prediction and spatial localization, providing a more comprehensive evaluation of model performance.

3.4. Implementation Detail

The network used in the experiment is Deeplab V3+, with ResNet-101 as the backbone. Images are randomly cropped to a size of 512 × 512 before being input into the training network, with a batch size of 1 and a training epoch of 200. The learning rate is set to a decayed learning rate, l r ( 1 i t e r / i t e r m a x ) 0.9 . The stochastic gradient descent (SGD) optimizer is used, with an initial learning rate set to 0.01 and a learning rate decay of 0.0005. The dataset is divided into 1/2, 1/4, 1/8, and 1/16 labelled images, with the remaining images being used as unlabelled images for training the model. The weight of the auxiliary loss is set to 0.4. When updating the parameter θ t , the EMA smoothing factor γ is set to 0.99.

4. Results and Discussion

4.1. Comparison Experiments

4.1.1. iSAID

Following the general dataset ratio settings in semi-supervised semantic segmentation, we use ratios of labelled to unlabelled datasets of 1/8, 1/4, and 1/2, as well as conducting experiments using the full dataset. For comparative experiments, all datasets used are kept entirely consistent. To mitigate the impact of randomness, we repeat each experiment under the same settings three times and take the average.
In terms of contrast method selection, we chose the classic Mean Teacher (MT) method for natural scenes, as well as the methods with superior performing methods, RanPaste and GCT. In the field of remote sensing images, we selected recently proposed methods that have shown good performances in semi-supervised semantic segmentation for remote sensing images: ICNet and PICS. The experimental results compared with these methods are shown in Table 1. At the same time, we used bold fonts to represent the best evaluation metrics in this series of comparative experiments. It can be seen that our method achieved the best performance in terms of the mIoU metric on 1/8, 1/4, and 1/2 dataset proportions, with respective results of 42.65%, 45.08%, and 49.11%. Though our metrics on the complete dataset didn’t reach the topmost rank, they are competitively close to the highest-performing PICS-I method—with our method’s mIoU trailing by merely 0.45%. Furthermore, our performance outstrips all of the other methods analyzed in the study.
The iSAID dataset consists of 15 categories, and within these 15 categories, the segmentation difficulty varies. For example, categories like Baseball Diamond and Tennis Court demonstrate better segmentation results, whereas under the same conditions, categories like Helicopter and Bridge prove to be much more difficult to segment. The mIoU for specific categories is shown in Table 2, where the dataset ratio is 1/4. From the results, it can be seen that in the difficult categories, our method has a distinct advantage over other methods, with a clear segmentation advantage in categories like Helicopter and Bridge. These results demonstrate the effectiveness of our improvements in feature space. Specific analyses will be elaborated in conjunction with the upcoming ablation studies.
The visualization of the segmentation results on the iSAID dataset is shown in Figure 3. What we’ve presented here are the results of the segmentation, performed on a quarter of the dataset. To better demonstrate the advantages of our segmentation, we have chosen to present segments that are challenging to separate, and we contrast these with the results from the PICS method. We chose the PICS method specifically because it exhibits the best performance in experiments aside from our own. As is evident, our method is the only one that can identify helicopters. Moreover, when considering other objects, our method visibly provides superior segmentation results.

4.1.2. Potsdam

Remote sensing scenes are diverse and complex. The categories included in the iSAID dataset cannot fully represent remote sensing images. Therefore, we need additional datasets to demonstrate the universality of our method on remote sensing images. Here, we choose the Potsdam dataset, which contains a large number of buildings and streets. The experimental setup is the same as the iSAID dataset. Table 3 shows the Potsdam segmentation performance of different methods under different dataset proportions.
From the results, it can be seen that our method also achieved the best overall performance on the Potsdam data. All comparison methods on the Potsdam dataset have superior performance. Although our method achieved the best performance in mIoU, with results of 79.33%, 85.01%, and 85.93% under the dataset proportions of 1/8, 1/4, and 1/2, there is no noticeable gap. This is because the targets in the Potsdam dataset are quite distinct and easy to classify. The IoU data of specific categories under the dataset proportion of 1/4 is shown in Table 4, where our method demonstrated the best performance in the Buildings, Trees, and Cars categories.
The visualization of the segmentation on the Potsdam dataset is shown in Figure 4.

4.2. Ablation Study

In this section, to delve deeper into the impact of each module on the overall method, and to conduct accurate quantitative assessments, we decided to perform a detailed ablation study on each module. In this series of experiments, we chose to use the iSAID dataset. The reason for this choice is that, compared to the Potsdam dataset, the images in the iSAID dataset are more challenging to accurately segment, and the differences in segmentation metrics can more effectively reflect the practical utility of each module. To ensure the fairness and consistency of the experiments, we uniformly used a 1/4 dataset ratio in all ablation studies.
We will begin by conducting a detailed ablation study on the overall modules. In our approach, there are several key modules that can be ablated, their functions and roles are as follows:
Feature Disturbed Module (FDM): The primary function of this module is to enhance the disturbance of features encoded by the student and teacher networks. Through this method, we can improve the robustness and the generalization ability of the model while ensuring network performance.
Contrastive Learning Module ( L c ): The main task of this module is to deeply optimize the feature space using a contrastive learning strategy. In this module, we have set an entropy threshold to filter negative keys, aiming to enhance the effectiveness and precision of contrastive learning. To verify the effectiveness of this strategy, we also need to conduct a detailed ablation study on this module.
Through such ablation studies, we can gain a deeper understanding and evaluation of the specific contributions and impacts of each module to the overall method, thereby better optimizing our model and approach. The results of the ablation study are shown in Table 5. The overall results of the ablation experiments confirm that each module positively contributes to the performance of the model. We ablated all modules, resulting in a basic model that served as our baseline for comparison. During the ablation of the contrastive learning module, we used the basic cross-entropy loss as the loss function.
On this basic model, we incrementally added different modules, starting with the FDM module. The results showed that just by adding this one module, the model’s mIoU increased by 2.32%. This result validates the effectiveness of our strategy of transferring consistency regularization criteria to feature disturbance. After separately adding the contrastive learning module, the mIoU increased by 1.11%. On this basis, we added an entropy threshold for auxiliary key-value screening, which further improved the mIoU by 0.69%. Although we added the contrastive learning module alone without introducing any feature-level disturbance, it is undeniable that contrastive learning optimizes the feature space, especially in the case of remote sensing images. Finally, we combined all of the modules to form our final model. This model achieved a mIoU increase of 3.85%. This result fully demonstrates the effectiveness of our modular design and the significant contribution of each module to the performance of the model.
To more clearly demonstrate the effects of our ablation study, we have additionally incorporated t-SNE visualizations using the Potsdam dataset, as illustrated in Figure 5. From these visualizations, it is evident that the feature disturbed module contributes to optimizing the feature space. Furthermore, employing contrastive learning on top of FDM significantly enhances the optimization of the feature space. This combination of techniques clearly delineates the improvements in feature representation and model performance.
Within each module, we have made meticulous configurations. For example, in the feature disturbed module, we opted for a combination of strong and weak disturbances, and determined how to set the entropy threshold to achieve the best result. To demonstrate and to validate the rationality of these settings, we conducted a series of thorough and detailed ablation experiments.

4.2.1. Ablation Study of FDM

The proposal of FDM was inspired by the optimization assistance of strong and weak enhancements at the image level. In its construction, we used a strong enhancement: VAT, and a weak enhancement: feature jittering. We set up the ablation experiment for FDM as follows: the teacher network and the student network each generate VAT for feature perturbation; the teacher network uses the feature jittering method, and the student network uses VAT, as in our original experimental setup; both the teacher network and the student network use feature jittering for feature perturbation. The experimental results are shown in Table 6.
The results from the Table 6 indicate that the simultaneous use of VAT and feature jittering achieved the best performance. However, when using VAT or feature jittering alone, the mIoU decreased by 2.82% and 1.63%, respectively. When using VAT alone, the features of both networks underwent significant perturbations. Although contrastive learning was used to match the feature space, the excessive randomness led to a more chaotic and disordered feature space, increasing the learning difficulty and degrading the performance. In contrast, when using feature jittering alone, the final mIoU score was quite close to the result without feature perturbations, indicating a certain degree of performance improvement. This is because when we applied feature jittering, the noise we set caused only slight perturbations to the features. Therefore, the difference between the perturbed feature maps and the original features was not significant. At the same time, it added a certain degree of randomness, which helped to maintain feature stability after network training, thereby improving model performance to some extent.
Our original experimental setup, namely the use of both VAT and feature jittering, showed markedly better performance than the above two scenarios. This validates the reasonableness of our method setup and also proves that using the results generated by weak augmentation to assist the results generated by strong augmentation in migrating to the feature space is equally applicable.

4.2.2. Ablation Study of the Entropy Threshold

During the process of feature key-value selection, we employed an entropy threshold as an auxiliary tool. The setting of this entropy threshold requires careful selection, for which we conducted a series of exhaustive ablation experiments. Our method involves calculating the entropy of all pixel points and then selecting a certain percentile as the entropy threshold. The advantage of this approach is that the threshold can adjust according to the overall change in entropy, rather than solely relying on a fixed threshold. This method ensures that our feature key-value selection is more flexible and adaptable, and better equipped to handle different data distributions and complexities. We experimented with multiple percentiles for the entropy threshold, including 0.4, 0.6, 0.7, 0.8, and 0.99. The final mIoU results are displayed in Table 7.
The experimental results show that selecting the 0.7 percentile as the entropy threshold yields the best results. When we filter key-values, we choose those that are below a certain entropy threshold, resulting in more accurate key-values. If the threshold is set too high, such as at the 0.8 and 0.99 percentile entropy values, many key-values that the model is uncertain about may still be selected, leading to classification errors, with their mIoUs being 0.14% and 0.49% lower than the optimal setting, respectively. This is because an excessively high threshold might affect the selection of positive keys, potentially misclassifying latent postive keys as negative keys, thus leading to a decrease in results. Conversely, if the key-value selection is too low, such as 0.4 and 0.6, it does not achieve better results than 0.7, with the mIoUs being 0.37% and 0.06% lower than the optimal setting, respectively. This is because, although some pixel points have high degrees of uncertainty, they may contain valuable information such as the boundary information of the positive category, which has high predictive uncertainty but is important for model training. Therefore, blindly lowering the threshold does not improve performance. Based on this, we chose a threshold percentile of 0.7, achieving the best results in model training and prediction.

5. Conclusions

In this paper, we have delved into the unique characteristics of remote sensing images and have introduced an innovative semi-supervised semantic segmentation framework. This framework synergistically integrates consistency regularization methods and contrastive learning techniques, significantly advancing the field of semi-supervised semantic segmentation. Initially, we establish a unique learning paradigm for consistency regularization, introducing perturbations at the feature level, and enhancing both strong and weak perturbations to maintain the stability of features after training. Secondly, for the feature space of the images, we introduce contrastive learning to achieve better partitioning and classification. By leveraging the entropy threshold to assist in feature selection, we are able to select more accurate positive keys and negative keys, making the results of contrastive learning more precise. Our method surpasses the state-of-the-art techniques on the Potsdam and iSAID datasets. Concerning the entropy threshold setting, our current model utilizes a static value. However, we recognize the potential for performance optimization through a dynamically adapting entropy threshold that evolves in tandem with model training. This insight forms the basis for our future research direction, where we aim to develop and validate an effective dynamic entropy threshold model, further pushing the boundaries of semi-supervised semantic segmentation in remote sensing imagery.

Author Contributions

Conceptualization, Y.X.; methodology, Y.X.; software, Y.X.; validation, Y.X., X.Q.; formal analysis, Y.X.; investigation, Y.X., Z.F., X.Q., X.L. and Y.G.; data curation, Y.X.; writing—original draft preparation, Y.X.; writing—review and editing, Y.X., Z.F., X.Q., X.L. and Y.G.; visualization, Y.X.; supervision, Z.F., X.L. and Y.G.; project administration, Z.F. and X.L.; funding acquisition, Z.F. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by the Strategic Priority Research Program of the Chinese Academy of Sciences, Grant No. XDA0310502, and the Future Star of Aerospace Information Research Institute, Chinese Academy of Sciences, Grant No. E3Z10701.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available in a publicly accessible repository that does not issue DOIs and are provided by third parties. The datasets that we use are available at https://captain-whu.github.io/iSAID/index.html (accessed on 1 June 2023) and https://seafile.projekt.uni-hannover.de/f/429be50cc79d423ab6c4/ (accessed on 1 June 2023).

Acknowledgments

We are grateful to Lv Chen for her valuable contributions in results generation and reporting.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, G.; Xie, X.; Han, J.; Guo, L.; Xia, G.S. Remote sensing image scene classification meets deep learning: Challenges, methods, benchmarks, and opportunities. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3735–3756. [Google Scholar] [CrossRef]
  2. Marmanis, D.; Schindler, K.; Wegner, J.D.; Galliani, S.; Datcu, M.; Stilla, U. Classification with an edge: Improving semantic image segmentation with boundary detection. ISPRS J. Photogramm. Remote Sens. 2018, 135, 158–172. [Google Scholar] [CrossRef]
  3. Mao, Y.; Chen, K.; Diao, W.; Sun, X.; Lu, X.; Fu, K.; Weinmann, M. Beyond single receptive field: A receptive field fusion-and-stratification network for airborne laser scanning point cloud classification. ISPRS J. Photogramm. Remote Sens. 2022, 188, 45–61. [Google Scholar] [CrossRef]
  4. Wei, H.; Zhang, Y.; Chang, Z.; Li, H.; Wang, H.; Sun, X. Oriented objects as pairs of middle lines. ISPRS J. Photogramm. Remote Sens. 2020, 169, 268–279. [Google Scholar] [CrossRef]
  5. Qi, X.; Mao, Y.; Zhang, Y.; Deng, Y.; Wei, H.; Wang, L. PICS: Paradigms Integration and Contrastive Selection for Semisupervised Remote Sensing Images Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–19. [Google Scholar] [CrossRef]
  6. Yang, Z.; Yan, Z.; Diao, W.; Zhang, Q.; Kang, Y.; Li, J.; Li, X.; Sun, X. Label Propagation and Contrastive Regularization for Semi-supervised Semantic Segmentation of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5609818. [Google Scholar]
  7. Zhang, B.; Zhang, Y.; Li, Y.; Wan, Y.; Guo, H.; Zheng, Z.; Yang, K. Semi-supervised deep learning via transformation consistency regularization for remote sensing image semantic segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 5782–5796. [Google Scholar] [CrossRef]
  8. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. Adv. Neural Inf. Process. Syst. 2017, 30, 1195–1204. [Google Scholar]
  9. Olsson, V.; Tranheden, W.; Pinto, J.; Svensson, L. ClassMix: Segmentation-Based Data Augmentation for Semi-Supervised Learning. In Proceedings of the Workshop on Applications of Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  10. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  11. French, G.; Laine, S.; Aila, T.; Mackiewicz, M.; Finlayson, G. Semi-supervised semantic segmentation needs strong, varied perturbations. arXiv 2019, arXiv:1906.01916. [Google Scholar]
  12. Chen, X.; Yuan, Y.; Zeng, G.; Wang, J. Semi-Supervised Semantic Segmentation with Cross Pseudo Supervision. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2613–2622. [Google Scholar] [CrossRef]
  13. Kim, J.; Min, Y.; Kim, D.; Lee, G.; Seo, J.; Ryoo, K.; Kim, S. Conmatch: Semi-supervised learning with confidence-guided consistency regularization. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part XXX. Springer: Berlin/Heidelberg, Germany, 2022; pp. 674–690. [Google Scholar]
  14. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  15. Zhang, Y.; Wu, Z.; Zhang, Y.; Guo, J.; Yang, P.; Chen, G.; Huang, Q.; Luo, P. Bootstrapping Semantic Segmentation with Regional Contrast. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14841–14850. [Google Scholar]
  16. Sohn, K.; Berthelot, D.; Carlini, N.; Zhang, Z.; Zhang, H.; Raffel, C.A.; Cubuk, E.D.; Kurakin, A.; Li, C.L. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Adv. Neural Inf. Process. Syst. 2020, 33, 596–608. [Google Scholar]
  17. Wang, J.X.; Chen, S.B.; Ding, C.H.; Tang, J.; Luo, B. Semi-supervised semantic segmentation of remote sensing images with iterative contrastive network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  18. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  20. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  21. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  22. Wu, H.; Zhang, J.; Huang, K. FastFCN: Rethinking Dilated Convolution in the Backbone for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October– 2 November 2019; pp. 2634–2642. [Google Scholar]
  23. Tao, H.; Duan, Q.; Lu, M.; Hu, Z. Learning Discriminative Feature Representation with Pixel-level Supervision for Forest Smoke Recognition. Pattern Recognit. 2023, 143, 109761. [Google Scholar] [CrossRef]
  24. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 7262–7272. [Google Scholar]
  25. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  26. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  27. Rasmus, A.; Valpola, H.; Honkala, M.; Berglund, M.; Raiko, T. Semi-Supervised Learning with Ladder Networks. Computer Science 2015, 9 (Suppl. S1), 1–9. [Google Scholar]
  28. Ouali, Y.; Hudelot, C.; Tami, M. Semi-Supervised Semantic Segmentation with Cross-Consistency Training. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12671–12681. [Google Scholar] [CrossRef]
  29. Ke, Z.; Qiu, D.; Li, K.; Yan, Q.; Lau, R.W. Guided collaborative training for pixel-wise semi-supervised learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 429–445. [Google Scholar]
  30. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  31. French, G.; Mackiewicz, M.; Fisher, M. Self-ensembling for visual domain adaptation. arXiv 2017, arXiv:1706.05208. [Google Scholar]
  32. Liu, Y.; Tian, Y.; Chen, Y.; Liu, F.; Belagiannis, V.; Carneiro, G. Perturbed and Strict Mean Teachers for Semi-supervised Semantic Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4248–4257. [Google Scholar] [CrossRef]
  33. Yang, L.; Zhuo, W.; Qi, L.; Shi, Y.; Gao, Y. St++: Make self-training work better for semi-supervised semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4268–4277. [Google Scholar]
  34. Hung, W.C.; Tsai, Y.H.; Liou, Y.C.; Lin, Y.Y.; Yang, M.H. Adversarial Learning for Semi-Supervised Semantic Segmentation. In Proceedings of the British Machine Vision Conference (BMVC), Newcastle, UK, 3–6 September 2018. [Google Scholar]
  35. Fu, X.; Peng, Y.; Liu, Y.; Lin, Y.; Gui, G.; Gacanin, H.; Adachi, F. Semi-supervised specific emitter identification method using metric-adversarial training. IEEE Internet Things J. 2023, 10, 10778–10789. [Google Scholar] [CrossRef]
  36. Hung, W.C.; Tsai, Y.H.; Liou, Y.T.; Lin, Y.Y.; Yang, M.H. Adversarial learning for semi-supervised semantic segmentation. arXiv 2018, arXiv:1802.07934. [Google Scholar]
  37. Alonso, I.; Sabater, A.; Ferstl, D.; Montesano, L.; Murillo, A.C. Semi-supervised semantic segmentation with pixel-level contrastive learning from a class-wise memory bank. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 8219–8228. [Google Scholar]
  38. Wang, W.; Zhou, T.; Yu, F.; Dai, J.; Konukoglu, E.; Gool, L.V. Exploring Cross-Image Pixel Contrast for Semantic Segmentation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 7283–7293. [Google Scholar] [CrossRef]
  39. Zhao, X.; Vemulapalli, R.; Mansfield, P.A.; Gong, B.; Green, B.; Shapira, L.; Wu, Y. Contrastive Learning for Label Efficient Semantic Segmentation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 10603–10613. [Google Scholar] [CrossRef]
  40. Zhou, Y.; Xu, H.; Zhang, W.; Gao, B.; Heng, P.A. C3-SemiSeg: Contrastive Semi-supervised Segmentation via Cross-set Learning and Dynamic Class-balancing. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 7016–7025. [Google Scholar] [CrossRef]
  41. Chen, X.; He, K. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2021; pp. 15750–15758. [Google Scholar]
  42. Lai, X.; Tian, Z.; Jiang, L.; Liu, S.; Zhao, H.; Wang, L.; Jia, J. Semi-supervised semantic segmentation with directional context-aware consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1205–1214. [Google Scholar]
  43. Yang, L.; Qi, L.; Feng, L.; Zhang, W.; Shi, Y. Revisiting weak-to-strong consistency in semi-supervised semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7236–7246. [Google Scholar]
  44. Kerdegari, H.; Razaak, M.; Argyriou, V.; Remagnino, P. Urban scene segmentation using semi-supervised GAN. In Proceedings of the Image and Signal Processing for Remote Sensing XXV. International Society for Optics and Photonics, Strasbourg, France, 9–11 September 2019; Volume 11155, p. 111551H. [Google Scholar]
  45. Zhang, H.; Hong, H.; Zhu, Y.; Zhang, Y.; Wang, P.; Wang, L. Semi-Supervised Semantic Segmentation of SAR Images Based on Cross Pseudo-Supervision. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; IEEE: Manhattan, NY, USA, 2022; pp. 1496–1499. [Google Scholar]
  46. Fang, B.; Li, Y.; Zhang, H.; Chan, J.C.W. Collaborative learning of lightweight convolutional neural network and deep clustering for hyperspectral image semi-supervised classification with limited training samples. ISPRS J. Photogramm. Remote Sens. 2020, 161, 164–178. [Google Scholar] [CrossRef]
  47. Hong, D.; Yokoya, N.; Xia, G.S.; Chanussot, J.; Zhu, X.X. X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data. ISPRS J. Photogramm. Remote Sens. 2020, 167, 12–23. [Google Scholar] [CrossRef] [PubMed]
  48. Takeru, M.; Shin-Ichi, M.; Shin, I.; Masanori, K. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1979–1993. [Google Scholar]
  49. Waqas Zamir, S.; Arora, A.; Gupta, A.; Khan, S.; Sun, G.; Shahbaz Khan, F.; Zhu, F.; Shao, L.; Xia, G.S.; Bai, X. isaid: A large-scale dataset for instance segmentation in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 28–37. [Google Scholar]
  50. ISPRS 2D Semantic Labeling Contest-Potsdam. Available online: https://www.isprs.org/education/benchmarks/UrbanSemLab/2d-sem-label-potsdam.aspx (accessed on 1 June 2023).
  51. Wang, J.X.; Chen, S.B.; Ding, C.H.; Tang, J.; Luo, B. RanPaste: Paste consistency and pseudo label for semisupervised remote sensing image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed method framework. This framework includes a student network and a teacher network. Both networks share the same architecture. The optimization of the framework is based on three loss functions. L s u p is generated from the predicted labels of the annotated images and the ground truth; L u n s u p is calculated from the predicted labels of two unannotated images that have undergone image enhancement; L c o n is generated from the predicted labels of two features-perturbed images calculated through contrastive learning.
Figure 1. Overview of the proposed method framework. This framework includes a student network and a teacher network. Both networks share the same architecture. The optimization of the framework is based on three loss functions. L s u p is generated from the predicted labels of the annotated images and the ground truth; L u n s u p is calculated from the predicted labels of two unannotated images that have undergone image enhancement; L c o n is generated from the predicted labels of two features-perturbed images calculated through contrastive learning.
Sensors 24 00730 g001
Figure 2. Entropy-assisted contrastive learning. The softmax probabilities of the labels predicted by the network pass through an entropy calculation module, obtaining the entropy value for each pixel. An entropy threshold is set, as shown by the red line in the figure. We select the pixels and their features below the entropy threshold as the key values for contrastive learning.
Figure 2. Entropy-assisted contrastive learning. The softmax probabilities of the labels predicted by the network pass through an entropy calculation module, obtaining the entropy value for each pixel. An entropy threshold is set, as shown by the red line in the figure. We select the pixels and their features below the entropy threshold as the key values for contrastive learning.
Sensors 24 00730 g002
Figure 3. iSAID dataset segmentation result visualization image.
Figure 3. iSAID dataset segmentation result visualization image.
Sensors 24 00730 g003
Figure 4. Potsdam dataset segmentation result visualization image.
Figure 4. Potsdam dataset segmentation result visualization image.
Sensors 24 00730 g004
Figure 5. t-SNE visualization of feature spaces on test images of Potsdam dataset.
Figure 5. t-SNE visualization of feature spaces on test images of Potsdam dataset.
Sensors 24 00730 g005
Table 1. Comparisons with the SOTA methods evaluated on the iSAID dataset.
Table 1. Comparisons with the SOTA methods evaluated on the iSAID dataset.
Method1/8 1/4 1/2 Full
mIoU(%) m F 1 (%) mIoU(%) m F 1 (%) mIoU(%) m F 1 (%) mIoU m F 1 (%)
MT [8]39.7656.90 41.9159.07 45.3362.38 49.9766.64
RanPaste [51]41.1158.27 42.3859.53 47.0664.00 50.2966.92
ICNet [17]42.1459.29 42.6759.82 46.8063.76 50.6567.24
GCT [29]40.0957.23 41.0358.19 46.9163.86 50.7467.32
PCIS [5]42.6359.78 44.2861.38 48.9165.69 53.9070.05
(ours )42.6559.80 45.0862.15 49.1165.87 53.4569.66
Table 2. Comparisons with the SOTA methods evaluated on the iSAID dataset.
Table 2. Comparisons with the SOTA methods evaluated on the iSAID dataset.
MethodSHRABDTCBCGTFBRLV
SV HC SP ST SBF PL HA mIoU(1/4)
MT [8]47.5356.2963.2764.7927.8430.329.0362.49
33.684.4369.4031.1740.5739.8447.9941.91
RanPaste [51]46.6350.5854.9869.3527.9929.399.3668.12
30.219.1466.4426.6852.1546.2048.4442.38
ICNet [17]51.4947.5666.4365.7624.7528.969.0465.28
35.728.8965.9721.4048.9849.0750.7242.67
GCT [29]49.1449.1144.9467.8824.6125.6011.6358.47
35.0710.7756.4825.9165.0042.0248.8141.03
PCIS [5]50.2049.3255.7670.4229.7528.1615.1365.46
34.1713.6568.6026.6553.7647.4755.7644.28
(ours)51.1249.4555.5571.4830.3929.5719.2365.81
34.3618.8368.5427.0154.6447.8652.4345.08
Table 3. Comparisons with the SOTA methods evaluated on the Potsdam dataset.
Table 3. Comparisons with the SOTA methods evaluated on the Potsdam dataset.
Method1/8 1/4 1/2
mIoU(%) m F 1 (%) mIoU(%) m F 1 (%) mIoU(%) m F 1 (%)
MT [8]78.9488.23 84.5291.61 85.1091.95
Ranpaste [51]77.9587.61 84.0191.31 85.2392.03
ICNet [17]78.6188.02 83.5991.06 85.0791.93
GCT [29]78.8088.14 84.1791.40 85.2292.02
PCIS [5]78.9588.24 84.6691.69 85.3692.10
(ours)79.3388.47 85.0191.90 85.9392.43
Table 4. Comparisons with the SOTA methods evaluated on the Potsdam dataset.
Table 4. Comparisons with the SOTA methods evaluated on the Potsdam dataset.
MethodImp.surf.BuildingsLow veg.TreeCarmIoU (1/4)
MT [8]90.5682.3679.9580.5789.1684.52
RanPaste [51]91.3182.3978.8579.2787.7284.01
ICNet [17]90.9482.4180.6080.5287.4784.39
GCT [29]91.4280.9478.6180.7688.2584.17
PCIS [5]91.2781.8280.6880.5888.9584.66
(ours)91.2282.8480.4581.3389.2185.01
Table 5. Ablation study on the effectiveness of components in our method.
Table 5. Ablation study on the effectiveness of components in our method.
FDM L c Entropy
Threshold
mIoU(%)m F 1 (%)
41.2358.39
43.5560.68
42.3459.49
43.0360.17
44.1861.28
45.0862.15
Table 6. Ablation study on the effectiveness of FDM.
Table 6. Ablation study on the effectiveness of FDM.
Feature PerturbationVAT & VATFJ & VATFJ & FJ
mIoU (%)42.2645.0843.45
m F 1 (%)59.4162.1560.58
Table 7. Ablation study on the effectiveness of entropy threshold.
Table 7. Ablation study on the effectiveness of entropy threshold.
Entropy
Threshold
0.40.60.70.80.99
mIoU (%)44.7145.0245.0844.9444.59
m F 1 (%)61.7962.0962.1562.0161.68
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xin, Y.; Fan, Z.; Qi, X.; Geng, Y.; Li, X. Enhancing Semi-Supervised Semantic Segmentation of Remote Sensing Images via Feature Perturbation-Based Consistency Regularization Methods. Sensors 2024, 24, 730. https://doi.org/10.3390/s24030730

AMA Style

Xin Y, Fan Z, Qi X, Geng Y, Li X. Enhancing Semi-Supervised Semantic Segmentation of Remote Sensing Images via Feature Perturbation-Based Consistency Regularization Methods. Sensors. 2024; 24(3):730. https://doi.org/10.3390/s24030730

Chicago/Turabian Style

Xin, Yi, Zide Fan, Xiyu Qi, Ying Geng, and Xinming Li. 2024. "Enhancing Semi-Supervised Semantic Segmentation of Remote Sensing Images via Feature Perturbation-Based Consistency Regularization Methods" Sensors 24, no. 3: 730. https://doi.org/10.3390/s24030730

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop