Next Article in Journal
Remote Sensing of Instantaneous Drought Stress at Canopy Level Using Sun-Induced Chlorophyll Fluorescence and Canopy Reflectance
Previous Article in Journal
Integrating Deep Learning and Hydrodynamic Modeling to Improve the Great Lakes Forecast
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semi-Supervised Cloud Detection in Satellite Images by Considering the Domain Shift Problem

1
Department of Aerospace and Geodesy, Data Science in Earth Observation, Technical University of Munich (TUM), 80333 Munich, Germany
2
School of Aerospace Science and Technology, Xidian University, Xi’an 710126, China
3
Remote Sensing Technology Institute, German Aerospace Center (DLR), 82234 Weßling, Germany
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(11), 2641; https://doi.org/10.3390/rs14112641
Submission received: 25 April 2022 / Revised: 25 May 2022 / Accepted: 30 May 2022 / Published: 31 May 2022
(This article belongs to the Topic Big Data and Artificial Intelligence)

Abstract

:
In terms of semi-supervised cloud detection work, efforts are being made to learn a promising cloud detection model via a limited number of pixel-wise labeled images and a large number of unlabeled ones. However, remote sensing images obtained from the same satellite sensor often show a data distribution drift problem due to the different cloud shapes and land-cover types on the Earth’s surface. Therefore, there are domain distribution gaps between labeled and unlabeled satellite images. To solve this problem, we take the domain shift problem into account for the semi-supervised learning (SSL) network. Feature-level and output-level domain adaptations are applied to reduce the domain distribution gaps between labeled and unlabeled images, thus improving predicted results accuracy of the SSL network. Experimental results on Landsat-8 OLI and GF-1 WFV multispectral images demonstrate that the proposed semi-supervised cloud detection network (SSCDnet) is able to achieve promising cloud detection performance when using a limited number of labeled samples and outperforms several state-of-the-art SSL methods.

Graphical Abstract

1. Introduction

With the development of the Earth observation technology, an increasing number of optical satellites are launched for Earth observation missions. Remote sensing images acquired from the optical satellites can serve environment protection [1], global climate change [2], hydrology [3], agriculture [4], urban development [5], and military reconnaissance [6]. However, since 60% earth’s surface is covered by clouds, the acquired optical remote sensing (RS) images are often contaminated by clouds [7]. In the field of meteorological, cloud information of RS images is useful in weather forecast [8], while, for earth surface observation missions, cloud coverage degrades the quality of satellite imagery. Therefore, it is important to improve RS images quality through cloud detection.
Over the past few decades, cloud detection from RS imagery has attracted much attention. Many advanced cloud detection technologies have been proposed. In this paper, we broadly categorize these methods into rule-based methods and machine learning-based methods. The rule-based methods are mostly developed from spectral/spatial domain [9,10,11,12]. These methods distinguish clouds from clear sky pixels by exploiting reflectance variations in visible, shortwave-infrared, and thermal bands. Rule-based methods have obvious flaws, i.e., they strongly depend on particular sensor models and have poor generalization performance. For example, Fmask algorithms [9,10,11] are developed for Sentinel-2 and Landsat 4/5/7/8 satellite images, while the multifeature combined (MFC) method [12] is developed for GF-1 wide field view (WFV) satellite images only. In addition, machine learning-based cloud detection methods have also attracted much attention due to their powerful data adaptability. The most representative machine learning-based cloud detection methods are maximum likelihood [13,14], support vector machine (SVM) [15,16], and neural network [17,18]. However, these methods heavily rely on hand-crafted features, such as color, texture, and morphological features, to distinguish clouds from clear sky pixels.
Recent years, with the development of deep learning, deep convolutional neural network (DCNN) methods have been rapidly developed and widely used for cloud detection from RS images. For example, U-Net and SegNet variants cloud detection frameworks [19,20,21,22,23], and multi-scale/level feature fusion cloud detection frameworks [24,25,26,27,28]. In addition, advanced convolutional neural network (CNN) models, such as CDnetV2 [7] and ACDnet [29], are developed for cloud detection from RS imagery with cloud–snow coexistence. To achieve real-time and onboard processing, lightweight neural networks, such as [30,31,32], are proposed for pixelwise cloud detection from RS images. However, most of the previous CNN-based cloud detection methods are based on supervised learning frameworks. Although these CNN-based cloud detection methods have achieved impressive performance, they heavily rely on a large number of training data with strong pixel-wise annotations. Some recent cloud detection works, such as unsupervised domain adaptation (UDA) [33,34] and domain translation strategy [35], have begun to explore how to avoid using pixel-wise annotations for cloud detection network training. However, these pixel-wise annotations free methods are not real label-free ones because they rely on other labeled datasets.
Obtaining data label is an expensive and time-consuming task, especially pixel-wise annotation. As illustrated in cityscapes dataset annotation work [36], it usually takes 1.5 h to label a pixel-wise annotation from a high-resolution urban scene image with pixel size of 1024 × 2048. For a remote sensing image with pixel size of 8 k × 8 k, it may take more hours to label a whole scene RS image according to such experience. Although it is easier to label the cloud pixel-wise samples individually, it may still take three to four hours for a tough case that contains a large number of tiny and thin clouds, which increases the heavy cost of manual labeling undoubtedly. In contrast, unlabeled RS images can be far more easily acquired than labeled ones [37]. Therefore, it desperately needs to exploit how to utilize a large number of unlabeled data to enhance the performance of cloud detection model.
In this paper, we proposed to use a semi-supervised learning (SSL) method [38,39] to train a cloud detection network. Because the SSL method is able to reduce the heavy cost of manual dataset labeling. In a semi-supervised segmentation framework, such as DAN [37] and s4GAN [40], the segmentation network (cloud detection network) is able to simultaneously take advantage of a large amount of unlabeled samples and a limited number of labeled examples for network’s parameter learning. The core of SSL method is a self-training [41] strategy, which is able to leverage pseudo-label generated from a large amount of unlabeled samples to supervise the segmentation network training [42]. This also means that accurate pseudo-label labeling is the key of self-training. Therefore, most advanced SSL methods, such as [40,43], focus on improving pseudo-labels of unlabeled samples to improve the performance of the SSL network.
SSL networks developed for tradition natural image segmentation, such as [38,39,40,43], may not achieve a promising performance for satellite images cloud detection due to RS images are different from traditional camera natural images. In addition, the data drift problem may appear between a limited number of labeled examples and a large number of unlabeled samples due to different cloud shapes and land-cover types on different satellite images. The SSL network trained with labeled samples is difficult to generalize to unlabeled samples due to the data drift problem. During training, the SSL network may produce prediction results with lower certainty when the network input with unlabeled data [40]. The prediction results of unlabeled samples has shown lower certainty and cannot generate accurate pseudo-labels, which makes self-training unfavorable for providing supervision signals of network training.
To solve this problem, we take the domain shift problem into account for the SSL framework. In this paper, inspired by unsupervised domain adaptation (UDA) method [44], we apply domain adaptation at the feature-level [45] and output-level [46,47] in the proposed semi-supervised cloud detection network (SSCDnet) to solve the data drift problem. In this paper, we use two available cloud cover validation datasets, i.e., Landsat-8 OLI (Operational Land Imager) [48] cloud cover validation dataset (https://landsat.usgs.gov/landsat-8-cloud-cover-assessment-validation-data, accessed on 24 April 2022) and GF-1 WFV [12] cloud and cloud shadow cover validation dataset (http://sendimage.whu.edu.cn/en/mfc-validation-data/, accessed on 24 April 2022), to comprehensively evaluate the proposed SSCDnet. These two datasets have been widely used for evaluating the performance of supervised CNN-based cloud detection methods [19,20,21,22,23,27].
In summary, the main contributions of this work are summarized as follows:
(i)
We propose a semi-supervised cloud detection framework, named SSCDnet, which learns knowledge from a limited number of pixel-wise labeled examples and a large number of unlabeled samples for cloud detection.
(ii)
We take the domain shift problem into account between labeled and unlabeled images and propose the feature-level and output-level domain adaptation method to reduce domain distribution gaps.
(iii)
We propose a double threshold pseudo-labeling method to obtain trustworthy pseudo label, which helps to avoid the effects of noise labels for self-training as much as possible and to further enhance the performance of SSCDnet.
This paper is organized as follows: in Section 2, we present the proposed SSCDnet in detail. Experimental datasets and networks training details are presented in Section 4. The experimental results and discussions are presented in Section 4 and Section 5, respectively, followed by conclusions in Section 6.

2. The Proposed Method

In this section, we provide a detailed introduction of the proposed SSCDnet, including the traditional semi-supervised segmentation framework, the proposed overall workflow of SSCDnet, feature/out-level domain adaptation, trustworthy pseudo label labelling, cloud detection network, and discriminator network structures.

2.1. Traditional Semi-Supervised Segmentation Framework

In a traditional semi-supervised segmentation framework [37,38,39,40], as shown in Figure 1, the segmentation network G simultaneously takes advantage of a large number of unlabeled samples and a limited number of labeled examples for a network’s parameter training. In this framework, there are two datasets, i.e., labeled dataset M l = { x l , y l } and unlabeled dataset M u = { x u } , where x l and x u are the input data of the segmentation network G and y l is a pixel-wise label of x l . p l and p u represent predicted results of x l and x u , respectively. p ^ u represents a pseudo label of p u .
During training, given an input labeled image x l , the segmentation network G is supervised by a standard cross-entropy loss L c e . When using the unlabeled data x u , the segmentation network is further supervised by a self-training loss L s t . That is, we use pseudo label p ^ u that generated from predicted result p u as the “ground truth” for self-training to enhance the semantic segmentation network G. Therefore, total training objective L G of the segmentation network G is defined as follows:
L G = L c e + λ s t L s t ,
where λ s t is weight used for minimizing the objective L G .
As shown in Figure 2, remote sensing images obtained from different places show large domain distribution gaps between each other due to different cloud shapes and land-cover types on Earth’s surface. Therefore, there is data drift between labeled samples and unlabeled ones in training dataset of SSL. In SSL framework, segmentation network G trained with labeled samples is hard to generalize to unlabeled ones due to data drift problems. It is difficult to produce a highly certain prediction result p u for unlabeled sample x u . Predicted results with low certainty leads to low quality pseudo-labels, thus affecting the supervision signal provided from the self-training loss L s t , and further affecting the performance of segmentation network G. Therefore, improving certainty of predicted results of unlabeled samples is the key to a semi-supervised learning framework.

2.2. Proposed Semi-Supervised Cloud Detection Framework

In this paper, we take the domain shift problem into account for the semi-supervised learning (SSL) framework to improve generalization of the segmentation network to generate trustworthy pseudo-label for self training. We improve a standard SSL network with the unsupervised domain adaptation (UDA) strategy and propose an improved semi-supervised cloud detection network as shown in Figure 3. UDA methods are able to help semantic segmentation networks to learn domain-invariant features at different representation layers, such as input-level (pixel-level) [49], feature-level [45], or output-level [46,47].
Different from the traditional SSL framework, we apply feature-level and output-level domain adaptations at the intermediate layers and the end layer of network, respectively, to reduce domain distribution gaps and improve the generalization performance of SSCDnet. The highly generalized network is able to generate highly certain predicted results for unlabeled samples. To further improve the quality of pseudo-label, instead of directly using the predicted results of unlabeled samples for network training, we proposed a trustworthy pseudo label labelling method to obtain a high-quality pseudo label. To be specific, similar to [40,43], we take advantage of the feedback information from output-level domain adaptation to obtain high-quality candidate labels. Then, we use a threshold strategy to obtain trustworthy regions from the high-quality candidate labels. Finally, trustworthy regions are considered as the ground-truth labels for network training through the self-training loss.
During training, given an input labeled image x l and its corresponding predicted result is p l , the cloud detection network is supervised by a standard cross-entropy loss L c e [27]. When inputting unlabeled data x u , segmentation network G predicts its corresponding result p u . Then, we generate the pseudo label p ^ u from predicted result p u by using proposed trustworthy pseudo label labelling method. Pseudo label p ^ u as the “ground truth” for self-training to enhance the semantic segmentation network G. Therefore, the semantic segmentation network is further supervised by self-training loss L s t and feature-level domain adaptation loss L f d a as well as output-level domain adaptation loss L o d a when inputting unlabeled data. The total loss of the semi-supervised cloud detection network, named SSCDnet, is defined as follows:
L G = L c e + λ s t L s t + λ o d a L o d a + λ f d a L f d a ,
where λ s t , λ o d a , and λ f d a are three regulation parameters used for minimizing the objective L G .
During training, we minimize L G for updating parameters of segmentation network G. The detailed information of the proposed feature/output-level domain adaptation strategies and trustworthy pseudo label labelling method will be introduced in following subsections.

2.3. Reducing Domain Distribution Gaps

2.3.1. Feature-Level Domain Adaptation

To reduce domain distribution gaps at the feature-level, we propose a class-relevant feature alignment (CRFA) strategy. As shown in Figure 4, we use predicted score maps of each class (i.e., cloud and background classes) as the attention maps to obtain class-relevant features. Then, we design a standard binary classification network as the discriminator to help segmentation network G to generate domain-invariant feature representations, thus helping to reduce domain distribution gaps between labeled samples and unlabeled ones at feature level. Therefore, the proposed CRFA domain adaptation consists of two stages: (i) class-relevant feature selection and (ii) class-relevant feature alignment.
To be specific, let H l k R W × H × C and H u k R W × H × C denote labeled and unlabeled samples’ features extracted from the k-th intermediate hidden layer of network G , respectively. Let p ˜ l i ( : , : , 1 ) and p ˜ l i ( : , : , 2 ) denote the spatial attention maps of cloud and background areas of labeled samples, respectively. Then, cloud-relevant feature C l k R W × H × C and background-relevant feature B l k R W × H × C of the labeled samples’ feature H l k R W × H × C are defined as follows:
C l k = H l k F p ˜ l i ( : , : , 1 ) ,
and
B l k = H l k F p ˜ l i ( : , : , 2 ) ,
where F ( · ) represents sampling operator, which includes down-sampling or up-sampling operators. Similarly, we are able to obtain cloud-relevant feature C u k R W × H × C and background-relevant feature B u k R W × H × C of unlabeled samples’ feature H u k R W × H × C .
After obtaining these class-relevant features, we input these features into discriminators as shown in Figure 4. For discriminator training, both discriminators D crfa ( k , 1 ) and D crfa ( k , 2 ) use cross-entropy domain classification loss [44] as the objective function. They are defined as follows:
L D crfa ( k , 1 ) = E [ log ( D crfa ( k , 1 ) ( C l k ) ) ] E [ log ( 1 D crfa ( k , 1 ) ( C u k ) ) ]
and
L D crfa ( k , 2 ) = E [ log ( D crfa ( k , 2 ) ( B l k ) ) ] E [ log ( 1 D crfa ( k , 2 ) ( B u k ) ) ] .
For segmentation network training, the adversarial objectives provided by discriminators ( D crfa ( k , 1 ) and D crfa ( k , 2 ) ) are defined as follows:
L D crfa ( k , 1 ) a d v = E [ log D crfa ( k , 1 ) ( C u k ) ]
and
L D crfa ( k , 2 ) a d v = E [ log D crfa ( k , 2 ) ( B u k ) ] .
Therefore, the class-relevant feature alignment loss L c r f a k of the k-th intermediate hidden feature is defined as follows:
L c r f a k = L D crfa ( k , 1 ) a d v + L D crfa ( k , 2 ) a d v .
Since the proposed feature-level domain adaptation is performed at multiple intermediate layers, the feature-level domain adaptation loss L f d a is provided from all intermediate layers’ CRFA losses { L c r f a k } k = 1 K , i.e.,
L f d a = k = 1 K L c r f a k .
where L f d a k is the feature domain adaptation loss of the k-th intermediate hidden feature.

2.3.2. Output-Level Domain Adaptation

Instead of directly aligning the predicted results, we align cloud and background objects images extracted from labeled and unlabeled samples, respectively, via a proposed class-relevant outputs alignment (CROA) method. Similar to the CRFA method, the CROA method is built on the standard adversarial learning framework. CROA consists of two discriminators, which are used to align cloud and background images, respectively. As shown in Figure 5, we use two standard binary classification networks as discriminators to reduce the domain distributions gaps between labeled and unlabeled datasets at the output level and improve the certainty of predicted results z u for unlabeled sample x u .
As shown in Figure 5, we use the cloud and background objects extracted from original images as the input data of different discriminators. To be specific, let z c u = p u ( : , : , 1 ) x u and z b u = p u ( : , : , 2 ) x u denote extracted cloud and background objects of unlabeled sample. Similarly, let z c l = p l ( : , : , 1 ) x l and z b l = p l ( : , : , 2 ) x l denote extracted cloud and background objects of the labeled sample, where ⊗ represents element-wise multiplication. For discriminator training, both discriminators D croa 1 and D croa 2 use cross-entropy domain classification loss [44] as the objective function. They are defined as follows:
L D croa 1 = E [ log ( D croa 1 ( z c l ) ] E [ log ( 1 D croa 1 ( z c u ) ) ] ,
and
L D croa 2 = E [ log ( D croa 1 ( z b l ) ] E [ log ( 1 D croa 1 ( z b u ) ) ] .
For segmentation network training, adversarial objectives provided by these discriminators are defined as follows:
L D c r o a 1 a d v = E [ log D c r o a 1 ( x c u ) ] ,
and
L D c r o a 2 a d v = E [ log D c r o a 2 ( x b u ) ] .
Therefore, the output-level domain adaptation objective L o d a is defined as follows:
L o d a = L D c r o a 1 a d v + L D c r o a 2 a d v .
As shown in Figure 6, we present the visualized experiment results on a GF-1 WFV image with or without applied domain adaptation. Results show that applying domain adaptation is able to improve the segmentation network to produce high certainly predicted results for unlabeled samples as shown in Figure 6c. High certainly predicted results ensure that we can obtain trustworthy pseudo label for self-training, thus improving the network’s cloud detection performance. More detailed information can be found in Section 4.1 (Ablation Study).

2.4. Trustworthy Pseudo Label Labelling

Pseudo label labelling is the main work for self-training in an SSL framework [38,39,40,43,50]. During self-training, pseudo labels predicted from the segmentation model serve as the “ground truth” to provide additional supervisory signals for network training, which makes the segmentation network able to leverage the unlabeled data. In this paper, we propose a double threshold pseudo-labelling method to efficiently obtain trustworthy regions from pseudo label. We use trustworthy regions as the ground truth for network training.

2.4.1. Candidate Labels Selection

As shown in Figure 7, we sample an unlabeled image x u into the segmentation network G, obtaining its corresponding predicted probability maps p u = G ( x u ) as well as extracted cloud and background objects z c u and z b u , respectively. We first use feedback information from the output-level domain adaptation to obtain candidate labels. Specifically, we select high-quality pseudo labels online based on two discriminator scores of output-level domain adaptation, i.e.,
D c r o a ( z c u ) > τ 1 , D c r o a ( z b u ) > τ 1 ,
where τ 1 is the threshold. Equation (16) is feedback information from the output-level domain adaptation. We treat these predicted labels satisfied Equation (16) as candidate labels. If not satisfied, we directly set self-training loss L s t = 0 .

2.4.2. Trustworthy Regions Selection from Candidate Labels

After we obtained the candidate labels, we set a confidence threshold τ 2 to discover trustworthy regions from the selected candidate labels, i.e.,
D c r o a 1 ( z c u ) > τ 1 , D c r o a 2 ( z b u ) > τ 1 , a n d p ( w , h , c ) u > τ 2 .
Equation (17) indicates that, if predicted probability of these regions is greater than τ 2 , we use these trustworthy regions as the ground-truth for self-training.

2.4.3. Self-Training Loss L s t

According to the above mentioned method, the self-training loss L s t is defined as:
L s t = p ^ ( w , h , c ) u log p ( w , h , c ) u , D c r o a 1 ( z c u ) > τ 1 , D c r o a 2 ( z b u ) > τ 1 , a n d p ( w , h , c ) u > τ 2 0 otherwise .
where p ^ u is the pseudo label generated from the prediction map p u by using a one-hot encoding scheme. p ( w , h , c ) u represents the predicted probability at the location (h,w) of the C-channel.

2.5. Network Architecture

2.5.1. Cloud Detection Network

Similar to most SSL approaches [40,43], we use DeepLabv2 [51] as our main cloud detection framework and resort to ResNet-101 [52] as the backbone to extract semantic segmentation information. DeepLabv2 uses atrous spatial pyramid pooling (ASPP) module [51], which incorporates multiple parallel dilated convolutional layers [53] with different sampling rates, to capture multi-scale features for robustness clouds detection. In this paper, we set the resolution of predicted probability map as 1/8 × 1/8 size of input image for fair comparison with previous semi-supervised works based on DeepLabv2, such as [37,40]. Then, we directly up-sample the predicted probability maps to the same size as the input images to obtain final predicted results.

2.5.2. Discriminator Network

In this paper, we design a standard binary classification network as the discriminator for both CRFA and CROA modules. Figure 8 illustrates the detailed structure of the designed discriminator. This discriminator contains four convolutional layers, a global average pooling layer, and a fully-connected layer. To be specific, there are four convolutional layers with 4 × 4 kernels. Four convolutional layers have { 256 , 192 , 128 , 64 } and { 64 , 128 , 256 , 512 } channels for CRFA and CROA modules, respectively. Each convolutional layer shares the same stride (stride = 2) and simultaneously followed by a Leaky-ReLU activation (slope = 0.2) and a dropout layer (dropout-rate = 0.5). After these convolution operations, a global average pooling layer and a fully-connected layer are designed to obtain confidence score for each input image.

3. Dataset and Experimental Settings

3.1. Experimental Dataset

In this paper, we use two available cloud cover validation datasets, i.e., Landsat-8 OLI cloud cover validation dataset [48] and GF-1 WFV cloud and cloud shadow cover validation dataset [12], to comprehensively evaluate the proposed SSCDnet. Table 1 shows the detailed information of Landsat-8 OLI and GF1-WFV multispectral images. Similar to [27], we follow the idea that the number of subimages with and without cloud in the training data should be balanced. Otherwise, the detection results would bias towards the majority. Therefore, we exclude some cloud-free and full cloud covered scenes. In addition, current cloud detection accuracy measurement is not robust when the cloud percentage is quite low [10]. A low cloud cover percentage in a scene may cause an apparent reduction in the cloud’s producer accuracy and user accuracy. Therefore, original large size images with cloud percentage less than 5% are usually removed in training and testing. Finally, we select 40 and 20 scenes images from both Landsat-8 and Gaofen-1 WFV dataset for network training and evaluation, respectively.
As illustrated in [54], CNNs are strongly biased towards recognizing textures for object classification. Therefore, we can select a limited number of channels data of multispectral data to evaluate the proposed SSCDnet. In this paper, we select channels 3 (green), 4 (red), and 5 (near-infrared) of Landsat-8 OLI data and channels 2 (green), 3 (red), and 4 (near-infrared) of Gaofen-1 WFV data for segmentation network training and testing. During training, all training data are cropped into subimages of pixel size 321 × 321. There are about 30 k and 75 k annotated sub-images in Landsat-8 OLI and GF-1 WFV training dataset, respectively. We select a portion of the data set for network supervised training and the remaining de-annotated portion for network self-training (unsupervised training). During testing, we divide the whole RS image into a series of sub-images with image size of 1200 × 1200 for network evaluation. The final detection result is obtained by merging results of sub-images.

3.2. Network Training Details and Parameters Setting

SSCDnet is trained under PyTorch framework (https://pytorch.org/, accessed on 24 April 2022). The operating system is Ubuntu 14.04 equipped with NVIDIA GTX 1080 Ti GPU. We optimize the segmentation (generator) and discriminator networks by using SGD algorithm [55] and Adam optimizer [56], respectively. During training, these two datasets share the following parameters settings. That is, learning rates for generator and discriminator network are 2.5 × 10 4 and 1 × 10 4 , respectively. Training decay policy is “poly” [57]. The number of mini-batch size, momentum, and weight decay are 4, 0.9, and 5 × 10 4 , respectively. SSCDnet is trained for 20 and 15 epochs on Landsat-8 OLI and GF-1 WFV dataset, respectively.
In addition, in order to improve the performance, we use the model pre-trained on the ImageNet dataset [58] to fine-tune the parameters of the backbone network (ResNet-101). We conduct feature-level domain adaptation tasks at the end of C o n v 4 _ x and C o n v 5 _ x residual blocks, i.e., K = 2 . We empirically set weight-parameters λ s t = 1.0 , λ f d a = 0.001 , and λ o d a = 0.1 for GF-1 dataset, set λ s t = 1.0 , λ f d a = 0.001 , and λ o d a = 0.01 for Landsat-8 OLI dataset. In addition, we empirically set threshold-parameters τ 1 = 0.6 and τ 2 = 0.55 to accurately generate trustworthy pseudo labels for both GF-1 and Landsat-8 OLI datasets.

4. Experimental Results

4.1. Ablation Study

To investigate the effectiveness of SSCDnet, we conduct a series of ablation studies on it. Five widely used quantitative metrics of RS images, i.e., mean intersection over union (MIoU), kappa coefficient (Kappa), overall accuracy (OA), producer accuracy (PA), and user accuracy (UA), are used to comprehensively measure the cloud detection results. All experiment results are obtained from the GF-1 WFV dataset.

4.1.1. Ablation Study on Loss Function

In this paper, the proposed SSCDnet is supervised by four loss functions, i.e., a standard cross-entropy loss L c e , self-training loss L s t , output-level domain adaptation loss L o d a , and feature-level domain adaptation loss L f d a . In Table 2, we list ablation results under different proportion of labeled samples ( 1 200 , 1 100 , 1 40 , 1 20 , and full labeled samples) to demonstrate the effects of each component loss.
Results in Table 2 show that the best performance is achieved by a combination of all optimal loss terms (we use { L c e , L s t , L o d a , L f d a } to represent this combination.), while the baseline framework supervised only by the standard cross-entropy loss L c e shows the worst performance. To be specific, a combination of L c e and L s t (i.e., { L c e , L s t } ) and combination of L c e , L s t , and L o d a (i.e., { L c e , L s t , L o d a } ) obtain significant improvement of +9.44% and +13.86% in Kappa, respectively, compared with only L c e loss when labeled sample proportion is 1 200 . In addition, { L c e , L s t } and { L c e , L s t , L o d a } also outperform L c e under other proportions of labeled sample settings, which shows that output-level domain adaptation and self-training strategies are able to improve the performance of segmentation network.
Based on { L c e , L s t , L o d a } , we further introduce feature domain adaptation loss L f d a to reduce feature-level domain distribution gaps between labeled and unlabeled datasets. Results in Table 2 show that { L c e , L s t , L o d a , L f d a } performs better than { L c e , L s t , L o d a } . In addition, it can be seen that the performance of { L c e , L s t , L o d a , L f d a } consistently outperforms those of other combinations under different proportions of labeled sample ( { 1 200 , 1 100 , 1 40 , 1 20 } ), which shows that the proposed { L c e , L s t , L o d a , L f d a } is a promising strategy for semi-supervised segmentation network.
In summary, results in Table 2 demonstrate that standard cross-entropy loss L c e , self-training loss L s t , output-level domain adaptation loss L o d a , and feature-level domain adaptation loss L f d a are beneficial for semi-supervised cloud detection work. Because applying domain adaptation strategy is able to reduce distribution gaps between labeled and unlabeled datasets and improve SSCDnet to generate trustworthy pseudo-labels for self-training, thus providing positive supervised signals for segmentation network learning. Combination of all these loss terms is able to achieve a state-of-the-art cloud detection performance when using a limited number of labeled samples.

4.1.2. Ablation Study on Feature-Level Domain Adaptation

In this paper, we use ResNet-101 [52] as the backbone network to extract feature maps and conduct domain adaptation study on the end features of C o n v 4 _ x , C o n v 5 _ x , C o n v 3 _ x , and C o n v 2 _ x residual blocks. In Table 3, we list the experiment results of ablation study on different intermediate layers. Experiment results show that applying feature domain adaptation on the end of C o n v 4 _ x and C o n v 5 _ x residual blocks, i.e., FDA_45, achieves the best performance. In contrast, applying feature domain adaptation on other layers’ features always shows worse performance than on C o n v 4 _ x and C o n v 5 _ x layers. Hence, we conduct the domain adaptation tasks at the end of C o n v 4 _ x and C o n v 5 _ x residual blocks (two intermediate layers K = 2 in feature-level domain adaptation). Furthermore, we find that applying domain adaptation tasks at above mentioned intermediate layers is able to achieve promising performance on the Landsat-8 OLI dataset.

4.1.3. Ablation Study on Self-Training’s Double Threshold

In this paper, we proposed a double threshold method to obtain trustworthy pseudo labels for network self-training. In Table 4, we investigate effectiveness of the proposed pseudo-labeling method.
In Table 4, we list the experiment results under different pseudo-labeling strategies, i.e., pseudo-labeling directly transforms from probability maps L s t ( N / A , N / A ) , pseudo-labeling based on a discriminator score of output-level domain adaptation L s t ( τ 1 , N / A ) , and pseudo-labeling based on both discriminator score of output-level domain adaptation and trustworthy regions selection L s t ( τ 1 , τ 2 ) , where τ 1 and τ 2 are used to obtain candidate labels and trustworthy regions, respectively. Results in Table 4 show that performance of the loss term L s t ( τ 1 , τ 2 ) is better than those of L s t ( τ 1 , N / A ) and L s t ( N / A , N / A ) , which demonstrates that trustworthy regions selection based on high confidence candidate label is a promising strategy for pseudo label selecting.

4.1.4. Hyper-Parameter Analysis

The proposed training objective L G has three balance weights, i.e., λ s t , λ o d a , and λ f d a . In this paper, since we use all the trustworthy regions’ pseudo labels for self-training, we directly set λ s t = 1.0 . Then, λ o d a and λ f d a are two important hyper-parameters that affect cloud detection results. In Table 5, we list a series of validation experimental results to investigate the impact of these hyper-parameters. As illustrated in Table 5, we first obtain the promising λ o d a by setting λ f d a = N / A (N/A means Not Applicable). Then, we obtain the promising λ f d a based on obtained promising λ o d a . Experimental results in Table 5 show that setting λ o d a = 0.10 and λ f d a = 0.001 is able to achieve promising performance. Similarly, we can obtain the promising hyper-parameter setting based on above mentioned methods on Landsat-8 OLI data cloud detection task.

4.2. Comparisons with State-of-the-Art Methods

4.2.1. Comparison Methods

For comprehensive evaluation, we compare deep adversarial network DAN [37], which focuses on training with both un-labeled and labeled images simultaneously to improve segmentation performance. Moreover, we compare two semi-supervised CNN-based semantic segmentation methods, i.e., Hung et al. [43] and s4GAN [40]. These two SSL methods are based on an adversarial network and achieve promising performance on PASCAL VOC 2012 [59] and Cityscapes datasets [36]. For fair comparison with the proposed SSCDnet, we use DeepLabv2 [51] and ResNet-101 [52] as a segmentation network and backbone of above-mentioned competing methods, respectively. In addition, baseline network DeepLabv2 [51] is also used as the competing method. During training, we retrain these CNN-based methods under their optimal parameter settings on Landast-8 OLI and GF-1 WFV datasets.

4.2.2. Results on GF-1 WFV Data

Table 6 shows the quantitative results in terms of average OA, MIoU, Kappa, PA, and UA on the GF-1 WFV testing dataset. We show these comparison results on four different proportions of labeled samples ( 1 200 , 1 100 , 1 40 , and 1 20 ). Meanwhile, we also give comparison results on fully labeled samples. Experiment results show that our proposed SSCDnet consistently outperforms these competing methods at different proportions of labeled samples. Notably, SSCDnet achieves 83.08% Kappa and 86.96% MIoU using only 0.5% (1/200) training data with pixel-wise annotation. Results of SSCDnet significantly outperform those of competing methods. SSCDnet also achieves the best performance on fully labeled data and shows a larger gain in terms of OA, MIoU, Kappa, PA, and UA than competing methods.
It can be seen that the baseline network, DeepLabV2 [51], trained with only labeled data, shows the worst results. DAN [37] shows better results than DeepLabV2 [51] due to it being able to effectively utilize unlabeled data for training. These semi-supervised methods, Hung et al. [43] and s4GAN [40] show better results than DeepLabV2 [51] and DAN [37] due to being able to learn knowledge from a limited number of labeled examples and a large number of additional unlabeled samples. Although these two methods show competitive cloud detection performance, there is still a gap in performance between them and the proposed SSCDnet. In addition, it also can be seen that all of the methods show promising cloud detection performance on GF-1 WFV data when using fully labeled data for network training. Even the worst baseline method DeepLabV2 [51] achieves 90.15% mIoU and 88.64% Kappa.
In Figure 9, we show qualitative results when labeled samples proportion is 1 200 . These images contain typical land-cover types, i.e., Figure 9a includes mountains, wetlands, farmland, and grass/crops, Figure 9b includes village, forest and ice/snow, Figure 9c includes water. Experiment results show that detection results of SSCDnet show more consistency with the ground-truth than those of other competing methods. Results of competing methods show a lot of misclassification pixels in thin cloud areas, while SSCDnet shows less misclassification pixels, which indicates that SSCDnet is able to achieve more promising performance on these imageries when using a limited number of labeled images.

4.2.3. Results on Landsat-8 OLI Data

In addition to the experiment results on the GF-1 data, we also conduct experiments on Landsat8 OLI data. In Table 7, we list quantitative results on the Landsat8 OLI testing dataset. Compared to other methods, the proposed SSCDnet also performs best on Landsat8 OLI data. For the low labeled sample’s proportion, such as 1 200 , SSCDnet is still able to achieve satisfactory results (90.77% MIoU and 88.72% Kappa). These results show consistency with the ground-truth. In contrast, performance of these competing methods is less than that of SSCDnet. When increasing the proportion of labeled samples, competing methods can improve their performance, but they are still inferior to SSCDnet. From Table 7, we find that the performance of SSCDnet at a labeled sample proportion of 1 20 approaches that of its full supervision and outperforms fully supervised DeepLabV2 [51] and DAN [37]. For fully labeled data, with the help of adversarial training and self-training strategies, SSCDnet still shows the best performance compared with other competing methods.
In Figure 10, we show qualitative results on three whole scene landsat-8 OLI images. These images contain typical land-cover types, i.e., Figure 10a includes mountains, forest, ice/snow, water, and wetlands areas. Figure 10b includes water, floating ice, urban, mountains, and forest areas. Figure 10c includes barren and desert areas. Experiment results are obtained when labeled sample proportion is 1 200 . Experiment results show that SSCDnet trained with a limited number of labeled samples can yield very competitive performance on Landsat-8 OLI data. It can be seen that results of these competing methods show a large number of misclassified pixels (red areas), especially in Figure 10c. There are a large number of misclassified pixels displayed in the thin cloud area. In contrast, SSCDnet works well on these images. Results of SSCDnet shows better consistency with the ground-truth and fewer misclassified pixels than competing methods. Overall, experiment results in Table 7 and Figure 10 show that the proposed SSCDnet is able to achieve promising cloud detection performance on Landsat-8 OLI data.
Table 7. Cloud extraction accuracy (%) of different comparison networks on Landsat-8 OLI data. All results are the averaged results on all the testing images.
Table 7. Cloud extraction accuracy (%) of different comparison networks on Landsat-8 OLI data. All results are the averaged results on all the testing images.
ProportionsMethodOAMIoUKappaPAUA
1 200 DeepLabV2 [51]91.2484.1681.7782.1294.53
DAN [37]92.6886.9782.8390.4788.78
Hung et al. [43]93.1487.0284.4991.5590.13
s4GAN [40]93.8888.7685.9193.3891.05
SSCDnet95.4891.2888.8795.6090.62
1 100 DeepLabV2 [51]92.5488.7685.0688.3089.41
DAN [37]93.9989.3285.7591.1189.45
Hung et al. [43]94.5890.7187.8993.8090.33
s4GAN [40]95.4091.3089.0294.1791.08
SSCDnet95.6391.7589.3893.8192.01
1 40 DeepLabV2 [51]93.1589.0685.6790.1689.90
DAN [37]94.3789.2686.7892.3889.77
Hung et al. [43]94.7990.9388.1193.2291.05
s4GAN [40]95.5491.3389.1993.1592.00
SSCDnet95.8292.0389.6193.1892.31
1 20 DeepLabV2 [51]94.1289.3386.4591.8590.36
DAN [37]95.2989.6186.9992.6790.14
Hung et al. [43]95.8391.0688.2092.1191.82
s4GAN [40]96.1691.9689.6992.0492.84
SSCDnet96.6792.1990.1392.5093.26
FullDeepLabV2 [51]96.5691.0389.2591.1091.26
DAN [37]96.6092.0989.5391.6293.11
Hung et al. [43]96.5792.1189.7792.2892.03
s4GAN [40]96.6392.0689.8292.1692.41
SSCDnet97.1992.6590.3492.7593.36

5. Discussion and Analysis

5.1. Robustness Analysis

To evaluate the robustness of the proposed method, we conducted a series of experiments as follows: (1) experiment results on the same area under different seasons and (2) experiment results on different land cover types.

5.1.1. Results on the Same Area under Different Seasons

In Figure 11, we present the cloud detection results on the same area under different seasons, i.e., Spring, Summer, Autumn, and Winter. Satellite images obtained from this area include different land cover types, such as mountain, village, urban, water, ocean, plant, and farmland areas. In Figure 11, overall accuracy (OA) of Spring, Summer, Autumn, and Winter images are 98.21%, 97.70%, 95.79%, and 96.03%, respectively. Experiment results show that the proposed semi-supervised cloud detection method SSCDnet is able to achieve a promising performance under different radiance (i.e., same area and different seasons). However, there is still a large number of misclassified pixels in the experiment results. These misclassified pixels are located mostly near the cloud object boundaries and thin cloud areas, which is also a difficult problem for most CNN-based cloud detection works, such as [7,19,27,33,34,35].

5.1.2. Results on Different Land Cover Types

In Figure 12, we present the experiment results on twelve sub images with different land cover types. Results show that these corresponding cloud detection results obtained by SSCDnet show consistency with the ground truth when our CNN model is trained with a limited number of labeled dataset (labeled proportion is 1 10 ), except for some misclassified pixels located near cloud object boundaries and thin cloud areas.
To be specific, for tough cases, such as barren/desert areas (Figure 12a,b) and urban areas (Figure 12g,h), SSCDnet is able to achieve a promising performance on these land cover types. In addition, SSCDnet also shows promising performance on water areas, such as lake (Figure 12i,j), river (Figure 12h), and ocean (Figure 12k,l) areas. Except for snow, buildings, and some white objects, few ground objects affect cloud detection, and SSCDnet can easily obtain promising results in some general cases, such as mountain/plant (Figure 12e,f) and farmland/village (Figure 12g,h) areas. In general, results in Figure 12 demonstrate that SSCDnet has a robust cloud detection performance on different land cover types.

5.2. Computational Complexity Analysis

To analyze the computational complexity of SSCDnet, we evaluate computational complexity of these networks with six evaluation criterions, which are floating point operations (FLOPs), number of trainable parameters, training time, training GPU memory usage, testing time, and testing GPU memory usage. In Table 8, we list results of different competing methods. FLOPs are calculated from input data with image size of 321 × 321. Training times of all methods are obtained from 5000 iterations. Training GPU memory size is obtained by setting batch size of 4 and image size of 321 × 321. Testing time is obtained by testing 20 scene Landsat-8 OLI satellite images with image size of 8 k × 8 k. Training GPU memory size is obtained by setting batch size of 1 and image size of 1200 × 1200.
During training, both the segmentation network and discriminator network are trained simultaneously. Discriminators of different SSL methods have different raw input data channels, which results in these models having different model parameters, different computational complexity and training times, and different GPU memory requirements. In Table 8, we can see that FLOPs, number of parameters, and training GUP memory usage of SSCDnet are higher than those of competing methods. This is because SSCDnet performs intermediate feature map domain adaptation alignments, while competing methods have no such operation. Feature map alignment requires more computations and GPU memories. In addition, the discriminator network for feature alignment further increases the number of training parameters. Luckily, the longest training time is not the SSCDnet but Hung et al. [43].
During testing, we only need to use the segmentation network to detect clouds instead of the discriminator network. Since all the methods use the same baseline segmentation network, i.e., DeepLabV2 [51], all these methods share the same testing times and GPU memory usage. In Table 8, we can see that it takes about 400 s to detect 20 scenes Landsat 8-OLI satellite images with image size of 8 k × 8 k. In other words, it takes 20 s to detect an image. 2849 MB GPU memory is required to process an image with size of 1200 × 1200.
Table 8. Computational complexity analysis of different semi-supervised methods and the baseline method (DeepLabV2 [51]).
Table 8. Computational complexity analysis of different semi-supervised methods and the baseline method (DeepLabV2 [51]).
MethodsGFLOPsParametersTraining TimeTraining GPUTesting TimeTesting GPU
DeepLabV2 [51]74.0343.94 M1118.9906 s10,079 MB400.7494 s2849 MB
DAN [37]76.6746.70 M2129.5402 s10,297 MB400.7494 s2849 MB
Hung et al. [43]76.6146.70 M10,178.0476 s10,297 MB400.7494 s2849 MB
s4GAN [40]76.6846.70 M2224.8972 s10,341 MB400.7494 s2849 MB
SSCDnet410.5661.91 M2771.0219 s10,435 MB400.7494 s2849 MB
The code for computational complexity is available from https://github.com/sovrasov/flops-counter.pytorch (accessed on 24 April 2022). 1 GFLOPs = 1 × 109 FLOPs. 1 M = 1 × 106. 1 MB = 1 × 106 bytes.

5.3. Limitations

Although SSCDnet achieves a promising cloud detection performance on both GF-1 WFV and Landsat-8 OLI data, there is still a large number of misclassified pixels in tough cases when using a limited number of labeled samples. Since urban and floating ice areas show the similar color or texture with the clouds, and thin cloud objects show few differences with underground objects, it is difficult for SSCDnet to handle these areas when using a limited number of labeled samples as shown in Figure 13, where Figure 13a,b are the results of GF-1 WFV data (GF1_WFV2_W70.8_N19.2_20140801_L2A0000292230), while Figure 13c,d are the results of Landsat-8 OLI dat (LC81180382014244LGN00).
Experiment results in Figure 13a,b show that there are many misclassified pixels at urban and floating ice areas when labeled sample proportion is 1 200 and 1 100 . When labeled sample proportion is more than 1 40 , the qualitative results remain basically stable and show consistency with the ground-truth. Results in Figure 13c,d indicate that capturing sharp and detailed object boundaries in thin cloud areas is still very difficult even trained with fully labeled samples. Adding a sufficient of thin cloud samples for network training may obtain good detection performance. In future work, we will focus on this point to further improve this work.

6. Conclusions

Semi-supervised learning is an effective training strategy, which is able to train a segmentation network by using a limited number of pixel-wise labeled samples and a large number of unlabeled ones. In this paper, we present a semi-supervised cloud detection network, named SSCDnet. Since there are domain distribution gaps between the labeled and unlabeled datasets, we take the domain shift problem into account for the semi-supervised learning framework and propose feature-/output-level domain adaptation strategy to reduce domain distribution gaps, thus improving SSCDnet to generate trustworthy pseudo label for unlabeled data. A high certain pseudo label provides positive supervised signals for segmentation network learning through self-training. Experimental results on GF-1 WFV and Landsat-8 OLI datasets demonstrate that SSCDnet is able to achieve promising performance by using a limited number of labeled samples. It shows great promise for practical application on new satellite RS imagery in the presence of less labeled data available.
Although SSCDnet shows good performance, there is still much room for improvement, such as hyper-parameters setting of loss function and threshold setting of pseudo-labeling. Different cloud detection datasets have different domain distributions. We need to update these parameters to achieve a promising performance on different datasets. In addition, different ground objects have different characteristics, and the performance of SSCDnet on other objects detection also needs to be further evaluated. In our future work, we will further evaluate this method on other cloud detection datasets and other object detection tasks. In addition, SSCDnet performs poorly on cloud boundaries and thin cloud regions, which requires our future efforts to improve it. In our future work, we will also explore how to utilize some auxiliary information, such as land use and land cover (LULC) map, water index map, and vegetation index map, to improve the cloud detection performance.
In general, a semi-supervised learning training strategy provides us with an effective way for cloud detection from RS images in the presence of less labeled data available. In addition, this strategy may provide us a promising way for other object detection tasks such as water, vegetation, and building detection.
In order to promote understanding of the paper’s technology, we released the code of SSCDnet. It is available at: https://github.com/nkszjx/SSCDnet (accessed on 24 April 2022).

Author Contributions

Conceptualization, J.G. and Z.L.; methodology, J.G.; validation, J.G. and Y.Z.; formal analysis, Z.L.; investigation, Q.X.; data curation, J.G.; writing—original draft preparation, J.G.; writing—review and editing, J.G. and Q.X.; supervision, X.Z.; project administration, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Sino-German (CSC-DAAD) Postdoc Scholarship Program under Grant 202006255045, in part by the German Federal Ministry of Education and Research (BMBF) in the framework of the international future AI Laboratory “AI4EO-Artificial Intelligence for Earth Observation: Reasoning, Uncertainties, Ethics and Beyond” under Grant 01DD20001, in part by the German Federal Ministry of Economics and Technology in the framework of the “National Center of excellence ML4Earth” under Grant 50EE2201C, in part by the European Research Council (ERC) through the European Union’s Horizon 2020 Research and Innovation Programme under Grant ERC-2016-StG-714087, in part by the Helmholtz Association through the Framework of Helmholtz AI under Grant ZT-I-PF-5-01-Local Unit “Munich Unit at Aeronautics, Space and Transport (MASTr)”, and in part by the Helmholtz Excellent Professorship “Data Science in Earth Observation Big Data Fusion for Urban Research” under Grant W2-W3-100.

Data Availability Statement

Not applicable.

Acknowledgments

We thank Huanfeng Shen’s team from Wuhan University, China, for their providing a Gaofen-1 WFV cloud cover validation dataset. We are also thankful for Earth Explorer of the United States Geological Survey for providing the Landsat-8 OLI cloud cover validation dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, S.; Wang, Q.; Li, Y.; Liu, S.; Wang, Z.; Zhu, L.; Wang, Z. An overview of satellite remote sensing technology used in China’s environmental protection. Earth Sci. Inform. 2017, 10, 137–148. [Google Scholar] [CrossRef]
  2. Yang, J.; Gong, P.; Fu, R.; Zhang, M.; Chen, J.; Liang, S.; Xu, B.; Shi, J.; Dickinson, R. The role of satellite remote sensing in climate change studies. Nat. Clim. Chang. 2013, 3, 875–883. [Google Scholar] [CrossRef]
  3. Schmugge, T.J.; Kustas, W.P.; Ritchie, J.C.; Jackson, T.J.; Rango, A. Remote sensing in hydrology. Adv. Water Resour. 2002, 25, 1367–1385. [Google Scholar] [CrossRef]
  4. Shanmugapriya, P.; Rathika, S.; Ramesh, T.; Janaki, P. Applications of remote sensing in agriculture—A Review. Int. J. Curr. Microbiol. Appl. Sci. 2019, 8, 2270–2283. [Google Scholar] [CrossRef]
  5. Kabisch, N.; Selsam, P.; Kirsten, T.; Lausch, A.; Bumberger, J. A multi-sensor and multi-temporal remote sensing approach to detect land cover change dynamics in heterogeneous urban landscapes. Ecol. Indic. 2019, 99, 273–282. [Google Scholar] [CrossRef]
  6. Bachagha, N.; Wang, X.; Luo, L.; Li, L.; Khatteli, H.; Lasaponara, R. Remote sensing and GIS techniques for reconstructing the military fort system on the Roman boundary (Tunisian section) and identifying archaeological sites. Remote Sens. Environ. 2020, 236, 111418. [Google Scholar] [CrossRef]
  7. Guo, J.; Yang, J.; Yue, H.; Tan, H.; Hou, C.; Li, K. CDnetV2: CNN-Based Cloud Detection for Remote Sensing Imagery with Cloud-Snow Coexistence. IEEE Trans. Geosci. Remote Sens. 2020, 59, 700–713. [Google Scholar] [CrossRef]
  8. Sliwa, B.; Falkenberg, R.; Liebig, T.; Piatkowski, N.; Wietfeld, C. Boosting vehicle-to-cloud communication by machine learning-enabled context prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3497–3512. [Google Scholar] [CrossRef] [Green Version]
  9. Zhu, Z.; Wang, S.; Woodcock, C.E. Improvement and expansion of the Fmask algorithm: Cloud, cloud shadow, and snow detection for Landsats 4–7, 8, and Sentinel 2 images. Remote Sens. Environ. 2015, 159, 269–277. [Google Scholar] [CrossRef]
  10. Qiu, S.; He, B.; Zhu, Z.; Liao, Z.; Quan, X. Improving Fmask cloud and cloud shadow detection in mountainous area for Landsats-8 images. Remote Sens. Environ. 2017, 199, 107–119. [Google Scholar] [CrossRef]
  11. Qiu, S.; Zhu, Z.; He, B. Fmask 4.0: Improved cloud and cloud shadow detection in Landsats 4–8 and Sentinel-2 imagery. Remote Sens. Environ. 2019, 231, 111205. [Google Scholar] [CrossRef]
  12. Li, Z.; Shen, H.; Li, H.; Xia, G.; Gamba, P.; Zhang, L. Multi-feature combined cloud and cloud shadow detection in GaoFen-1 wide field of view imagery. Remote Sens. Environ. 2017, 191, 342–358. [Google Scholar] [CrossRef] [Green Version]
  13. Alonso-Montesinos, J.; Martínez-Durbán, M.; del Sagrado, J.; del Águila, I.; Batlles, F. The application of Bayesian network classifiers to cloud classification in satellite images. Renew. Energy 2016, 97, 155–161. [Google Scholar] [CrossRef]
  14. Xu, L.; Wong, A.; Clausi, D.A. A novel Bayesian spatial–temporal random field model applied to cloud detection from remotely sensed imagery. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4913–4924. [Google Scholar] [CrossRef]
  15. Latry, C.; Panem, C.; Dejean, P. Cloud detection with SVM technique. In Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, Barcelona, Spain, 23–27 July 2007; pp. 448–451. [Google Scholar]
  16. Ishida, H.; Oishi, Y.; Morita, K.; Moriwaki, K.; Nakajima, T.Y. Development of a support vector machine based cloud detection method for MODIS with the adjustability to various conditions. Remote Sens. Environ. 2018, 205, 390–407. [Google Scholar] [CrossRef]
  17. Hughes, M.J.; Hayes, D.J. Automated detection of cloud and cloud shadow in single-date Landsat imagery using neural networks and spatial post-processing. Remote Sens. 2014, 6, 4907–4926. [Google Scholar] [CrossRef] [Green Version]
  18. Jang, J.d.; Viau, A.A.; Anctil, F.; Bartholomé, E. Neural network application for cloud detection in SPOT VEGETATION images. Int. J. Remote Sens. 2006, 27, 719–736. [Google Scholar] [CrossRef]
  19. Jeppesen, J.H.; Jacobsen, R.H.; Inceoglu, F.; Toftegaard, T.S. A cloud detection algorithm for satellite imagery based on deep learning. Remote Sens. Environ. 2019, 229, 247–259. [Google Scholar] [CrossRef]
  20. Wieland, M.; Li, Y.; Martinis, S. Multi-sensor cloud and cloud shadow segmentation with a convolutional neural network. Remote Sens. Environ. 2019, 230, 111203. [Google Scholar] [CrossRef]
  21. Chai, D.; Newsam, S.; Zhang, H.K.; Qiu, Y.; Huang, J. Cloud and cloud shadow detection in Landsat imagery based on deep convolutional neural networks. Remote Sens. Environ. 2019, 225, 307–316. [Google Scholar] [CrossRef]
  22. Drönner, J.; Korfhage, N.; Egli, S.; Mühling, M.; Thies, B.; Bendix, J.; Freisleben, B.; Seeger, B. Fast cloud segmentation using convolutional neural networks. Remote Sens. 2018, 10, 1782. [Google Scholar] [CrossRef] [Green Version]
  23. Lopez, J.; Santos, S.; Atzberger, C.; Torres, D. Convolutional Neural Networks for Semantic Segmentation of Multispectral Remote Sensing Images. In Proceedings of the 2018 IEEE 10th Latin-American Conference on Communications (LATINCOM), Guadalajara, Mexico, 14–16 November 2018; pp. 1–5. [Google Scholar]
  24. Zhan, Y.; Wang, J.; Shi, J.; Cheng, G.; Yao, L.; Sun, W. Distinguishing cloud and snow in satellite images via deep convolutional network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1785–1789. [Google Scholar] [CrossRef]
  25. Morales, G.; Huamán, S.G.; Telles, J. Cloud Detection in High-Resolution Multispectral Satellite Imagery Using Deep Learning. In Proceedings of the International Conference on Artificial Neural Networks, Rhodes, Greece, 4–7 October 2018; pp. 280–288. [Google Scholar]
  26. Li, Z.; Shen, H.; Cheng, Q.; Liu, Y.; You, S.; He, Z. Deep learning based cloud detection for medium and high resolution remote sensing images of different sensors. ISPRS J. Photogramm. Remote Sens. 2019, 150, 197–212. [Google Scholar] [CrossRef] [Green Version]
  27. Yang, J.; Guo, J.; Yue, H.; Liu, Z.; Hu, H.; Li, K. CDnet: CNN-Based Cloud Detection for Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6195–6211. [Google Scholar] [CrossRef]
  28. Shao, Z.; Pan, Y.; Diao, C.; Cai, J. Cloud Detection in Remote Sensing Images Based on Multiscale Features-Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4062–4076. [Google Scholar] [CrossRef]
  29. Chen, Y.; Weng, Q.; Tang, L.; Liu, Q.; Fan, R. An Automatic Cloud Detection Neural Network for High-Resolution Remote Sensing Imagery With Cloud–Snow Coexistence. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6004205. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Iwasaki, A.; Xu, G.; Song, J. Cloud detection on small satellites based on lightweight U-net and image compression. J. Appl. Remote Sens. 2019, 13, 026502. [Google Scholar] [CrossRef]
  31. Zhang, J.; Li, X.; Li, L.; Sun, P.; Su, X.; Hu, T.; Chen, F. Lightweight U-Net for cloud detection of visible and thermal infrared remote sensing images. Opt. Quantum Electron. 2020, 52, 397. [Google Scholar] [CrossRef]
  32. Li, J.; Wu, Z.; Hu, Z.; Jian, C.; Luo, S.; Mou, L.; Zhu, X.X.; Molinier, M. A Lightweight Deep Learning-Based Cloud Detection Method for Sentinel-2A Imagery Fusing Multiscale Spectral and Spatial Features. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–19. [Google Scholar] [CrossRef]
  33. Guo, J.; Yang, J.; Yue, H.; Li, K. Unsupervised Domain Adaptation for Cloud Detection Based on Grouped Features Alignment and Entropy Minimization. IEEE Trans. Geosci. Remote Sens. 2021, 1–13. [Google Scholar] [CrossRef]
  34. Guo, J.; Yang, J.; Yue, H.; Liu, X.; Li, K. Unsupervised Domain-Invariant Feature Learning for Cloud Detection of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5405715. [Google Scholar] [CrossRef]
  35. Guo, J.; Yang, J.; Yue, H.; Chen, Y.; Hou, C.; Li, K. Cloud Detection From Remote Sensing Imagery Based on Domain Translation Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5000805. [Google Scholar] [CrossRef]
  36. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T. The Cityscapes Dataset for Semantic Urban Scene Understanding. arXiv 2016, arXiv:1604.01685. [Google Scholar]
  37. Zhang, Y.; Yang, L.; Chen, J.; Fredericksen, M.; Hughes, D.P.; Chen, D.Z. Deep Adversarial Networks for Biomedical Image Segmentation Utilizing Unannotated Images. In Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI 2017), Quebec City, QC, Canada, 11–13 September 2017; pp. 408–416. [Google Scholar]
  38. Hong, S.; Noh, H.; Han, B. Decoupled Deep Neural Network for Semi-supervised Semantic Segmentation. arXiv 2015, arXiv:1506.04924. [Google Scholar]
  39. Mostafa S, I.; Arash, V.; Mani, R.; William G, M. Semi-Supervised Semantic Image Segmentation with Self-correcting Networks. arXiv 2020, arXiv:1811.07073. [Google Scholar]
  40. Mittal, S.; Maxim, T.; Thomas, B. Semi-Supervised Semantic Segmentation with High- and Low-level Consistency. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1369–1379. [Google Scholar] [CrossRef] [Green Version]
  41. Massih-Reza, A.; Vasilii, F.; Loic, P.; Emilie, D.; Yury, M. Self-Training: A Survey. arXiv 2022, arXiv:2202.12040. [Google Scholar]
  42. Rajat, R.; Alexis, B.; Honglak, L.; Benjamin, P.; Andrew, Y.N. Self-taught learning: Transfer learning from unlabeled data. In Proceedings of the Twenty-fourth International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007. [Google Scholar]
  43. Hung, W.; Tsai, Y.; Liou, Y.; Lin, Y.Y.; Yang, M.H. Adversarial Learning for Semi-Supervised Semantic Segmentation. arXiv 2018, arXiv:1802.07934. [Google Scholar]
  44. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  45. Ren, Z.; Lee, Y.J. Cross-Domain Self-Supervised Multi-task Feature Learning Using Synthetic Imagery. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 762–771. [Google Scholar]
  46. Tuan-Hung, V.; Himalaya, J.; Maxime, B.; Matthieu, C.; Patrick, P. ADVENT: Adversarial Entropy Minimization for Domain Adaptation in Semantic Segmentation. arXiv 2019, arXiv:1811.12833. [Google Scholar]
  47. Tsai, Y.; Sohn, K.; Schulter, S.; Chandraker, M. Domain Adaptation for Structured Output via Discriminative Patch Representations. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1456–1465. [Google Scholar]
  48. Foga, S.; Scaramuzza, P.L.; Guo, S.; Zhu, Z.; Dilley, R.D., Jr.; Beckmann, T.; Schmidt, G.L.; Dwyer, J.L.; Hughes, M.J.; Laue, B. Cloud detection algorithm comparison and validation for operational Landsat data products. Remote Sens. Environ. 2017, 194, 379–390. [Google Scholar] [CrossRef] [Green Version]
  49. Chen, Y.; Lin, Y.; Yang, M.; Huang, J. CrDoCo: Pixel-Level Domain Transfer With Cross-Domain Consistency. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1791–1800. [Google Scholar]
  50. Souly, N.; Spampinato, C.; Shah, M. Semi Supervised Semantic Segmentation Using Generative Adversarial Network. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5689–5697. [Google Scholar]
  51. Chen, L.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  53. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2016, arXiv:1511.07122. [Google Scholar]
  54. Geirhos, R.; Rubisch, P.; Michaelis, C.; Bethge, M.; Wichmann, F.A.; Brendel, W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv 2018, arXiv:1811.12231. [Google Scholar]
  55. Bengio, Y. Practical recommendations for gradient-based training of deep architectures. In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 2012; pp. 437–478. [Google Scholar]
  56. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980v9. [Google Scholar]
  57. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  58. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  59. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. Available online: http://host.robots.ox.ac.uk/pascal/VOC/voc2012/ (accessed on 24 April 2022).
Figure 1. The traditional semi-supervised segmentation framework.
Figure 1. The traditional semi-supervised segmentation framework.
Remotesensing 14 02641 g001
Figure 2. Remote sensing images obtained from the same satellite sensor (GF-1 satellite) at different places. Where (a) GF1_WFV2_E127.2_N45.9_20140704_L2A00003099760, (b) GF1_WFV1_E102.0_N28.0_20140302_L2A00002804760, (c) GF1_WFV1_E99.4_N36.3_20140716_L2A00002789230, and (d) GF1_WFV3_E89.3_N35.6_20140702_L2A00008451540.
Figure 2. Remote sensing images obtained from the same satellite sensor (GF-1 satellite) at different places. Where (a) GF1_WFV2_E127.2_N45.9_20140704_L2A00003099760, (b) GF1_WFV1_E102.0_N28.0_20140302_L2A00002804760, (c) GF1_WFV1_E99.4_N36.3_20140716_L2A00002789230, and (d) GF1_WFV3_E89.3_N35.6_20140702_L2A00008451540.
Remotesensing 14 02641 g002
Figure 3. The detailed structure of the proposed SSCDnet.
Figure 3. The detailed structure of the proposed SSCDnet.
Remotesensing 14 02641 g003
Figure 4. Structure of the proposed class-relevant feature alignment module.
Figure 4. Structure of the proposed class-relevant feature alignment module.
Remotesensing 14 02641 g004
Figure 5. Structure of the proposed CROA.
Figure 5. Structure of the proposed CROA.
Remotesensing 14 02641 g005
Figure 6. Experiment results on an unlabeled remote sensing image. (a) input GF-1 WFV image; (b) un-applied domain adaptation result, and (c) applied domain adaptation result.
Figure 6. Experiment results on an unlabeled remote sensing image. (a) input GF-1 WFV image; (b) un-applied domain adaptation result, and (c) applied domain adaptation result.
Remotesensing 14 02641 g006
Figure 7. The double threshold pseudo-labelling method for self-training.
Figure 7. The double threshold pseudo-labelling method for self-training.
Remotesensing 14 02641 g007
Figure 8. Structure of the proposed discriminator network.
Figure 8. Structure of the proposed discriminator network.
Remotesensing 14 02641 g008
Figure 9. Comparison of cloud detection results of different methods on GF-1 WFV dataset with labeled sample proportion of 1/200. Image ID of (a), (b), and (c) are GF1_WFV2_W102.1_N37.6_20140517_L2A0000244678, GF1_WFV3_E114.1_N2.1_20151011_L2A0001094727, and GF1_WFV3_E87.8_N2.1_20140316_L2A0000184430, respectively.
Figure 9. Comparison of cloud detection results of different methods on GF-1 WFV dataset with labeled sample proportion of 1/200. Image ID of (a), (b), and (c) are GF1_WFV2_W102.1_N37.6_20140517_L2A0000244678, GF1_WFV3_E114.1_N2.1_20151011_L2A0001094727, and GF1_WFV3_E87.8_N2.1_20140316_L2A0000184430, respectively.
Remotesensing 14 02641 g009
Figure 10. Comparison of cloud extraction results of different methods on Landsat-8 OLI dataset with a labeled sample proportion of 1/200. The image ID of (a), (b), and (c) are LC80650182013237LGN00, LC80430122014214LGN00, and LC81990402014267LGN00, respectively.
Figure 10. Comparison of cloud extraction results of different methods on Landsat-8 OLI dataset with a labeled sample proportion of 1/200. The image ID of (a), (b), and (c) are LC80650182013237LGN00, LC80430122014214LGN00, and LC81990402014267LGN00, respectively.
Remotesensing 14 02641 g010
Figure 11. Cloud extraction results of SSCDnet on Landsat-8 OLI data under different seasons (labeled proportion is 1 20 ). Where, (a) is the Spring season (ID: LC08_L1TP_122033_20160411_20170326_01_T1), (b) is the Summer season (ID: LC08_L1TP_122033_20180706_20180717_01_T1), (c) is the Autumn season(ID: LC08_L1TP_122033_20181111_20181127_01_T1), and (d) is the Winter season (ID: LC08_L1TP_122033_20150103_20170415_01_T1).
Figure 11. Cloud extraction results of SSCDnet on Landsat-8 OLI data under different seasons (labeled proportion is 1 20 ). Where, (a) is the Spring season (ID: LC08_L1TP_122033_20160411_20170326_01_T1), (b) is the Summer season (ID: LC08_L1TP_122033_20180706_20180717_01_T1), (c) is the Autumn season(ID: LC08_L1TP_122033_20181111_20181127_01_T1), and (d) is the Winter season (ID: LC08_L1TP_122033_20150103_20170415_01_T1).
Remotesensing 14 02641 g011
Figure 12. Cloud extraction results (1200 × 1200) of SSCDnet on Landsat-8 OLI data under different land cover types. Where, (a,b) are barren/desert areas. (c,d) are mountain/plant areas. (e,f) are farmland/villages areas. (g,h) are urban/river areas. (i,j) are mountain/lake areas. (k,l) are ocean areas.
Figure 12. Cloud extraction results (1200 × 1200) of SSCDnet on Landsat-8 OLI data under different land cover types. Where, (a,b) are barren/desert areas. (c,d) are mountain/plant areas. (e,f) are farmland/villages areas. (g,h) are urban/river areas. (i,j) are mountain/lake areas. (k,l) are ocean areas.
Remotesensing 14 02641 g012
Figure 13. Cloud extraction results (1200 × 1200) of SSCDnet on GF-1 WFV and Landsat-8 OLI data under different labeled sample proportions. (a,b) are the results of GF-1 WFV data (GF1_WFV2_W70.8_N19.2_20140801_L2A0000292230), while (c,d) are the results of Landsat-8 OLI data (LC81180382014244LGN00).
Figure 13. Cloud extraction results (1200 × 1200) of SSCDnet on GF-1 WFV and Landsat-8 OLI data under different labeled sample proportions. (a,b) are the results of GF-1 WFV data (GF1_WFV2_W70.8_N19.2_20140801_L2A0000292230), while (c,d) are the results of Landsat-8 OLI data (LC81180382014244LGN00).
Remotesensing 14 02641 g013
Table 1. The detailed information of Landsat-8 OLI and GF1-WFV multispectral images.
Table 1. The detailed information of Landsat-8 OLI and GF1-WFV multispectral images.
SensorSpectral BandWavelength Range ( μ m)Spatial Resolution (m)
Landsat-8 OLIBand 1 (coastal)0.433–0.45330
Band 2 (blue)0.450–0.51530
Band 3 (green)0.525–0.60030
Band 4 (red)0.630–0.68030
Band 5 (NIR)0.845–0.88530
Band 6 (SWIR1)1.560–1.66030
Band 7 (SWIR2)2.100–2.30030
Band 8 (Pan)0.500–0.68015
Band 9 (Cirrus)1.360–1.39030
GF-1 WFVBand 1 (blue)0.450–0.52016
Band 2 (green)0.520–0.59016
Band 3 (red)0.630–0.69016
Band 4 (NIR)0.770–0.89016
Table 2. Cloud extraction accuracy (%) of different ablation networks on GF-1 WFV data.
Table 2. Cloud extraction accuracy (%) of different ablation networks on GF-1 WFV data.
Proportions L c e
L s t ×
L o d a ××
L f d a ×××
1 200 OA90.2293.5695.3095.51
MIoU71.7783.2286.9587.54
Kappa68.2178.1582.0782.69
PA67.0373.6780.0281.28
UA95.5196.4094.5793.55
1 100 OA91.3493.8995.7395.85
MIoU74.1084.1188.0588.90
Kappa71.5178.9083.3483.92
PA69.7174.3281.5282.34
UA94.9395.9994.2893.36
1 40 OA92.2394.5296.0296.24
MIoU76.0286.0389.0189.30
Kappa73.5880.8584.7785.01
PA75.0880.1884.4085.49
UA93.4393.3091.5992.24
1 20 OA93.1995.1196.5596.90
MIoU79.8887.1690.4690.93
Kappa75.7583.7786.2586.97
PA80.6982.2284.8785.89
UA92.3092.1092.7292.24
FullOA96.9197.1197.1797.19
MIoU90.1591.5091.6791.71
Kappa88.6489.4989.3589.37
PA87.2888.6788.8488.90
UA92.0592.3392.5192.32
where 1 200 , 1 100 , 1 40 , and 1 20 are the fractions of the total training images in the dataset that are used as labeled data, and the rest of the data was used without labels. Bold indicates maximum value in this paper.
Table 3. Ablation study of domain adaptation on different intermediate layers (proportion = 1 200 ).
Table 3. Ablation study of domain adaptation on different intermediate layers (proportion = 1 200 ).
MethodsOAMIoUKappaPAUA
FDA_595.4087.1182.5380.2894.68
FDA_4595.5187.5482.6981.2893.55
FDA_34595.1386.1781.6479.1994.89
FDA_234595.4187.0682.9381.7393.54
Table 4. Ablation study on double threshold (proportion = 1 200 ).
Table 4. Ablation study on double threshold (proportion = 1 200 ).
MethodsOAMIoUKappaPAUA
L s t ( N / A , N / A ) 94.1285.0281.4379.3894.74
L s t ( τ 1 , N / A ) 95.0286.3381.9580.1193.30
L s t ( τ 1 , τ 2 ) 95.3086.9582.0780.0294.57
Table 5. The study on hyper-parameter λ s t , λ o d a , and λ f d a (proportion = 1 200 ).
Table 5. The study on hyper-parameter λ s t , λ o d a , and λ f d a (proportion = 1 200 ).
λ st λ oda λ fda OAMIoUKappaPAUA
1.00.05N/A93.0583.6680.8077.9094.51
1.00.10N/A93.4884.7481.5179.0293.61
1.00.15N/A93.1284.1081.1878.6094.19
1.00.100.1092.2681.1272.6168.5394.79
1.00.100.0194.9685.8681.4380.0693.48
1.00.100.00195.5187.5482.6981.2893.55
1.00.100.00595.4087.1182.6880.3594.86
Table 6. Cloud extraction accuracy (%) of different comparison networks on GF-1 WFV data. All results are the averaging results on all testing images.
Table 6. Cloud extraction accuracy (%) of different comparison networks on GF-1 WFV data. All results are the averaging results on all testing images.
ProportionMethodsOAMIoUKappaPAUA
1 200 DeepLabV2 [51]90.2271.7768.2167.0395.51
DAN [37]92.5979.9374.6572.0894.59
Hung et al. [43]93.0181.2075.1176.5994.02
s4GAN [40]93.4481.7975.6277.3193.38
SSCDnet95.5187.5482.6981.2893.55
1 100 DeepLabV2 [51]91.3474.1071.5169.7194.93
DAN [37]93.1382.4177.5675.8493.66
Hung et al. [43]94.2083.8878.5676.1193.02
s4GAN [40]94.6084.9179.1476.4792.63
SSCDnet95.8588.9083.9282.3493.36
1 40 DeepLabV2 [51]92.2376.0273.5875.0893.43
DAN [37]93.8983.6380.1581.2992.53
Hung et al. [43]94.4185.0581.5981.8092.76
s4GAN [40]95.3086.8082.4782.2093.00
SSCDnet96.2489.3085.0185.4992.24
1 20 DeepLabV2 [51]93.1979.8875.7580.6992.30
DAN [37]94.3285.0581.5282.2192.93
Hung et al. [43]94.9486.1282.5382.4493.09
s4GAN [40]95.6187.1883.6382.9693.37
SSCDnet96.9090.9386.9785.8992.24
FullDeepLabV2 [51]96.1190.1587.6487.2892.15
DAN [37]97.0990.3788.7288.1992.43
Hung et al. [43]97.1290.5788.7988.5092.71
s4GAN [40]97.1690.8888.8388.5792.60
SSCDnet97.1991.7189.3788.9092.32
where 1 200 , 1 100 , 1 40 , and 1 20 are the fractions of the total training images in the dataset that are used as labeled data, and the rest of the data are used without labels.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, J.; Xu, Q.; Zeng, Y.; Liu, Z.; Zhu, X. Semi-Supervised Cloud Detection in Satellite Images by Considering the Domain Shift Problem. Remote Sens. 2022, 14, 2641. https://doi.org/10.3390/rs14112641

AMA Style

Guo J, Xu Q, Zeng Y, Liu Z, Zhu X. Semi-Supervised Cloud Detection in Satellite Images by Considering the Domain Shift Problem. Remote Sensing. 2022; 14(11):2641. https://doi.org/10.3390/rs14112641

Chicago/Turabian Style

Guo, Jianhua, Qingsong Xu, Yue Zeng, Zhiheng Liu, and Xiaoxiang Zhu. 2022. "Semi-Supervised Cloud Detection in Satellite Images by Considering the Domain Shift Problem" Remote Sensing 14, no. 11: 2641. https://doi.org/10.3390/rs14112641

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop