Next Article in Journal
Forest Age Mapping Using Landsat Time-Series Stacks Data Based on Forest Disturbance and Empirical Relationships between Age and Height
Next Article in Special Issue
MFTSC: A Semantically Constrained Method for Urban Building Height Estimation Using Multiple Source Images
Previous Article in Journal
Analysis of East Asia Wind Vectors Using Space–Time Cross-Covariance Models
Previous Article in Special Issue
Improvements in Forest Segmentation Accuracy Using a New Deep Learning Architecture and Data Augmentation Technique
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cloud Removal in Remote Sensing Using Sequential-Based Diffusion Models

Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(11), 2861; https://doi.org/10.3390/rs15112861
Submission received: 11 April 2023 / Revised: 24 May 2023 / Accepted: 26 May 2023 / Published: 31 May 2023

Abstract

:
The majority of the optical observations collected via spaceborne optical satellites are corrupted by clouds or haze, restraining further applications of Earth observation; thus, exploring an ideal method for cloud removal is of great concern. In this paper, we propose a novel probabilistic generative model named sequential-based diffusion models (SeqDMs) for the cloud-removal task in a remote sensing domain. The proposed method consists of multi-modal diffusion models (MmDMs) and a sequential-based training and inference strategy (SeqTIS). In particular, MmDMs is a novel diffusion model that reconstructs the reverse process of denosing diffusion probabilistic models (DDPMs) to integrate additional information from auxiliary modalities (e.g., synthetic aperture radar robust to the corruption of clouds) to help the distribution learning of main modality (i.e., optical satellite imagery). In order to consider the information across time, SeqTIS is designed to integrate temporal information across an arbitrary length of both the main modality and auxiliary modality input sequences without retraining the model again. With the help of MmDMs and SeqTIS, SeqDMs have the flexibility to handle an arbitrary length of input sequences, producing significant improvements only with one or two additional input samples and greatly reducing the time cost of model retraining. We evaluate our method on a public real-world dataset SEN12MS-CR-TS for a multi-modal and multi-temporal cloud-removal task. Our extensive experiments and ablation studies demonstrate the superiority of the proposed method on the quality of the reconstructed samples and the flexibility to handle arbitrary length sequences over multiple state-of-the-art cloud removal approaches.

Graphical Abstract

1. Introduction

In recent decades, massive remote sensing data have been collected by Earth-observing satellites, and such data have started to play an important role in a variety of tasks, including environmental monitoring [1], economic development mapping [2], land cover classification [3], and agricultural monitoring [4,5]. However, remote sensing images are often blocked by haze or clouds [6], which impedes the data processing and analysis for target monitoring tasks. Therefore, it is valuable and pivotal to explore the approaches for reconstructing the data corrupted by clouds for subsequent data analysis and employment.
In general, cloud removal can be seen as a special type of inpainting task that fills the missing areas of remote sensing data corrupted by clouds with new and suitable content. Prior approaches to cloud removal can be classified into two main types according to the source of information used for the reconstruction: multi-modal approaches and multi-temporal approaches [7]. In order to expand the information sources, multi-modal approaches [8,9,10,11,12,13,14] have been developed to reconstruct cloud-covered pixels via information translated from synthetic aperture radar (SAR) data or other modal data more robust to the corruption of clouds [15]. Traditional multi-modal approaches [8,9] utilize the digital number of SAR as an indicator to find the repair pixel. Eckardt et al. [9] introduce the term closest feature vector (CFV), combining the closest spectral fit (CSF) algorithm [16] with the synergistic application of multi-spectral satellite images and multi-frequency SAR data. With the wide application of deep learning and the rapid development of generative models, Gao et al. [14] first translate the SAR images into simulated optical images in an object-to-object manner by a specially designed convolutional neural network (CNN) and then fuse the simulated optical images together with the SAR images and the cloudy optical images by a generative adversarial network (GAN) to reconstruct the corrupted area. In contrast to methods using a single time point of observations, multi-temporal approaches [17,18,19,20,21,22] attempt a temporal reconstruction of cloudy observations by means of inference across time series, utilizing the information from other cloud-free time point as a reference, based on the fact that the extent of cloud coverage over a particular region is variable over time and seasons [6]. Traditional multi-temporal approaches [18,19,20] employ hand-crafted filters such as mean and median filters to generate the cloud-covered parts using a large number of images over a specific area. For instance, Ramoino et al. [20] conduct cloud removal using plenty of Sentinel-2 images taken every 6–7 days across a time period of three months. In terms of approaches that utilize deep learning techniques, Sarukkai et al. [17] propose a novel spatiotemporal generator network (STGAN) to better capture correlations across multiple images over an area, leveraging multi-spectral information (i.e., RGB and IR bands of Sentinel-2) to generate a cloud-free image. However, these image reconstruction approaches do not leverage multi-modal information and require a large number of mostly cloud-free images taken over an unchanging landscape, greatly limiting their usability and applications.
Meanwhile, much of the early work on cloud removal used datasets containing simulated cloudy observations, copying cloudy pixel values from one image to another clear-free one [23], but could not precisely reproduce the statistic of satellite images containing natural cloud occurrences [12]. Recently, Ebel et al. [7] curated a new real-world dataset called SEN12MS-CR-TS, which contains both multi-temporal and multi-modal globally distributed satellite observations. They also proposed a sequence-to-point cloud removal method based on 3-D CNN (we denote it as Seq2point) to integrate information across time and within different modalities. However, this method lacks a probabilistic interpretation and the flexibility to handle input sequences of arbitrary length. It just uses the ResNet-based [24] branch and a 3-D CNN structure as a generator to combine the feature maps in the time span and needs to be retrained in a great amount of time when the length of the input sequence changes.
Overall, existing approaches have at least one of three major shortages: (1) They do not use globally distributed real-world datasets, leading to the degraded generalizability of the methods. (2) They are not designed to fully leverage both multi-modal and multi-temporal information to reconstruct the corrupted regions. (3) They lack a probabilistic interpretation and flexibility to handle an arbitrary length of the input sequences.
In this paper, we propose a novel method, sequential-based diffusion models (SeqDMs), for the cloud-removal task in remote sensing by integrating information across time and within different modalities. As GANs are known to suffer from the instability training process [25], we choose the denoising diffusion probabilistic models (DDPMs) [26] with a better probabilistic interpretation and a more powerful capability of capturing the data distribution as the backbone model. In particular, we propose novel diffusion models, multi-modal diffusion models (MmDMs), which reconstruct the reverse process of DDPMs to integrate additional information from the auxiliary modalities (e.g., SAR or other modalities robust to the corruption of clouds) to help the distribution learning of main modality (i.e., spaceborne optical satellites data). Since the standard DDPMs training and inference strategy processes the samples only in a single time point, we introduce an improved training and inference strategy named sequential-based training and inference strategy (SeqTIS) to integrate information across time from both main modality and auxiliary modalities input sequences. It is worth noting that SeqDMs have the flexibility to handle an arbitrary length of the input sequences without retraining the model, which significantly reduces the training time cost. We conduct adequate experiments and ablation studies on a globally distributed SEN12MS-CR-TS dataset [7] to evaluate our method and justify its design. We also compare with other state-of-the-art cloud removal approaches to show the superiority of the proposed method.

2. Preliminaries: Denoising Diffusion Probabilistic Models

Our proposed method for cloud removal is based on DDPMs [26]. Here, we first introduce the definition and properties of this type of generative model. The DDPMs define a Markov chain of a diffusion process controlled by a variance schedule { β t ( 0 , 1 ) } t = 1 T to transform the input sample x 0 to white Gaussian noise x T N ( 0 , I ) in T diffusion time steps. The symbol t represents the diffusion time steps in the diffusion process of DDPMs. We use the symbol l to represent the sample index in sequential data later. In order to distinguish ’time step’ in the diffusion process of the model and in sequential data, we describe it as ’diffusion time step’ and ’sequential time step’, separately. Each step in the diffusion process is given by:
q ( x t | x t 1 ) = N ( x t ; 1 β t x t 1 , β t I ) .
The sample x t is obtained by slowly adding i.i.d. Gaussian random noise with variance β t at diffusion time step t and scaling the previous sample x t 1 with 1 β t according to the variance schedule. A notable property of the diffusion process is that it admits sampling x t at an arbitrary diffusion time step t from the input x 0 in a closed form as
q ( x t | x 0 ) = N ( x t ; α ¯ t x 0 , ( 1 α ¯ t ) I ) ,
where α t : = 1 β t and α ¯ t : = s = 1 t α s .
The inference process (the generation direction) works by sampling a random noise vector x T and then gradually denoising it until reaches a high-quality output sample x 0 . To implement the inference process, the DDPMs are trained to reverse the process in Equation (1). The reverse process is modeled by a neural network that predicts the parameters μ θ ( x t , t ) and Σ θ ( x t , t ) of a Gaussian distribution:
p θ ( x t 1 | x t ) = N ( x t 1 ;   μ θ ( x t , t ) , Σ θ ( x t , t ) ) .
The learning objective for the model in Equation (3) is derived by considering the variational lower bound,
E [ log p θ ( x 0 ) ] E q [ log p θ ( x 0 : T ) q ( x 1 : T | x 0 ) ] = E q [ log p ( x T ) t 1 log p θ ( x t 1 | x t ) q ( x t | x t 1 ) ] = L V L B ,
and this objective function can be further decomposed as:
L V L B = E q [ D K L ( q ( x T | x 0 ) | | p ( x T ) ) L T + t > 1 D K L ( q ( x t 1 | x t , x 0 ) | | p θ ( x t 1 | x t ) ) L t 1 log p θ ( x 0 | x 1 ) L 0 ] .
It is noteworthy that the term L t 1 trains the network in Equation (3) to perform one reverse diffusion step. Furthermore, the posterior q ( x t 1 | x t ) is tractable when conditioned on x 0 , and it allows for a closed form expression of the objective since q ( x t 1 | x t , x 0 ) is also a Gaussian distribution:
q ( x t 1 | x t , x 0 ) = N ( x t 1 ;   μ ˜ t ( x t , x 0 ) , β ˜ t I ) ,
where
μ ˜ t ( x t , x 0 ) = α ¯ t 1 β t 1 α ¯ t x 0 + α t ( 1 α ¯ t 1 ) 1 α ¯ t x t
and
β ˜ t = 1 α ¯ t 1 1 α ¯ t β t .
Instead of directly predicting μ ˜ t , a better way is to parametrize the model by predicting the cumulative noise ε that is added to the current intermediate sample x t :
μ θ ( x t , t ) = 1 α t ( x t β t 1 α t ¯ ε θ ( x t , t ) ) ,
then, the following simplified training objective is derived from the term L t 1 in Equation (5):
L s i m p l e = E t , x 0 , ε [ | | ε ε θ ( α ¯ t x 0 + 1 α ¯ t ε , t ) | | 2 ] .
As introduced by Nichol and Dhariwal [27], learning variance Σ θ ( x t , t ) in Equation (3) of the reverse process helps to improve the log-likelihood and reduce the number of sampling steps. Since L s i m p l e does not depend on Σ θ ( x t , t ) , they define a new hybrid objective:
L h y b r i d = L s i m p l e + λ L V L B .
Furthermore, they make DDPMs achieve better sample quality than the current state-of-the-art generative models by finding a better architecture through a series of ablations [28]. Hence, we base our proposed method on DDPMs.

3. Materials and Methods

In this section, we describe in detail the proposed method sequential-based diffusion models (SeqDMs) for cloud removal in remote sensing, which consists of two components. In Section 3.1, we first present novel diffusion models, multi-modal diffusion models (MmDMs), which leverage a sequence of auxiliary modal data as additional information to learn the distribution of main modality. In Section 3.2, we introduce an improved training and inference strategy named sequential-based training and inference strategy (SeqTIS) for cloud removal to integrate information across time from both the main modality (i.e., optical satellite data) and auxiliary modalities (e.g., SAR or other modalities more robust to the corruption of clouds) input sequences.

3.1. Multi-Modal Diffusion Models

Unlike prior diffusion models [26,27,28], multi-modal diffusion models (MmDMs) leverage auxiliary modalities data as additional inputs to learn the distribution of the main modality, which will complement the partial missing information of the main modality during the inference process (cloud-removal process). The graphical model for MmDMs is shown in Figure 1.
We denote a sequence of multi-modal input data as { X , A 1 , . . . , A n , . . . , A N } , which consists of one main modality X, as well as N auxiliary modalities A. Since optical satellite data are susceptible to haze or clouds and SAR or other modalities are more robust against these influences [6,15], we consider optical satellite data as the main modality X and SAR or other modalities as auxiliary modalities A in this paper. The most important feature of MmDMs is the powerful ability to capture the distribution of X, and the ability to complement the missing information of X with A during the inference process, leading to a better performance of cloud removal in remote sensing.
The diffusion process of MmDMs is similar to that of DDPMs; it involves individually transforming each modal input sample to white Gaussian noise according to the same variance schedule { β t ( 0 , 1 ) } t = 1 T in T diffusion time steps. Each diffusion step of the main-modality sample X 0 is given by:
q ( X 1 : T | X 0 ) : = t = 1 T q ( X t | X t 1 ) , q ( X t | X t 1 ) = N ( X t ; 1 β t X t 1 , β t I ) ,
and the diffusion step for the nth auxiliary modal sample A 0 n is given by:
q ( A 1 : T n | A 0 n ) : = t = 1 T q ( A t n | A t 1 n ) , q ( A t n | A t 1 n ) = N ( A t n ; 1 β t A t 1 n , β t I ) .
Instead of directly learning a neural network p θ ( X t 1 | X t ) to approximate the intractable posterior distribution q ( X t 1 | X t ) described in Equation (5) of DDPMs, MmDMs add the auxiliary modalities as conditions to the reverse process:
p θ ( X 0 : T | A 1 : T 1 : N ) : = p ( X T ) t = 1 T p θ ( X t 1 | X t , A t 1 : N ) ,
p θ ( X t 1 | X t , A t 1 : N ) = N ( X t 1 ;   μ θ ( X t , A t 1 : N , t ) , Σ θ ( X t , A t 1 : N , t ) ) .
Then, the training objective for the model in Equation (15) is derived by the variational lower bound on negative log likelihood:
E [ log p θ ( X 0 ) ] E q [ log p θ ( X 0 : T | A 1 : T 1 : N ) q ( X 1 : T | X 0 ) ] = E q [ log p ( X T ) t 1 log p θ ( X t 1 | X t , A t 1 : N ) q ( X t | X t 1 ) ] = L V L B ,
and it can be further rewritten as:
L V L B = E q [ D K L ( q ( X T | X 0 ) | | p ( X T ) ) L T + t > 1 D K L ( q ( X t 1 | X t , X 0 ) | | p θ ( X t 1 | X t , A t 1 : N ) ) L t 1 log p θ ( X 0 | X 1 , A 1 1 : N ) L 0 ] .
Based on Equations (6)–(10) and the property of the diffusion process, we can derive a new version of the simplified training objective from the term L t 1 in Equation (17):
L s i m p l e ( θ ) = E t , X 0 , ε x , A 0 1 : N , ε a 1 : N [ | | ε x ε θ ( α ¯ t X 0 + 1 α ¯ t ε x , α ¯ t A 0 1 : N + 1 α ¯ t ε a 1 : N , t ) | | 2 ] ,
where ε x is the cumulative noise added to the main modal sample X 0 and ε a 1 : N are the noises individually added to N auxiliary modal samples A 0 1 : N .

3.2. Sequential-Based Training and Inference Strategy

To better reconstruct the missing information corrupted by clouds or haze, we introduce sequential-based training and inference strategy (SeqTIS) to integrate the information across time from both main modality and auxiliary modalities. SeqTIS contains a temporal training strategy and a conditional inference strategy, both of which use a reusable module called sequential data fusion module. For ease of description, we extend the multi-modal input data { X , A 1 , . . . , A n , . . . , A N } in a prior section into a multi-modal and multi-temporal version { X L , A 1 _ L , . . . , A n _ L , . . . , A N _ L } , where X L = { x 1 , . . . , x l . . . , x L } and A n _ L = { a n _ 1 , . . . , a n _ l , . . . , a n _ L } are the time series of length L. Corresponding to the multi-modal and multi-temporal input data, we also have a ground truth sequence { X ^ , A ^ 1 , . . . , A ^ n , . . . A ^ N } , which contains cloud-free main modality X ^ . All of the modules and processes in SeqTIS are described below.

3.2.1. Sequential Data Fusion Module

As shown in Figure 2, the sequential data fusion modules are used to integrate the information across time in each modality and are designed separately for either main modality or auxiliary modalities.
Since the auxiliary modalities A 1 : N are not influenced by clouds or haze, the sequential data fusion module for auxiliary modality is simply designed to individually diffuse data in each sequential time step l into a certain diffusion time step t and then calculate an average weighted value of that diffusion time step to integrate information across time. The nth auxiliary modality sequence A n _ L is processed by the sequential data fusion module for auxiliary modality as follows:
a t n _ l N ( α ¯ t a 0 n _ l , ( 1 α ¯ t ) I ) = α ¯ t a 0 n _ l + ( 1 α ¯ t ) ε a n _ l , l = 1 , 2 , . . . , L , n = 1 , 2 , . . . , N ,
where ε a n _ l N ( 0 , I ) is a cumulative noise added to the current intermediate sample a t n _ l . After diffusing the data in each sequential time step l, we calculate the average weighted value of diffusion time step t as follows:
A ˜ t n = l = 1 L a t n _ l L , n = 1 , 2 , . . . , N ,
which integrates the information of sequence A n _ L across time.
Since the main modality X L = { x 1 , . . . , x l . . . , x L } is susceptible to missing information due to clouds or haze, the sequential data fusion module for the main modality is designed to maintain the known regions’ (cloud-free pixels) information of each sequential time step l as much as possible, which is quite different from that for auxiliary modalities. In order to model the spatial-temporal extent of clouds, the binary cloud masks M L = { m 1 , . . . , m l . . . , m L } are computed on-the-fly for each main modality data in X L via the cloud detector of s2cloudless [29]. The pixel value 1 of m l indicates a cloud-free pixel, and value 0 indicates a cloudy pixel. The main modality sequence X L is processed by the sequential data fusion module for the main modality to the diffusion time step t 1 as follows:
x t 1 l N ( α ¯ t 1 x 0 l , ( 1 α ¯ t 1 ) I ) = α ¯ t 1 x 0 l + ( 1 α ¯ t 1 ) ε x l , l = 1 , 2 , . . . , L ,
where ε x l is the noise added to the intermediate sample x t 1 l . Then, we maintain the known regions information of x t 1 l by:
x ˜ t 1 l = x t 1 l m l ,
where ⊙ indicates the pixel-wise multiplication. Finally, we calculate a weighted value of known regions at diffusion time step t 1 for each pixel according to the frequency of value 1, which occurs in masks M L throughout the whole time step L:
X t 1 k n o w n = l = 1 L x ˜ t 1 l l = 1 L m l + s ,
where s is a small offset (set to 10 19 ) to prevent the unknown regions’ pixels divided by 0.

3.2.2. Conditional Inference Strategy

Before the training strategy, we first describe in detail the conditional inference strategy of SeqTIS, whose overview is shown in Figure 2.
The goal of cloud removal is to reconstruct the pixels corrupted by clouds or haze in optical satellite data using information from known regions (cloud-free pixels) or other modalities as a condition. In order to obtain as many as possible known regions for cloud removal, the multi-modal and multi-temporal sequences { X L , A 1 _ L , . . . , A n _ L , . . . , A N _ L } are used as inputs to the models during the inference process (cloud-removal process).
Since each reverse step in Equation (15) from X t to X t 1 depends on both the main modality X t and the auxiliary modalities A t 1 : N , we need to integrate information across the time of each modality sequence first and then alter the known regions, as long as the correct properties of the corresponding distribution can be maintained.
In order to integrate the information of the main modality X L at diffusion time step t 1 , we use the sequential data fusion module for the main modality expressed by Equations (21)–(23) to obtain the known regions’ information X t 1 k n o w n . Then, we use the sequential data fusion modules for auxiliary modality expressed by Equations (19) and (20) to obtain the information integrated value A ˜ t 1 : N . After that, we can obtain the unknown regions (corrupted by clouds or haze) information at t 1 by using both X t and A ˜ t 1 : N as inputs:
X t 1 u n k n o w n N ( μ θ ( X t , A ˜ t 1 : N , t ) , Σ θ ( X t , A ˜ t 1 : N , t ) )
To utilize the information from the known regions as fully as possible, we define a region mask M ˜ for altering the known regions as follows:
M ˜ = Θ ( l = 1 L m l ) ,
where Θ is a pixel-wise indicator, meaning that the output value will be 1 if the pixel value at the corresponding position is not 0; otherwise, it will be 0. Finally, we can keep the known regions’ information and obtain the next reverse step intermediate X t 1 as follows:
X t 1 = M ˜ X t 1 k n o w n + ( 1 M ˜ ) X t 1 u n k n o w n
The conditional inference strategy allows us to integrate temporal information across the arbitrary length of input sequences without retraining the model, which significantly reduces the training cost and increases the flexibility of inference. Algorithm 1 displays the above complete procedure of the conditional inference strategy.
Algorithm 1 Conditional inference (cloud removal) strategy of SeqTIS
1:
X T N ( 0 , I )
2:
for t = T , . . . , 1 do
3:
    for  l = 1 , . . . , L  do
4:
         ε x l N ( 0 , I ) if t > 1 , else ε x l = 0
5:
         x t 1 l N ( α ¯ t 1 x 0 l , ( 1 α ¯ t 1 ) I ) = α ¯ t 1 x 0 l + ( 1 α ¯ t 1 ) ε x l
6:
         x ˜ t 1 l = x t 1 l m l
7:
        for  n = 1 , . . . , N  do
8:
            ε a n _ l N ( 0 , I ) if t > 1 , else ε a n _ l = 0
9:
            a t n _ l N ( α ¯ t a 0 n _ l , ( 1 α ¯ t ) I ) = α ¯ t a 0 n _ l + ( 1 α ¯ t ) ε a n _ l
10:
        end for
11:
    end for
12:
     A ˜ t n = l = 1 L a t n _ l L , where n = 1 , 2 , . . . , N
13:
     X t 1 u n k n o w n N ( μ θ ( X t , A ˜ t 1 : N , t ) , Σ θ ( X t , A ˜ t 1 : N , t ) )
14:
     X t 1 k n o w n = l = 1 L x ˜ t 1 l l = 1 L m l + s
15:
     M ˜ = Θ ( l = 1 L m l )
16:
     X t 1 = M ˜ X t 1 k n o w n + ( 1 M ˜ ) X t 1 u n k n o w n
17:
end for
18:
return X 0

3.2.3. Temporal Training Strategy

Unlike RePaint [30], which only uses the pre-trained unconditional DDPMs based on RGB dataset as a prior, it is necessary to train MmDMs based on multi-spectral satellite data from the beginning. Therefore, we propose a specific training strategy, the temporal training strategy, to accurately capture the real distribution of cloud-free main modality q ( X ^ ) and to force the models to fully leverage the auxiliary modalities’ information to deal with the extreme absence of the main modality.
In order to capture the distribution of q ( X ^ ) as a cloud removal prior, we leverage the powerful distribution capture capability of MmDMs by using the cloud-free samples { X ^ , A ^ 1 : N } in training split as inputs and optimize the parameter θ in the neural networks as follows:
θ [ | | ε x ^ ε θ ( α ¯ t X ^ 0 + 1 α ¯ t ε x ^ , α ¯ t A ^ 0 1 : N + 1 α ¯ t ε a ^ 1 : N , t ) | | 2 + λ L V L B ] .
where ε x ^ N ( 0 , I ) is the noise added to the input sample X ^ 0 .
Due to the probability of each sample of cloudy input sequence X L being severely corrupted by clouds, we have to force the models to make full use of the information from the auxiliary modalities by training with the cloudy sequences as well. We need to process X L and A ( 1 : N ) _ L in the training split by the sequential data fusion modules to obtain the known regions’ information X t k n o w n and A ˜ t 1 : N at diffusion time step t. Then, we use the Gaussian random noise N ( 0 , I ) to fill the unknown regions and optimize the parameter θ as follows:
θ [ | | ε x ^ ε θ ( X t , A ˜ t 1 : N , t ) | | 2 + λ L V L B ] .
Algorithm 2 displays the complete procedure of the temporal training strategy in detail.
Algorithm 2 Temporal training strategy of SeqTIS
1:
repeat
2:
     X ^ 0 q ( X ^ )
3:
     A ^ 0 n , where n = 1 , 2 , . . . , N           ▹ coregistered and paired with X ^ 0
4:
     X 0 L = { x 0 1 , . . . , x 0 l . . . , x 0 L }                ▹ corresponding to X ^ 0
5:
     A 0 ( 1 : N ) _ L = { a 0 ( 1 : N ) _ 1 , . . . , a 0 ( 1 : N ) _ l , . . . , a 0 ( 1 : N ) _ L }      ▹ coregistered and paired with X 0 L
6:
     t Uniform ( { 1 , . . . , T } )
7:
     ε x ^ N ( 0 , I ) if t > 1 , else ε x ^ = 0
8:
     ε a ^ n N ( 0 , I ) if t > 1 , else ε a ^ n = 0 , where n = 1 , 2 , . . . , N
9:
    Take gradient descent step on
10:
     θ [ | | ε x ^ ε θ ( α ¯ t X ^ 0 + 1 α ¯ t ε x ^ , α ¯ t A ^ 0 1 : N + 1 α ¯ t ε a ^ 1 : N , t ) | | 2 + λ L V L B ]
11:
    for  l = 1 , . . . , L  do
12:
         ε x l N ( 0 , I ) if t > 1 , else ε x l = 0
13:
         x t l N ( α ¯ t x 0 l , ( 1 α ¯ t ) I ) = α ¯ t x 0 l + ( 1 α ¯ t ) ε x l
14:
         x ˜ t l = x t l m l
15:
        for  n = 1 , . . . , N  do
16:
            ε a n _ l N ( 0 , I ) if t > 1 , else ε a n _ l = 0
17:
            a t n _ l N ( α ¯ t a 0 n _ l , ( 1 α ¯ t ) I ) = α ¯ t a 0 n _ l + ( 1 α ¯ t ) ε a n _ l
18:
        end for
19:
    end for
20:
     A ˜ t n = l = 1 L a t n _ l L , where n = 1 , 2 , . . . , N
21:
     X t u n k n o w n N ( 0 , I )
22:
     X t k n o w n = l = 1 L x ˜ t l l = 1 L m l + s
23:
     M ˜ = Θ ( l = 1 L m l )
24:
     X t = M ˜ X t k n o w n + ( 1 M ˜ ) X t u n k n o w n
25:
    Take gradient descent step on
26:
     θ [ | | ε x ^ ε θ ( X t , A ˜ t 1 : N , t ) | | 2 + λ L V L B ]
27:
until converged

4. Results

To verify the feasibility of our method on the cloud-removal task in remote sensing domain, we conduct sufficient experiments on a public real-world dataset for cloud removal.

4.1. Dataset Description

This real-world dataset named SEN12MS-CR-TS [7] is a globally distributed dataset for multi-modal and multi-temporal cloud removal in remote sensing domain. It contains paired and co-registered sequences of spaceborne radar measurements practically unaffected by clouds, as well as cloud-covered and cloud-free multi-spectral optical satellite observations. Complementary to the radar modality’s cloud-robust information, historical satellite data are collected via Sentinel-1 and Sentinel-2 satellites from European Space Agency’s Copernicus mission, respectively. The Sentinel satellite provides public access data and is among the most prominent satellites for Earth observation.
Statistically, it contains observations covering 53 globally distributed regions of interest (ROIs) and registers 30 temporally aligned SAR Sentinel-1 as well as optical multi-spectral Sentinel-2 images throughout the whole year of 2018 in each ROI. Each band of every observation is upsampled to 10-m resolution (i.e., to the native resolution of Sentinel-2’s bands 2, 3, 4, and 8), and then full-scene images from all ROIs are sliced into 15,578 nonoverlapping patches of dimensions 256 × 256 px 2 with 30 time samples for every S1 and S2 measurement. The approximate cloud coverage of all data is about 50%, from clear-view images (e.g., used as ground truth), over semi-transparent haze, or small clouds to dense and wide cloud coverage.

4.2. Evaluation Metrics

We evaluate the quantitative performance in terms of normalized root mean squares error (NRMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [31], and spectral angle mapper (SAM) [32], defined as follows:
N R M S E ( x , y ) = 1 C · H · W c = h = w = 1 C , H , W ( x c , h , w y c , h , w ) 2 ,
P S N R ( x , y ) = 20 · log 10 ( 1 N R M S E ( x , y ) ) ,
S S I M ( x , y ) = ( 2 μ x μ y + ϵ 1 ) ( 2 σ x y + ϵ 2 ) ( μ x + μ y + ϵ 1 ) ( σ x + σ y + ϵ 2 ) ,
S A M ( x , y ) = cos 1 ( c = h = w = 1 C , H , W ( x c , h , w · y c , h , w ) c = h = w = 1 C , H , W x c , h , w 2 · c = h = w = 1 C , H , W y c , h , w 2 ) ,
with images x, y compared via their respective pixel values x c , h , w ,   y c , h , w [ 0 , 1 ] , dimensions C = 13 , H = W = 256 , which means μ x , μ y ; standard deviations σ x , σ y ; covariance σ x y ; and constants ϵ 1 , ϵ 2 , to stabilize the computation. NRMSE belongs to the class of pixel-wise metrics and quantifies the average discrepancy between the target and the prediction. PSNR is evaluated on the whole image and quantifies the signal-to-noise ratio of the prediction as a reconstruction of the target image. SSIM is another image-wise metric that builds on PSNR and captures the SSIM of the prediction to the target in terms of perceived change, contrast, and luminance [31]. The SAM measure is a third image-wise metric that provides the spectral angle between the bands of two multi-spectral images [32]. For further analysis, the pixel-wise metric NRMSE is evaluated in three manners: (1) over all pixels of the target image (as per convention), (2) only over cloud-covered pixels (visible in neither of any input optical sample) to measure reconstruction of noisy information, and (3) only over cloud-free pixels (visible in at least on input optical sample) quantifying the preservation of information. The pixel-wise masking is performed according to the cloud mask given by the cloud detector of s2cloudless [29].

4.3. Baseline Methods

We compare our proposed method with the following baseline methods: (1) Least cloudy: we just take the least-cloudy input optical observation and forward it without further modification to be compared against the cloud-free target image. This provides a benchmark of how hard the cloud-removal task is with respect to the extent of cloud-coverage present in the data. (2) Mosaicing: we perform a mosaicing method that averages the values of pixels across cloud-free time points, thereby integrating information across time. That is, for any pixel, if there is a single clear-view time point, then its value is copied; for multiple cloud-free samples, the mean is formed, and in case no cloud-free time point exists, then a value of 0.5 is taken as a proxy. The mosaicing technique provides a benchmark of how much information can be integrated across time from multi-spectral optical observations exclusively. (3) STGAN [17]: Spatio-temporal generator networks (STGANs) are proposed to generate a cloud-free image from the given sequence of cloudy images, which only leverage the RGB and IR bands of the optical observation. (4) Seq2point [7]: Seq2point denotes the sequence-to-point cloud removal method builds on the generator of STGAN, replacing the pairwise concatenation of 2D feature maps in STGAN by stacking features in the temporal domain, followed by 3D CNNs.

4.4. Implementation Details

To make a fair comparison, we train all versions of SeqDMs by an Adamw [33] optimizer with a learning rate of 0.0001 and utilize the half-precision (i.e., FP16) training technique [34] to obtain significant computational speedup and memory consumption reduction. The architecture of the neural network used in MmDMs is obtained by modifying the input channels suitable for the multi-modal information based on that in [28], which is a U-Net [35] model using a stack of residual layers and downsampling convolutions, followed by a stack of residual layers with upsampling convolutions, with skip connections connecting the layers with the same spatial size. In addition, we use global attention layers at the 32 × 32 , 16 × 16 , and 8 × 8 resolutions with 4 attention heads, 128 base channels, 2 residual blocks per resolution, BigGAN up/downsampling, and adaptive group normalization. In order to make a consistent comparison with the above compared methods, the modalities S1 and S2 are, respectively, value-clipped within the intervals of [−25,0] and [0,10,000] and then normalized to the range [−1,1] for stable training. We set the sequence length L = 3 in the training split for the temporal training strategy and train the models with batch size 1 for 10 epochs on GPUs RTX3090 for roughly five days. All of the other compared methods are also trained on SEN12MS-CR-TS according to the training protocol of [7].

4.5. Experimental Results

To evaluate performance and generalization of the proposed method, we use the whole test split over all of the continents of SEN12MS-CR-TS containing S2 observations from the complete range of cloud coverage (between 0% and 100%). Table 1 compares the results of our proposed method with the baseline methods detailed in Section 4.3, entirely trained and inferred with the input sequences of length L = 3 .
Since mosaicing directly averages the values on cloud-free time points for each pixel to integrate information across time, it does not perform well on imagery structure (e.g., perceived change, contrast, and luminance) as well as multi-spectral structures, while it has very little noise with the highest PSNR, indicating the maximum amount of information can be integrated.
The results also show that the proposed method SeqDMs outperforms the baselines in the majority of pixel-wise metrics and greatly exceeds Seq2point [7] in the image-wise metric PSNR, except for SSIM and SAM (where Seq2point [7] is a little better). This demonstrates that SeqDMs can obtain reconstructed samples with superior image quality due to its powerful ability of distribution capture and can typically outperform trivial solutions to the multi-modal multi-temproal cloud removal problem. The examples of the reconstructed outcomes for the considered baselines on four different samples from the test split are illustrated in Figure 3. The considered cases are cloud-free, partly cloudy, cloud-covered with no visibility except for one single time point, and cloud-covered with no visibility at any time point. The illustrations show that SeqDMs can perfectly maintain any cloud-free pixels of the input sequences and leverage the distribution of the known regions to generate the cloudy pixels.
However, SeqDMs fall short of capturing the imagery structure or multi-spectral structure when the input sequences are quiet short, in terms of SSIM and SAM metrics. The input sequences are quite short, indicating that input samples are collected in a relatively concentrated period. This produces three challenges for the cloud removal using SeqDMs with the powerful ability of distribution capture: (1) Cloud coverage might be quiet high, resulting in severe information loss. (2) The temporal shift between input samples and cloud-free target image might be large, causing significant differences in the perceived change, contrast, or luminance. (3) The robustness of exceptional data due to equipment failure might be weak, misleading the model to a wrong inference direction that is completely different from the cloud-free target image.
To overcome the above challenges, we consider the inference process (i.e., cloud-removal process) over longer input sequences by using the conditional inference strategy detailed in Section 3.2.2 to integrate more information across time without retraining the SeqDMs again. Table 2 reports the performance of our proposed method SeqDMs and Seq2point [7] inferred with input sequences of length L = 3 , 4 , 5 , respectively. It is worth noting that SeqDMs need to be trained only once with sequences of length L = 3 , while Seq2point [7] needs to be trained with sequences of length L = 3 , 4 , 5 , respectively. The results indicate that inferring with longer input sequences can significantly improve the performance of SeqDMs in terms of reconstructed quality, imagery structure, and multi-spectral structure and can easily outperform the baseline methods in majority metrics with much less training cost, except for SAM.
To further understand the benefit of conditional inference strategy, Table 3 reports the performance of SeqDMs as a function of cloud coverage, inferred with input sequences of length L = 3 , 4 , 5 . The cloud coverage is calculated by averaging the extent of clouds of each images in the sequence of length L = 3 . The results show that longer input sequences can significantly improve the performance, especially in the cases of extreme cloud coverage. Figure 4 shows the performance histograms of SeqDMs in terms of PSNR, SSIM, NRMSE(cloudy), and SAM; it visualizes the significant improvements in the extreme cloud coverage cases by increasing input sequences length L. In addition, the cloud-removal performance is highly dependent on the percentage of the cloud coverage. While the performance decrease is not strictly monotonous with an increase in cloud coverage, a strong association still persists.
Finally, we conduct an ablation experiment to assess the benefit of utilizing the temporal training strategy of SeqTIS. Table 4 compares the results of our propose method SeqDMs with an ablation version not using temporal training strategy (i.e., only trained with Equation (27)). The comparison demonstrates that using the whole version of temporal training strategy of SeqTIS leads to a higher quality in the cloud-removal task.

5. Discussion

The main contribution of this paper is in the development of the sequential-based diffusion models (SeqDMs), which is a novel probabilistic generative model for the cloud-removal task of optical satellite imagery. It consists of two parts, multi-modal diffusion models (MmDMs) and sequential-based training and inference strategy (SeqTIS). MmDMs are novel diffusion models that reconstruct the reverse process of DDPMs to integrate additional information from the auxiliary modalities (e.g., SAR or other modalities robust to the corruption of clouds or haze) in order to help the distribution learning of the main modality (i.e., optical satellite imagery). Although the main modality typically is susceptible to the influences of clouds or haze, MmDMs are capable of capturing the distribution of the main modality by conditioning the missing information with the auxiliary modalities during the training and inference process. SeqTIS is an improved training and inference strategy specifically for MmDMs, which allows us to integrate temporal information across arbitrary length of the both main modality and auxiliary modalities input sequences without retraining the model again. With the help of the MmDMs and SeqTIS, our proposed method SeqDMs outperform several other state-of-the-art multi-modal multi-temporal cloud removal methods and have the flexibility to handle the arbitrary length of input sequences, producing significant improvements with only one or two additional input samples and greatly reducing the time cost of model training, as detailed in Table 1 and Table 2. This work serves as an important stepping stone for cloud removal by integrating information across time and data modalities to achieve improved interpretability, model flexibility, and generalizability. However, due to the powerful distribution capture capability of MmDMs and the direct information combination of known regions and unknown regions by SeqTIS, knowing how to more effectively enhance the robustness of the exceptional data in sequence and more efficiently extract useful information based on the transparency of the cloud is still crucial to fundamentally improve the performance of the proposed method rather than inferring over longer input sequences.

6. Conclusions

This paper proposes SeqDMs, a novel probabilistic generative model for the cloud-removal task in remote sensing. Unlike other popular generative models, our method introduces a novel diffusion model by reconstructing the reverse process of DDPMs to integrate additional information from the auxiliary modalities and utilizes a specialized training and inference strategy to handle sequences of an arbitrary length without retraining the model again. Our extensive experiments demonstrate that our method outperforms several state-of-the-art methods in cloud removal with excellent interpretability and flexibility. In the future, we will pursue the direction of studying the characteristics of each band of the multi-spectral optical satellite imagery to extract more helpful information for further reducing the semantic gap between the reconstructions and target images.

Author Contributions

Conceptualization, X.Z.; investigation, X.Z. and K.J.; methodology, X.Z. and K.J.; supervision, K.J.; validation, X.Z.; visualization, X.Z.; writing—original draft, X.Z.; and writing—review and editing, X.Z. and K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Data Availability Statement

The remote sensing dataset SEN12MS-CR-TS were downloaded from: https://patricktum.github.io/cloud_removal/sen12mscrts/ (accessed on 6 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vakalopoulou, M.; Karantzalos, K.; Komodakis, N.; Paragios, N. Building detection in very high resolution multispectral data with deep learning features. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 1873–1876. [Google Scholar]
  2. Jean, N.; Burke, M.; Xie, M.; Davis, W.M.; Lobell, D.B.; Ermon, S. Combining satellite imagery and machine learning to predict poverty. Science 2016, 353, 790–794. [Google Scholar] [CrossRef] [PubMed]
  3. Nataliia, K.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar]
  4. Kussul, N.; Skakun, S.; Shelestov, A.; Lavreniuk, M.; Yailymov, B.; Kussul, O. Regional scale crop mapping using multi-temporal satellite imagery. In Proceedings of the 36th International Symposium on Remote Sensing of Environment (ISRSE36), Berlin, Germany, 11–15 May 2015; pp. 45–52. [Google Scholar]
  5. Castillo, J.A.A.; Apan, A.A.; Maraseni, T.N.; Salmo, S.G. Estimation and mapping of above-ground biomass of mangrove forests and their replacement land uses in the philippines using sentinel imagery. ISPRS J. Photogramm. Remote Sens. 2017, 134, 70–85. [Google Scholar] [CrossRef]
  6. King, M.D.; Platnick, S.; Menzel, W.P.; Ackerman, S.A.; Hubanks, P.A. Spatial and temporal distribution of clouds observed by MODIS onboard the Terra and Aqua satellites. IEEE Trans. Geosci. Remote Sens. 2013, 51, 3826–3852. [Google Scholar] [CrossRef]
  7. Ebel, P.; Xu, Y.; Schmitt, M.; Zhu, X.X. SEN12MS-CR-TS: A Remote-Sensing Data Set for Multimodal Multitemporal Cloud Removal. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  8. Hoan, N.T.; Tateishi, R. Cloud removal of optical image using SAR data for ALOS applications. Experimenting on simulated ALOS data. J. Remote Sens. Soc. Jpn. 2009, 29, 410–417. [Google Scholar]
  9. Eckardt, R.; Berger, C.; Thiel, C.; Schmullius, C. Removal of optically thick clouds from multi-spectral satellite images using multi-frequency SAR data. Remote Sens. 2013, 5, 2973–3006. [Google Scholar] [CrossRef]
  10. Grohnfeldt, C.; Schmitt, M.; Zhu, X. A conditional generative adversarial network to fuse SAR and multispectral optical data for cloud removal from sentinel-2 images. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium (IGARSS 2018), Valencia, Spain, 22–27 July 2018; pp. 1726–1729. [Google Scholar]
  11. Meraner, A.; Ebel, P.; Zhu, X.X.; Schmitt, M. Cloud removal in Sentinel-2 imagery using a deep residual neural network and SAR-optical data fusion. ISPRS J. Photogramm. Remote Sens. 2020, 166, 333–346. [Google Scholar] [CrossRef]
  12. Ebel, P.; Meraner, A.; Schmitt, M.; Zhu, X.X. Multisensor data fusion for cloud removal in global and all-season Sentinel-2 imagery. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5866–5878. [Google Scholar] [CrossRef]
  13. Ebel, P.; Schmitt, M.; Zhu, X.X. Cloud removal in unpaired Sentinel-2 imagery using cycle-consistent GAN and SAR-optical data fusion. In Proceedings of the 2020 IEEE International Geoscience and Remote Sensing Symposium (IGARSS 2020), Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2065–2068. [Google Scholar]
  14. Gao, J.; Yuan, Q.; Li, J.; Zhang, H.; Su, X. Cloud removal with fusion of high resolution optical and SAR images using generative adversarial networks. Remote Sens. 2020, 12, 191. [Google Scholar] [CrossRef]
  15. Bamler, R. Principles of synthetic aperture radar. Surv. Geophys. 2000, 21, 147–157. [Google Scholar] [CrossRef]
  16. Meng, Q.; Borders, B.E.; Cieszewski, C.J.; Madden, M. Closest spectral fit for removing clouds and cloud shadows. Photogramm. Eng. Remote Sens. 2009, 75, 569–576. [Google Scholar] [CrossRef]
  17. Sarukkai, V.; Jain, A.; Uzkent, B.; Ermon, S. Cloud removal in satellite images using spatiotemporal generative networks. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 1785–1794. [Google Scholar]
  18. Helmer, E.H.; Ruefenacht, B. Cloud-free satellite image mosaics with regression trees and histogram matching. Photogramm. Eng. Remote Sens. 2005, 71, 1079–1089. [Google Scholar] [CrossRef]
  19. Tseng, D.C.; Tseng, H.T.; Chien, C.L. Automatic cloud removal from multi-temporal spot images. Appl. Math. Comput. 2008, 205, 584–600. [Google Scholar] [CrossRef]
  20. Ramoino, F.; Tutunaru, F.; Pera, F.; Arino, O. Ten-meter sentinel-2a cloud-free composite—Southern Africa 2016. Remote Sens. 2017, 9, 652. [Google Scholar] [CrossRef]
  21. Oehmcke, S.; Chen, T.H.K.; Prishchepov, A.V.; Gieseke, F. Creating cloud-free satellite imagery from image time series with deep learning. In Proceedings of the 9th ACM SIGSPATIAL International Workshop on Analytics for Big Geospatial Data, Seattle, WA, USA, 3 November 2020; pp. 1–10. [Google Scholar]
  22. Zhang, Q.; Yuan, Q.; Li, Z.; Sun, F.; Zhang, L. Combined deep prior with low-rank tensor SVD for thick cloud removal in multitemporal images. ISPRS J. Photogramm. Remote Sens. 2021, 177, 161–173. [Google Scholar] [CrossRef]
  23. Rafique, M.U.; Blanton, H.; Jacobs, N. Weakly supervised fusion of multiple overhead images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1479–1486. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. Salimans, T.; Goodfellow, I.; Zaremba, W.; Cheung, V.; Radford, A.; Chen, X. Improved techniques for training gans. In Proceedings of the Advances in Neural Information Processing Systems 29 (NeurIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  26. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020; pp. 6840–6851. [Google Scholar]
  27. Nichol, A.Q.; Dhariwal, P. Improved denoising diffusion probabilistic models. In Proceedings of the 38th International Conference on Machine Learning, Virtual. 18–24 July 2021; pp. 8162–8171. [Google Scholar]
  28. Dhariwal, P.; Nichol, A.Q. Diffusion models beat gans on image synthesis. In Proceedings of the Advances in Neural Information Processing Systems 34 (NeurIPS 2021), Virtual. 6–14 December 2021; pp. 8780–8794. [Google Scholar]
  29. Improving Cloud Detection with Machine Learning. Available online: https://medium.com/sentinel-hub/improving-cloud-detection-with-machine-learning-c09dc5d7cf13 (accessed on 10 October 2019).
  30. Lugmayr, A.; Danelljan, M.; Romero, A.; Yu, F.; Timofte, R.; Gool, L.V. Repaint: Inpainting using denoising diffusion probabilistic models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022), New Orleans, LA, USA, 21–24 June 2022; pp. 11461–11471. [Google Scholar]
  31. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  32. Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W.; Heidebrecht, K.B.; Shapiro, A.T.; Barloon, P.J.; Goetz, A.F.H. The spectral image processing system (SIPS)-interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  33. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. In Proceedings of the 7th International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  34. Micikevicius, P.; Alben, J.; Garcia, D.; Ginsburg, B.; Houston, M.; Kuchaiev, O.; Venkatesh, G.; Wu, H. Mixed precision training. In Proceedings of the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI 2015), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
Figure 1. Graphical model for multi-modal diffusion models (MmDMs), which consists of one main modality X and N auxiliary modalities A. In order to reverse the diffusion process of X, it learns a neural network p θ ( X t 1 | X t , A t 1 : N ) to approximate the intractable posterior distribution q ( X t 1 | X t ) .
Figure 1. Graphical model for multi-modal diffusion models (MmDMs), which consists of one main modality X and N auxiliary modalities A. In order to reverse the diffusion process of X, it learns a neural network p θ ( X t 1 | X t , A t 1 : N ) to approximate the intractable posterior distribution q ( X t 1 | X t ) .
Remotesensing 15 02861 g001
Figure 2. The overview of conditional inference strategy in sequential-based training and inference strategy (SeqTIS). The symbol ∗ indicates the pixel-wise multiplication.
Figure 2. The overview of conditional inference strategy in sequential-based training and inference strategy (SeqTIS). The symbol ∗ indicates the pixel-wise multiplication.
Remotesensing 15 02861 g002
Figure 3. Exemplary input sequences, reconstructed outcomes, and cloud-free target images for some baselines reported in Table 1 in four considered cases. Columns: Four different samples from the test split. The four considered cases are cloud-free, partly cloudy, cloud-covered with no visibility except for one single time point, and cloud-covered with no visibility in any time point. Rows: Three samples of input sequences, reconstructed outcomes of mosaicing, Seq2point [7], and SeqDMs, as well as the cloud-free target image.
Figure 3. Exemplary input sequences, reconstructed outcomes, and cloud-free target images for some baselines reported in Table 1 in four considered cases. Columns: Four different samples from the test split. The four considered cases are cloud-free, partly cloudy, cloud-covered with no visibility except for one single time point, and cloud-covered with no visibility in any time point. Rows: Three samples of input sequences, reconstructed outcomes of mosaicing, Seq2point [7], and SeqDMs, as well as the cloud-free target image.
Remotesensing 15 02861 g003aRemotesensing 15 02861 g003b
Figure 4. The performance histograms of SeqDMs inferred with input sequences of length L = 3 , 4 , 5 in terms of PSNR, SSIM, NRMSE(cloudy), and SAM.
Figure 4. The performance histograms of SeqDMs inferred with input sequences of length L = 3 , 4 , 5 in terms of PSNR, SSIM, NRMSE(cloudy), and SAM.
Remotesensing 15 02861 g004
Table 1. Quantitative evalutation of the proposed method SeqDMs with baseline methods in terms of normalized root mean squared error (NRMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [31], and spectral angle mapper (SAM) [32] metrics. All methods are trained and inferred with the input sequences of length L = 3 .
Table 1. Quantitative evalutation of the proposed method SeqDMs with baseline methods in terms of normalized root mean squared error (NRMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [31], and spectral angle mapper (SAM) [32] metrics. All methods are trained and inferred with the input sequences of length L = 3 .
ModelNRMSE(All) ↓NRMSE(Cloudy)NRMSE(Clear)PSNR ↑SSIM ↑SAM ↓
least cloudy0.0790.0820.03122.980.8150.213
mosaicing0.0620.0640.03631.680.8110.250
STGAN0.0570.0590.05025.420.8180.219
Seq2point0.0510.0520.04026.680.8360.186
SeqDMs(proposed)0.0450.0460.02628.070.8270.223
The ↓ means lower is better, and ↑ means the higher is better.
Table 2. Quantitative evalutation of the proposed method SeqDMs and Seq2point [7] inferred with input sequences of length L = 3 , 4 , 5 in terms of NRMSE, PSNR, SSIM [31], and SAM [32] metrics. It is worth noting that SeqDMs need to be trained only once with sequences of length L = 3 , while Seq2point [7] needs to be trained with sequences of length L = 3 , 4 , 5 , respectively.
Table 2. Quantitative evalutation of the proposed method SeqDMs and Seq2point [7] inferred with input sequences of length L = 3 , 4 , 5 in terms of NRMSE, PSNR, SSIM [31], and SAM [32] metrics. It is worth noting that SeqDMs need to be trained only once with sequences of length L = 3 , while Seq2point [7] needs to be trained with sequences of length L = 3 , 4 , 5 , respectively.
ModelNRMSE(All) ↓NRMSE(Cloudy)NRMSE(Clear)PSNR ↑SSIM ↑SAM ↓
Seq2point (L = 3)0.0510.0520.04026.680.8360.186
SeqDMs (L = 3)0.0450.0460.02628.070.8270.223
Seq2point (L = 4)0.0490.0500.04127.100.8450.172
SeqDMs (L = 4)0.0370.0380.02428.310.8470.201
Seq2point (L = 5)0.0480.0480.03227.070.8460.178
SeqDMs (L = 5)0.0380.0380.01728.210.8460.198
The ↓ means lower is better, and ↑ means the higher is better.
Table 3. Performance of our proposed method SeqDM as a function of cloud coverage, inferred with input sequences of length L = 3 , 4 , 5 . The cloud coverage is calculated by averaging the extent of clouds of each images in the sequence of length L = 3 .
Table 3. Performance of our proposed method SeqDM as a function of cloud coverage, inferred with input sequences of length L = 3 , 4 , 5 . The cloud coverage is calculated by averaging the extent of clouds of each images in the sequence of length L = 3 .
Cloud CoverageLNRMSE(All) ↓NRMSE(Cloudy)NRMSE(Clear)PSNR ↑SSIM ↑SAM ↓
0–10%30.0230.0300.02332.310.9070.118
40.0240.0280.02131.810.9050.120
50.0240.0260.01531.740.9070.120
10–20%30.0350.0350.03428.000.8630.161
40.0340.0330.03828.170.8650.160
50.0330.0330.02928.480.8710.157
20–30%30.0330.0320.03429.300.8540.138
40.0320.0320.03829.250.8550.138
50.0320.0310.02129.610.8610.135
30–40%30.0330.0320.03229.860.8750.156
40.0330.0330.02829.080.8730.160
50.0330.0330.02729.080.8750.160
40–50%30.0450.0450.04226.540.8260.227
40.0420.0420.03726.960.8340.223
50.0420.0420.03426.930.8340.222
50–60%30.0420.0420.02328.400.8320.217
40.0410.0410.02027.910.8410.217
50.0410.0410.03027.910.8410.217
60–70%30.0380.0380.02929.360.8380.217
40.0380.0380.02427.880.8390.220
50.0400.040-27.680.8320.220
70–80%30.0480.048-24.600.8030.268
40.0400.040-27.130.8420.228
50.0450.045-26.680.8220.225
80–90%30.0500.050-24.580.7790.285
40.0380.038-28.600.8340.216
50.0400.040-28.480.8330.209
90–100%30.1040.104-18.800.6090.434
40.0590.059-24.520.7550.288
50.0540.054-24.960.7820.261
The ↓ means lower is better, and ↑ means the higher is better.
Table 4. Comparison of the proposed method SeqDMs ( L = 3 ) using the temporal training strategy of SeqTIS versus an ablation version not using the temporal training strategy (i.e., only trained with Equation (27)) in terms of NRMSE, PSNR, SSIM [31], and SAM [32] metrics.
Table 4. Comparison of the proposed method SeqDMs ( L = 3 ) using the temporal training strategy of SeqTIS versus an ablation version not using the temporal training strategy (i.e., only trained with Equation (27)) in terms of NRMSE, PSNR, SSIM [31], and SAM [32] metrics.
ModelNRMSE(All) ↓NRMSE(Cloudy)NRMSE(Clear)PSNR ↑SSIM ↑SAM ↓
SeqDM (no temporal training strategy)0.0480.0500.02627.900.8210.223
SeqDM (with temporal training strategy)0.0450.0460.02628.070.8270.223
The ↓ means lower is better, and ↑ means the higher is better.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, X.; Jia, K. Cloud Removal in Remote Sensing Using Sequential-Based Diffusion Models. Remote Sens. 2023, 15, 2861. https://doi.org/10.3390/rs15112861

AMA Style

Zhao X, Jia K. Cloud Removal in Remote Sensing Using Sequential-Based Diffusion Models. Remote Sensing. 2023; 15(11):2861. https://doi.org/10.3390/rs15112861

Chicago/Turabian Style

Zhao, Xiaohu, and Kebin Jia. 2023. "Cloud Removal in Remote Sensing Using Sequential-Based Diffusion Models" Remote Sensing 15, no. 11: 2861. https://doi.org/10.3390/rs15112861

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop