Next Article in Journal
Surface Deformation Monitoring and Prediction of InSAR-Hybrid Deep Learning Model for Subsidence Funnels
Previous Article in Journal
Nighttime Remote Sensing Analysis of Lit Fishing Boats: Fisheries Management Challenges in the South China Sea (2013–2022)
Previous Article in Special Issue
A Cross-Spatial Differential Localization Network for Remote Sensing Change Captioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Conditional Diffusion Model for SAR Despeckling

1
School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing 100081, China
2
Terahertz Science Application Center (TSAC), Beijing Institute of Technology, Zhuhai 519088, China
3
Shanghai Institute of Satellite Engineering, Shanghai 201100, China
4
Yangtze Delta Region Academy in Jiaxing, Beijing Institute of Technology, Jiaxing 314019, China
5
Department of Automation, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(17), 2970; https://doi.org/10.3390/rs17172970
Submission received: 1 August 2025 / Revised: 21 August 2025 / Accepted: 26 August 2025 / Published: 27 August 2025

Abstract

Speckle noise inherent in Synthetic Aperture Radar (SAR) images severely degrades image quality and hinders downstream tasks such as interpretation and target recognition. Existing despeckling methods, both traditional and deep learning-based, often struggle to balance effective speckle suppression with structural detail preservation. Although Denoising Diffusion Probabilistic Models (DDPMs) have shown remarkable potential for SAR despeckling, their computational overhead from iterative sampling severely limits practical applicability. To mitigate these challenges, this paper proposes the Efficient Conditional Diffusion Model (ECDM) for SAR despeckling. We integrate the cosine noise schedule with a joint variance prediction mechanism, accelerating the inference speed by an order of magnitude while maintaining high denoising quality. Furthermore, we integrate wavelet transforms into the encoder’s downsampling path, enabling adaptive feature fusion across frequency bands to enhance structural fidelity. Experimental results demonstrate that, compared to a baseline diffusion model, our proposed method achieves an approximately 20-fold acceleration in inference and obtains significant improvements in key objective metrics. This work contributes to real-time processing of diffusion models for SAR image enhancement, supporting practical deployment by mitigating prolonged inference in traditional diffusion models through efficient stochastic sampling.

1. Introduction

Synthetic Aperture Radar (SAR) acquires electromagnetic scattering information from the Earth’s surface by transmitting and receiving microwave signals. Operating independently of illumination and meteorological conditions, it has become an indispensable technology for critical remote sensing applications, including disaster monitoring, vegetation structure inversion, and urban deformation detection [1]. However, the coherent nature of SAR imaging generates speckle noise through random interference of signals within each resolution cell. Speckle not only degrades the signal-to-noise ratio (SNR) of the imagery but also disrupts the local statistical homogeneity of scene features, obscuring structural edges and fine-grained textures [2]. By distorting class boundaries and masking target signatures, speckle systematically impairs the accuracy and reliability of critical downstream tasks such as land-cover classification, target detection [3], and change detection [4]. Therefore, developing advanced algorithms that achieve effective speckle suppression while maximally preserving structural and radiometric fidelity is paramount for enhancing the performance and robustness of high-level SAR applications.
Historically, research in SAR despeckling has progressed along three primary paradigms: local statistical filtering, transform-domain filtering, and non-local filtering. Local statistical filters operate by performing adaptive pixel-wise smoothing within a localized window [5]. This class of algorithms is exemplified by seminal works such as the Lee [6], Frost [7], Kuan [8], and Gamma-MAP [9] filters. While computationally efficient, their inherent locality imposes a fundamental trade-off: in homogeneous regions, they tend to blur subtle textures, whereas in heterogeneous areas, they often fail to preserve sharp edges and strong scatterers. The transform-domain paradigm for despeckling often begins within a homomorphic framework, where a logarithmic transform is first applied to linearize the multiplicative speckle into an additive noise model [10]. Subsequently, multi-scale geometric transforms [11] are employed to decompose the image, allowing for coefficient shrinkage to suppress noise while preserving structural information. To address the non-adaptive nature of conventional hard/soft thresholding, subsequent research pivoted towards more sophisticated statistical modelling. These strategies ranged from deriving adaptive shrinkage operators based on probability mixture models [12] to imposing spatial contextual constraints on wavelet coefficients via Markov Random Fields [13] and even transplanting classical Wiener filtering theory into the stationary wavelet domain to seek an optimal estimate [14]. However, a critical limitation of these “model-driven” methods is their reliance on fixed statistical priors, which are often inadequate for capturing the full spectrum of complex heterogeneity in real SAR scenes [15]. Non-local methods exploit Non-Local Self-Similarity by globally identifying similar patches for weighted filtering. The Non-Local Means algorithm [16] pioneered this approach, denoising via weighted averaging of similar patches to preserve fine details. For multiplicative speckle in SAR images, the Probabilistic Patch-Based (PPB) method [17] incorporated Gamma-distributed likelihood, aligning patch similarity with the noise model. This evolved into SAR-BM3D [18], which integrated a SAR-optimized similarity metric into the Block-Matching and 3D (BM3D) collaborative filtering framework, effectively balancing texture preservation and speckle suppression. However, this series of methods is typically limited by filtering parameter design and similarity criteria, incurring high computational costs [19]. These inherent bottlenecks have driven the shift toward data-driven deep learning paradigms.
In recent years, deep learning has become the benchmark for numerous image processing tasks [20]. SAR-CNN [21], proposed by Chierchia et al., pioneers the application of Convolutional Neural Networks (CNNs) to SAR image despeckling, employing a 17-layer residual network to generate despeckled images. Subsequently, ID-CNN [22] leverages synthetic speckle in optical images for supervision, optimizing a combined Euclidean loss and Total Variation (TV) regularizer to achieve smoother restoration without logarithmic transformation. SAR-DRN [23] integrates dilated convolutions into residual blocks with skip connections to address vanishing gradients, enhancing robustness in high-noise scenarios. MONet [24] employs a composite loss function integrating pixel-wise Mean Squared Error (MSE), Kullback–Leibler (KL) divergence, and a term preserving strong scatterers, enabling synergistic optimization of texture preservation and speckle statistical consistency across homogeneous, heterogeneous, and extremely heterogeneous regions. Introduced in 2017 [25], the Transformer framework established an effective approach for capturing long-range dependencies. Trans-SAR [26] pioneered the use of the hierarchical Vision Transformer for SAR despeckling, outperforming contemporary CNN-based models. Subsequently, SAR-CAM [27] introduced a Continuous Attention Module (CAM) and a Contextual Block (CB), significantly enhancing the representation of critical features.
Due to the absence of genuine speckle-free SAR images, early methods primarily generated synthetic samples by overlaying Gamma noise on optical images for supervised training. Subsequently, SAR2SAR [28] exploited noise redundancy to develop an unsupervised framework, eliminating the need for clean ground truth. Speckle2Void [29] utilized blind-spot CNNs and Bayesian posterior reconstruction for single-image self-supervision. Beyond unsupervised approaches, researchers have investigated high-SNR “pseudo ground truth” from multi-temporal images of the same scene to enable supervised training. For example, Vitale et al. [30] proposed a two-stage training strategy: pre-training on synthetic data followed by domain-adaptive fine-tuning via multi-temporal interferometric phase fusion.
In recent years, the proliferation of generative models in image processing has highlighted the potential of Generative Adversarial Networks (GANs) for SAR despeckling, optimizing perceptual quality and distributional consistency via adversarial learning between generator and discriminator [31]. More recently, Denoising Diffusion Probabilistic Models (DDPMs), prized for stable training and superior detail synthesis, have been applied to SAR despeckling. SAR-DDPM [32] pioneered DDPM adoption for this task. Building thereon, Diffusion SAR developed a conditional DDPM in the log-Yeo-Johnson transform domain [33]. Concurrently, Pan et al. [34] incorporated Swin Transformer Blocks into the noise predictor and introduced Pixel-Shuffle Down-sampling (PD) Refinement to address the domain gap.
However, existing methods—whether CNN- or Transformer-based—still struggle to balance noise suppression with information fidelity (i.e., preserving textures and structures). Although diffusion models have achieved breakthroughs in quantitative metrics, they introduce new challenges: artifacts and blurring persist in fine detail reconstruction, and the enormous inference overhead from thousands of iterative sampling steps poses a major barrier to practical application. Thus, developing a SAR despeckling model that simultaneously ensures high image fidelity (in terms of textures and structures) and high inference efficiency remains a core challenge urgently requiring resolution in this field.
To address the aforementioned challenges, this paper proposes the Efficient Conditional Diffusion Model, a framework that systematically enhances the comprehensive performance of SAR image despeckling through three key contributions:
An Efficient Inference Framework: We design a denoising network that jointly predicts noise components and variance parameters, enabling accelerated sampling through optimized timestep scheduling. This approach achieves 20× speedup in inference time while maintaining denoising quality. A Highly Effective Network Architecture: We integrate discrete wavelet transforms into the encoder’s downsampling stages, providing lossless frequency decomposition and multi-scale feature representation. This design preserves structural information more effectively during the despeckling process. An Efficacious Training Strategy: We employ pre-training on synthetic data followed by fine-tuning on real multi-temporal SAR images, effectively bridging the domain gap between synthetic training data and real-world SAR scenarios.
The remainder of this paper is organized as follows. Section 2 reviews the mathematical modeling of SAR speckle noise and the foundational principles of Denoising Diffusion Probabilistic Models (DDPMs). Section 3 details our proposed method, including its network architecture, efficient inference strategy, and loss function design. Section 4 presents comprehensive experimental results, encompassing quantitative comparisons and qualitative assessments on both synthetic and real-world SAR datasets. Section 5 conducts a series of ablation studies to systematically validate the effectiveness of key components in our model and the accelerated sampling strategy. Finally, Section 6 concludes the paper and outlines potential future research directions.

2. Related Works

2.1. SAR Speckle Model

While enabling high-resolution imaging through coherent processing, Synthetic Aperture Radar (SAR) inherently produces speckle noise. This noise arises from the coherent superposition of echoes from multiple elementary scatterers within a resolution cell, resulting in random interference due to their varying phases. Consequently, speckle noise exhibits multiplicative characteristics, being coupled with signal intensity, and is spatially correlated. It is mathematically modeled as [35]:
X = Y N
where X denotes the SAR image corrupted by speckle noise; Y represents the ideal noise-free reflectivity; and N signifies the speckle noise generated during the SAR imaging process. The speckle noise N typically follows a Gamma distribution, with its probability density function (PDF) expressed as:
P N = L L N L 1 e L N Γ L
where L represents the Equivalent Number of Looks (ENL), Γ L denotes the Gamma function, and the noise N is characterized by a mean of 1 and a variance of 1 / L . Given the complex statistical properties of multiplicative speckle in SAR imagery, a comprehensive evaluation of denoising algorithm performance must consider four key dimensions: speckle suppression, structural fidelity, radiometric fidelity, and artifact suppression [36].

2.2. Diffusion Models for SAR Despeckling

Denoising Diffusion Probabilistic Models (DDPMs) are a class of deep generative models inspired by non-equilibrium thermodynamics [37]. They define a forward diffusion process and a learnable reverse denoising process, both modeled as Markov chains, as shown in Figure 1.
The forward diffusion process gradually adds noise to the input data. It is modeled as a predefined Markov chain, starting from an input image x 0 and iteratively injecting Gaussian noise at each step until the signal is fully obscured. The forward transition distribution for a single step is defined as:
q x t | x t 1 = N x t ; 1 β t x t 1 , β t I
where β t 0 , 1 is a pre-defined variance schedule that governs the noise level at each timestep t . This transition kernel indicates that the state x t is obtained by scaling the preceding state x t 1 by 1 β t and adding Gaussian noise with variance β t . Through the reparameterization trick, a noisy version of the initial state x 0 at an arbitrary timestep t can be directly sampled:
q x t | x 0 = N x t ; α ¯ t x 0 , 1 α ¯ t I
where α t = 1 β t , α ¯ t = s = 1 t α s . For a sufficiently large T and a well-designed schedule, x T converges to a standard isotropic Gaussian distribution N 0 , I . This property yields a closed-form expression for generating training samples, which forms the cornerstone of the entire training procedure:
x t = α ¯ t x 0 + 1 α ¯ t ϵ
In the task of Synthetic Aperture Radar (SAR) image despeckling, our objective is to learn a mapping from a noisy observation y to its corresponding noise-free ground truth x 0 . The Conditional Denoising Diffusion Probabilistic Model (Conditional DDPM) offers a powerful generative framework for this purpose. Its core idea is that, since the true reverse process q x t 1 | x t , y is intractable due to its dependence on the entire data distribution, the model approximates it via a learned parameterized reverse Markov chain. This chain starts from standard Gaussian noise x T N 0 , I and, guided explicitly by the condition y , progressively recovers x 0 . Each step in this reverse process is modeled as a conditional Gaussian distribution parameterized by a neural network:
p θ x t 1 | x t , y = N x t 1 ; μ θ x t , y , t , Σ θ x t , y , t
Theoretically, the true posterior q x t 1 | x t , x 0 is also Gaussian, providing justification for our parameterized form. Ho et al. [37] showed that, with a specific parameterization, the mean μ θ can be derived from a neural network ϵ θ trained to predict the noise ϵ :
μ θ x t , y , t = 1 α t x t β t 1 α ¯ t ϵ θ x t , y , t
For the predictive variance Σ θ , the original work set it as a fixed hyperparameter, such as Σ θ = β ˜ t I or Σ θ = β t I   . In SAR despeckling, subsequent studies like Perera et al. [32] and Pan et al. [34] also adopted this fixed-variance approach.
The model’s training objective is derived by optimizing the variational lower bound (VLB). As demonstrated by Ho et al. [37], this objective can be simplified to an equivalent, more intuitive loss function, whose core principle is to train a network for predicting the added noise at each timestep. For our conditional task, this loss function is formulated as:
L s i m p l e = E t , x 0 , y , ϵ ϵ ϵ θ α ¯ t x 0 + 1 α ¯ t ϵ , y , t 2

2.3. Comparison of Linear and Cosine Noise Schedules

The noise schedule β t is a critical design choice that governs model performance and training dynamics. It precisely dictates the signal decay trajectory during the forward process, thereby profoundly impacting the quality of the diffusion process and the stability of training [38]. The linear and cosine schedules are two predominant strategies, which, as depicted in Figure 2, exhibit marked differences in their control of the Signal-to-Noise Ratio (SNR). The linear schedule, originally proposed by Ho et al. [37], is defined as:
β t = β min + t 1 T 1 β max β min
where β min and β max are the minimum and maximum noise levels, respectively, typically set to β min = 10 4 and β max = 0.02 . The cumulative product for the linear schedule is given by:
α ¯ t = s = 1 t 1 β s = s = 1 t 1 β min s 1 T 1 β max β min
While the linear schedule maintains low noise levels in early timesteps, the noise escalates too rapidly in later stages. This uneven decay trajectory causes the image’s structural information to be prematurely overwhelmed by noise, posing an inherent challenge to accurate recovery. To address these limitations, Nichol [39] introduced the cosine schedule, which employs a non-linear function to achieve a smoother Signal-to-Noise Ratio (SNR) transition.
α ¯ t = f t f 0             f t = cos 2 t / T + s 1 + s π 2
Here, s denotes a small offset (typically s   =   0.008 ) to ensure numerical stability. The noise variance at each timestep is defined as:
β t = 1 α ¯ t α ¯ t 1 = 1 f t f t 1
The core advantage of this design is its nearly linear, smooth decay of signal energy, avoiding the abrupt information loss typical of linear schedules. As validated by Nichol and Dhariwal [39], models with a cosine schedule achieve performance comparable to state-of-the-art (SOTA) generative models in log-likelihood and FID scores, significantly improving sample quality. This indicates that an optimized scheduling design enhances the model’s ability to learn the underlying data distribution effectively [40].

3. Methodology

3.1. Network Architecture

The pipeline of our proposed efficient conditional diffusion model is illustrated in Figure 3. We first define a forward diffusion process q x t | x t 1 starting from a clean image x 0 . This process employs a cosine schedule to progressively inject Gaussian noise into x 0 over T discrete timesteps. At a randomly sampled timestep t , we obtain a noisy sample x t , which is then concatenated with the conditioning speckled image y along the channel dimension to form the input tensor.
During inference, the model performs the reverse denoising process, starting from pure Gaussian noise x T . Guided by the conditional image y , it employs the learned reverse transition distribution p θ x t 1 | x t , y , iteratively denoising from t = T to t = 1 . To mitigate the computational inefficiency inherent in conventional diffusion models, we introduce an efficient stochastic sampling strategy, significantly reducing the number of sampling steps and enhancing inference speed while maintaining denoising quality. Further details are provided in Section 3.3.
To implement this procedure, we design a deep neural network, as illustrated in Figure 4, based on the classic U-Net architecture with a symmetric encoder–decoder structure and a bottleneck layer. For efficient multi-scale feature extraction, the network uses a base channel dimension of 128, with encoder channel multipliers set as 1 ,   1 ,   2 ,   3 ,   4 ; the decoder mirrors this setup.
The encoder path extracts deep semantic features via cascaded feature extraction blocks (Blocks A–C). It integrates a wavelet-based downsampling module at the initial high-resolution stage (Block A), progressively mapping the input image to a lower-resolution feature space. This approach enhances high-frequency detail representation and captures multi-scale contextual information. Symmetrically, the decoder path (Blocks D–F) employs upsampling operations and skip connections to fuse fine-grained details with deep semantics, enabling precise image structure reconstruction. The network applies group normalization (GN) and Sigmoid Linear Unit (SiLU) activation functions throughout. A key innovation is the dual-branch output, simultaneously predicting the noise component ϵ θ and the variance component v θ . The predicted noise follows Equation (7), while we avoid direct variance prediction for Σ θ . Instead, the network outputs an interpolation weight vector v θ x t , y , t , interpolating in the log-domain between theoretical variance bounds:
log Σ θ x t , y , t = v θ x t , y , t log β t + 1 v θ x t , y , t log β ˜ t
where β t is the forward diffusion variance and β ˜ t is the posterior variance lower bound. This parameterization constrains predicted variances to a theoretically valid range, mitigating numerical instability associated with direct prediction.
The detailed architectures of the core components in our network—the residual block, self-attention block, and output block—are illustrated in Figure 5. The fundamental feature extraction unit (Figure 5a) is a time-conditional residual block adapted from BigGAN [41], consisting of GN, SiLU activation, and a 3 × 3 convolution. We select GN instead of Batch Normalization (BN) for its better stability in small-batch training, which is typical in the high-noise regime of diffusion models. The SiLU activation is chosen to exploit its self-gating mechanism for improved representational capacity, while its smoothness ensures superior gradient flow compared to ReLU. The timestep t is mapped to a high-dimensional embedding vector t e using sinusoidal positional encoding. This embedding is then processed and injected into each residual block to provide temporal context regarding the current diffusion stage.
A self-attention block (Figure 5b) is incorporated into the low-resolution bottleneck layers to capture global dependencies. By computing Query (Q), Key (K), and Value (V) matrices, it dynamically enhances the modeling of large-scale structures and textures. Finally, the output block (Figure 5c), comprising GN, SiLU, and a 3 × 3 convolution, is employed in each of the two output branches to predict the noise ϵ θ and variance v θ , respectively.

3.2. Wavelet-Based Downsampling Module

In our proposed network, the downsampling operation within the encoder path is essential for enlarging the receptive field and extracting multi-scale features. Conventional downsampling methods, which typically reduce spatial resolution by discarding or averaging local information, inevitably lead to the loss of fine details in SAR imagery. To mitigate this, we introduce a downsampling module predicated on the 2D Discrete Wavelet Transform (2D-DWT), applied solely during the high-resolution stages of the encoder path. The central tenet of this approach is to transform the challenging task of preserving spatial details into a more tractable problem of differentiating features in the channel domain. Specifically, it is designed to enable the network to distinguish noise-related high-frequency components from those representing terrain features, such as edges and textures. As research by [42] indicates, wavelet-based downsampling significantly enhances the capture and reconstruction of small objects and boundaries. In the context of SAR despeckling, this implies that for minute structures like buildings, well-defined agricultural boundaries, roads, and topographical edges, our wavelet-based downsampling furnishes the network with richer feature information. As depicted in Figure 6, our proposed module is composed of two primary components:
Lossless Feature Encoding Stage: In this stage, the input feature map X C × H × W undergoes DWT decomposition. We select the Haar wavelet and utilize a symmetric padding mode. This process decomposes X into one low-frequency approximation sub-band X L L (capturing macroscopic structures) and three high-frequency detail sub-bands X H L , X L H , X H H (encoding horizontal, vertical, and diagonal details, respectively). Subsequently, these four sub-bands are concatenated along the channel dimension to form an information-preserving tensor X 4 C × H 2 × W 2 . This operation, while halving the spatial resolution, transfers the entire spatial information content into the channel dimension. This decomposition method ensures the lossless nature of the feature encoding process, providing an information-rich foundational representation for subsequent multi-scale feature learning. Adaptive Feature Learning: First, the concatenated feature tensor X is fed into a 1 × 1 convolutional layer, whose core function is to adaptively learn the optimal fusion of the low-frequency overview and the multi-directional high-frequency details. Next, a GN layer is employed to stabilize training dynamics and enhance the model’s generalization capability. Finally, the SiLU activation function introduces requisite non-linearity to boost the model’s expressive power. Collectively, this sequence of operations ensures that the module outputs a robust and information-rich multi-scale feature representation, providing a high-quality input for subsequent network layers.

3.3. Efficient Stochastic Sampling

In the domain of unconditional image generation, a series of efficient acceleration paradigms have been proposed to overcome the slow iterative sampling bottleneck of the original DDPM. Among these, Denoising Diffusion Implicit Models (DDIM) [40] stand out as one of the most representative works. The core idea of DDIM is to construct a non-Markovian forward process, which enables the derivation of an equivalent training objective. This, in turn, allows setting the variance of the random term in the reverse process, σ t 2 , to zero. Consequently, the generation process becomes a fully deterministic path, enabling consistent generation from the same noise input to the same image output while substantially reducing the number of sampling steps.
However, for SAR image despeckling, we argue that directly applying DDIM’s deterministic sampling path is suboptimal. A purely deterministic recovery path lacks stochastic exploration capability. When processing complex SAR images, this limitation can result in over-smoothing, loss of critical textures, and edge artifacts, as it fails to flexibly handle the intricate stochastic nature of speckle noise.
Inspired by the “learnable variance” concept introduced by Nichol et al. [39] in generative tasks, we propose an adaptive stochastic denoising path. In our framework, the reverse process variance is dynamically learned by the model from the data, rather than being a fixed hyperparameter. We formulate SAR image despeckling as a conditional generation task, centered on a noise prediction network ϵ θ x t , y , t conditioned on the noisy image y . Unlike standard models, our network not only predicts the noise ϵ θ but also jointly predicts a variance modulation parameter v θ . Our accelerated sampling strategy is implemented as follows:
For a model trained over T steps, we perform inference using a sparse timestep subsequence S = τ K , τ K 1 , , τ 1 of length K , where T τ K > τ K 1 > > τ 1 > 0 , instead of the full sequence. This subsequence is generated by uniform interval sampling across 1 , T , ensuring representative coverage of the entire diffusion process. For any timestep τ i in S , its cumulative signal coefficient α ¯ τ i is directly indexed from the pre-computed sequence { α ¯ t } t = 1 T of the full T step process. This indexing ensures that the τ i -th step on the sparse path shares the exact Signal-to-Noise Ratio (SNR) characteristics as the τ i -th step on the full path. Based on this, to achieve an effective transition from x τ i to x τ i 1 , we re-derive the single-step transition variance β τ i and the posterior variance β ˜ τ i :
β τ i = 1 α ¯ τ i α ¯ τ i 1 ,         β ˜ τ i = 1 α ¯ τ i 1 1 α ¯ τ i β τ i
The above definition ensures that each step along the sparse sampling trajectory strictly adheres to the statistical properties of the original diffusion process. Consequently, the single-step reverse transition from a state x τ i to its predecessor x τ i 1 is implemented by the following reparameterization formula:
x τ i 1 = 1 α τ i x τ i 1 α τ i 1 α ¯ τ i ϵ θ x τ i , y , τ i + σ τ i z
This transition comprises two components: a deterministic denoising term predicted by the noise estimation network ϵ θ x τ i , y , τ i , and a stochastic noise term σ τ i z , with z N 0 , I representing standard Gaussian noise. The crux of our method lies in dynamically and adaptively modeling the variance σ τ i 2 . Specifically, we parameterize it as a covariance matrix Σ θ x τ i , y , τ i , predicted directly by the network as:
Σ θ x τ i , y , τ i = exp v θ x τ i , y , τ i log β τ i + 1 v θ x τ i , y , τ i log β ˜ τ i
This covariance matrix is dynamically interpolated between the theoretical upper bound β τ i and lower bound β ˜ τ i of the variance at timestep τ i . This interpolation is guided by the variance interpolation weight v θ x τ i , y , τ i , which is predicted concurrently by the network. This adaptive variance mechanism imparts considerable flexibility to the model.
The two acceleration strategies differ fundamentally in their sampling paradigms. DDIM achieves acceleration by introducing a non-Markovian diffusion process, allowing deterministic generative paths and meaningful semantic interpolation in latent space. In contrast, our proposed method retains the intrinsic stochasticity of the Markovian diffusion process, achieving efficiency through adaptive variance prediction and cosine timestep scheduling. This preserves sampling randomness, which is particularly beneficial for addressing complex speckle noise degradation and maintaining structural fidelity in SAR imagery.

3.4. Loss Function

To achieve high-fidelity restoration of SAR images, we formulate a hybrid objective function for our proposed conditional diffusion model. The total objective function, L total , is defined as a weighted sum of two loss terms:
L total = L simple + λ vlb L vlb
Here, L simple is the primary noise prediction loss, L vlb is the variational lower bound (VLB) loss, and λ vlb is a weighting coefficient that balances their contributions. L simple guides ϵ θ to predict the Gaussian noise ϵ at timestep t using MSE loss.
L simple = E t , x 0 , ϵ , y ϵ ϵ θ x t , t , y 2
In this equation, t 1 , , T is a uniformly sampled timestep and ϵ N 0 , I is standard Gaussian noise. To enable the model to learn the reverse process variance Σ θ x t , t , y , we introduce the variational lower bound loss term L vlb . This term is formulated by minimizing the KL divergence between the “true” reverse posterior distribution q x t 1 | x t , x 0 and the model-parameterized distribution p θ x t 1 | x t , y :
L vlb = E t , x 0 , y D KL q x t 1 | x t , x 0 p θ x t 1 | x t , y
The magnitude of the VLB loss, L vlb , is intrinsically tied to the discretization granularity of the diffusion process (i.e., the total number of steps T ). To mitigate potential gradient imbalance, we propose a scale-normalized weighting strategy:
λ vlb = T 1000
This strategy linearly scales the weight L vlb with T , ensuring stable gradient contributions from L vlb across different T configurations and facilitating robust variance learning. For the T = 1000 setup used in our study, this yields λ vlb = 1.0 .

4. Experiments and Results

4.1. Datasets

To enhance the model’s adaptability to diverse terrain types and complex scenarios, we constructed the Multi-Scene SAR Synthetic Dataset (MS-SAR) by selecting representative scenes—urban, mountainous, and agricultural—from the DSIFN [43] and Sentinel-2 [44] optical datasets. We generated noise samples corresponding to different ENLs ( L   =   1 ,   2 ,   4 ) to simulate varied imaging conditions. Each sample pair consists of a noisy SAR-simulated image (input) and a clean optical image (ground truth), with all patches uniformly cropped to 256 × 256 pixels. The dataset is partitioned into training (3000 pairs), validation (200 pairs), and test sets (100 pairs), ensuring balanced representation across all representative scenes in each subset.
Despite the convenience of optical-based synthetic data for supervised training of SAR despeckling models, a fundamental limitation remains: the domain gap between optical and real SAR data. This gap arises from fundamentally different imaging physics—optical imaging relies on passive spectral reflectance, while SAR relies on active microwave backscattering—resulting in systematic biases in texture and structural representations between synthetic and real SAR data [45]. To bridge this gap, we utilized a multi-temporal real SAR dataset for fine-tuning, derived from Sentinel-1 C-band Ground Range Detected (GRD) products. This dataset includes 10 co-registered VV-polarized Interferometric Wide (IW) swath images covering the Toronto region in Canada, with a spatial resolution of 5   m × 20   m . A pseudo-noise-free reference image was generated via multi-temporal averaging, using a single-date image (5 September 2022) as the noisy input paired with the fused reference as the supervised ground truth [46]. This pair was cropped into 1000 256 × 256 patches for training. Additionally, we incorporated Sentinel-1A Single Look Complex (SLC) products for testing, encompassing diverse terrains including urban, mountainous, rural, and agricultural areas in Beijing, China, and Toronto, Canada. These data were acquired in IW mode with VV polarization and a spatial resolution of 3   m × 15   m .

4.2. Experiments Preparation

All deep learning experiments in this study were conducted on an NVIDIA A800 80 GB PCIe GPU, running on the Ubuntu operating system. The implementations were developed in Python 3.8 using the PyTorch 1.8.0 deep learning framework. For model training, we utilized the AdamW optimizer with momentum coefficients β 1 = 0.9 and β 2 = 0.999 . The initial learning rate was set to η 0 = 5 × 10 5 , with the learning rate scheduled using the Cosine Annealing with Warm Restarts strategy. The initial restart period was T 0 = 20 epochs, with subsequent periods scaled by a multiplicative factor T mult = 2 , and the minimum learning rate set to η min = 1 × 10 6 . Training was conducted for a total of 400 epochs. The diffusion process was configured with 1000 timesteps, employing a cosine noise schedule where the noise variance β t ranged from 1 × 10 6 to 1 × 10 2 . The model was first trained for 300 epochs on a synthetic dataset derived from optical images, using a batch size of 16. Subsequently, fine-tuning was performed for 100 epochs on a multi-temporal SAR dataset constructed from real SAR images, with a batch size of 8. During inference, we applied the same cosine noise schedule and used 50 sampling timesteps to generate the final results of the proposed method.

4.3. Methods for Comparison

To validate the reliability and effectiveness of the proposed method, we conducted comprehensive comparisons with seven established SAR image despeckling methods: PPB, BM3D, SAR-CNN, SAR-DRN, SAR-Trans, SAR-ON [47], SAR-CAM, and SAR-DDPM. Among these, PPB and BM3D are classical non-local approaches. For PPB, we used a patch size of 7 × 7 pixels and a search range of 21 × 21 pixels. For BM3D, block-matching was performed with a block size of 12 × 12 pixels and a stride of 2 pixels, while the search window for similar blocks was set to 39 × 39 pixels. The remaining methods are neural network-based. SAR-CNN, SAR-DRN, and SAR-ON are based on convolutional neural networks (CNNs). SAR-CAM employs successive attention blocks concatenated to progressively extract features. SAR-Trans utilizes a hierarchical encoder based on the Pyramid Vision Transformer architecture for multi-scale feature extraction. In contrast to SAR-DDPM, which relies on a standard 1000-timestep sampling procedure, our proposed method achieves significant computational efficiency gains by reducing the sampling steps to only 50 during inference, while preserving superior despeckling performance.

4.4. Experimental Results on Synthetic Datasets

To comprehensively evaluate the performance of the proposed method, detailed experimental verification was conducted on the MS-SAR dataset, encompassing three typical land-cover scenes (urban, mountainous, and farmland) under different numbers of looks ( L   =   1 ,   2 ,   4 ). The peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) were employed as objective evaluation metrics. Specifically, PSNR quantitatively assesses despeckling performance by calculating the mean squared error between reconstructed and reference images, expressed in decibels (dB), where higher values indicate superior denoising quality. SSIM comprehensively evaluates luminance, contrast, and structural similarity, ranging from 0 to 1, with values closer to 1 indicating higher structural fidelity. Additionally, qualitative visual analyses were performed to verify the balance achieved by each method between detail preservation and noise suppression.
Table 1 presents the quantitative evaluation results on the MS-SAR dataset, providing a comprehensive comparative analysis with seven state-of-the-art SAR despeckling algorithms. Optimal results are highlighted in bold. Traditional algorithms (PPB and BM3D) exhibit significantly inferior performance under single-look conditions. Among the deep learning-based methods, SAR-Trans performs notably well in farmland scenes but shows limited adaptability to complex structural textures, indicating strong scene dependency. SAR-ON and SAR-CAM demonstrate relatively balanced cross-scene generalization capability, occasionally surpassing the baseline diffusion model SAR-DDPM under certain conditions. In contrast, the proposed method consistently achieves the best performance across all test scenarios and look-number conditions, delivering average performance gains of 0.67–1.43 dB in PSNR and 0.01–0.06 in SSIM compared to the second-best baseline model.
The imaging results under single-look conditions are shown in Figure 7, where the original images are severely degraded by speckle noise. While the traditional algorithms PPB and SAR-BM3D exhibit some noise suppression capabilities, they suffer from over-smoothing and detail loss, respectively. Among the deep learning methods, SAR-DRN shows prominent residual noise artifacts; SAR-Trans, SAR-CAM, and SAR-ON display mild blurring in complex textured regions; and SAR-DDPM demonstrates strong edge preservation and noise suppression but falls short in detail recovery and structural fidelity. In contrast, our proposed method achieves superior noise suppression while effectively preserving edge information and textural details. The results for 2-look processing are shown in Figure 8. Although PPB and BM3D maintain adequate noise suppression, the over-smoothing issue persists, resulting in loss of structural details. SAR-DRN, SAR-Trans, and SAR-ON exhibit limited recovery capabilities in complex urban areas. Our proposed method attains optimal noise suppression alongside superior restoration of urban structures. Under the 4-look conditions illustrated in Figure 9, the noise level is further reduced, and all methods show markedly improved despeckling performance, though differences remain evident. SAR-DRN and SAR-CAM produce noticeable artifacts in densely built-up areas, degrading image quality. SAR-Trans and SAR-ON exhibit over-smoothing, leading to partial loss of structural details. While SAR-DDPM offers reasonable noise suppression, it is affected by intensity biases. Our proposed method effectively suppresses noise while optimally preserving building contours, demonstrating exceptional structural fidelity.
In summary, the comparative results against traditional methods and multiple baselines indicate that our method holds significant advantages in edge preservation, structural integrity, and recovery of complex textural details.

4.5. Experimental Results on Real SAR Datasets

To further validate the effectiveness and generalization capability of the proposed method in real-world scenarios, both the proposed and comparative methods were evaluated on various scenes within a real SAR image test set. This section presents the experimental results for farmland, urban, and mountainous areas, aiming to assess each algorithm’s despeckling performance across homogeneous, heterogeneous, and strongly heterogeneous regions in SAR images. Evaluations include both subjective visual assessments and objective quantitative metrics.
The experimental evaluation was conducted on a complex SAR urban image encompassing building clusters, road networks, and homogeneous areas, with despeckling results visualized in Figure 10. PPB provides insufficient noise suppression in urban scenes, retaining noticeable granular artifacts, as evidenced by substantial textural residues in its ratio image; SAR-BM3D induces over-smoothing, severely blurring architectural contours and road boundaries. Deep learning-based methods, while improved, retain shortcomings: SAR-DRN and SAR-Trans enhance background noise suppression but introduce subtle halo artifacts, with their residual maps showing evident edge remnants; SAR-ON and SAR-CAM perform well in strong scattering regions yet inadequately preserve fine details; SAR-DDPM demonstrates robust denoising but generates textural artifacts in homogeneous areas, with its ratio image revealing mismatches between the generated prior and actual speckle statistics. In contrast, our proposed method achieves superior despeckling across complex urban road networks and homogeneous regions, while faithfully preserving key geometric structures and radiometric characteristics.
The experimental results for the mountainous scene are presented in Figure 11. This scene is characterized by rich topographic details, including steep ridges and smooth slope faces. Consistent with the findings in urban areas, traditional filtering algorithms struggle to achieve an optimal balance between noise suppression and structural fidelity. Specifically, SAR-DRN, SAR-Trans, and SAR-ON exhibit over-smoothing in extensive slope areas, resulting in markedly softened ridge lines, reduced contrast in subtle gullies, and diminished topographic relief. Their corresponding residual maps reveal prominent ridge contours, indicating a systematic radiometric bias in the signal-noise separation process, whereby essential topographic gradient information is mistakenly categorized as noise. Although SAR-CAM and SAR-DDPM yield relatively superior despeckling outcomes, their residual maps still exhibit faint but discernible structural leakage. In stark contrast, our method’s superiority is evident: it preserves key topographical features without artifacts, leaving a residual map that is virtually free of structural leakage.
The farmland scene results in Figure 12 clearly reveal the inherent limitations of conventional methods: the PPB algorithm leaves significant granular artifacts in the processed image, while SAR-BM3D causes severe over-smoothing. CNN-, attention-, and Transformer-based models all fail to effectively decouple the clean signal from the observed speckle, with their residual maps showing prevalent structural remnants—particularly linear traces along farmland boundaries—manifesting as varying degrees of over-smoothing. As the baseline, SAR-DDPM introduces pronounced brightness shifting in homogeneous regions, with its ratio map exhibiting significant structural remnants that confirm systematic bias in the despeckling process, severely compromising radiometric reliability. In contrast, our proposed method delivers artifact-free smoothing in homogeneous regions while preserving sharp edges and strong scatterers. Its ratio map most closely approximates the ideal random speckle field among deep learning approaches, providing compelling evidence of its superior despeckling performance.
To objectively quantify the qualitative observations, we performed quantitative evaluations on despeckling results across three typical scenarios, with the best method in red bold and the second-best in red italics, as shown in Table 2. Due to the lack of noise-free references for real SAR images, we adopted a four-dimensional no-reference framework assessing performance in speckle suppression, radiometric fidelity, edge preservation, and overall quality. The Equivalent Number of Looks (ENL) quantifies speckle suppression via the squared mean-to-variance ratio in homogeneous regions [48]; higher ENL indicates better suppression. Radiometric fidelity is assessed using Mean of Image (MOI) and Mean of Ratio (MOR) [49]: MOI evaluates bias via pre- and post-despeckling mean ratios in homogeneous regions, while MOR detects distortion through ratio image means. We analyzed 5 regions of interest per image; ideally, MOI and MOR approach 1, with deviations signaling bias [50]. The Edge Preservation Degree based on Ratio of Averages (EPD-ROA) [51] measures structural consistency via local mean ratios pre- and post-filtering, including vertical and horizontal components; values near 1 denote better edge/detail preservation. The M-index [52] combines first-order (speckle suppression, mean preservation) and second-order (structure preservation) residuals for a holistic score; values near 0 indicate optimal balance between suppression and preservation.
Analysis of the quantitative results from Table 2 yields the following conclusions:
Urban Scene: In the urban scene, characterized by the highest structural complexity, our proposed method attains the lowest M-index (2.3356), underscoring its superior overall performance. This aligns closely with the visual observations of sharp building contours and clear road boundaries. Regarding radiometric fidelity, our method’s MOI (0.9832) and MOR (1.0104) are remarkably close to the ideal value of 1, quantitatively confirming its efficacy in alleviating radiometric distortions common in methods like SAR-DDPM and SAR-Trans. Although SAR-CAM achieves the highest ENL (206.76), indicating strong smoothing capability, its suboptimal EPD-ROA score reflects a compromise in structural detail preservation.
Mountainous Scene: In the mountainous scene, SAR-Trans secures the second-best ENL and EPD-ROA, yet its M-index is outperformed by our method. Our approach achieves the optimal M-index (2.8668) and EPD-ROA, demonstrating adaptability to heterogeneous terrains. Moreover, it exhibits robust radiometric fidelity (MOI = 0.9789, MOR = 1.0104), surpassing the notable biases in baselines such as SAR-DDPM (MOI = 1.0764).
Farmland Scene: In the farmland scene, dominated by large homogeneous regions, the key differentiators among methods are radiometric fidelity and overall performance. While SAR-ON and SAR-BM3D marginally surpass our method in ENL, our approach excels in other metrics and secures the lowest M-index (2.986). Additionally, its MOI (1.0126) and MOR (0.9885) closely approximate the ideal value of 1, mitigating the radiometric shifts prevalent in baselines like SAR-DDPM.
In conclusion, a comprehensive quantitative evaluation across three typical land-cover types reveals that our proposed method consistently achieves optimal or near-optimal performance on key metrics. This not only quantitatively validates the method’s robustness but also elucidates its core advantage: efficiently suppressing speckle noise in SAR images while preserving the fidelity of complex structures (e.g., edges and textures) and maintaining radiometric accuracy.

5. Discussion

5.1. Ablation Study

To validate the effectiveness of key components in our diffusion-based network, we conducted ablation studies to quantitatively assess the contributions of the wavelet-based downsampling module and joint noise-variance prediction. Forty images with L   =   1 ,   2 ,   4 were randomly selected from the MS-SAR test set, with results summarized in Table 3.
To assess the joint prediction strategy ( ϵ θ + v θ ), we compared Experiment 1 (predicting only ϵ θ ) with Experiment 2. Across degradation levels, Experiment 2′s PSNR and SSIM consistently exceed Experiment 1′s, confirming that simultaneous v_\theta prediction improves reconstruction stability and precision.
To verify the wavelet-based module’s contribution, we compared Experiment 3 (Haar wavelet downsampling) against Experiment 2 (conventional strided convolution). Experiment 3 outperforms Experiment 2 across all noise levels, validating that our lossless module preserves high-frequency information for decoupled multi-scale representations, enhancing despeckling performance and structural fidelity. Inference times between Experiments 3 and 2 show no perceptible overhead, confirming the module’s efficiency and plug-and-play nature.
Comparing Experiments 3–5 reveals the Haar wavelet’s superior performance across all look conditions over db4 and bior2.2. Haar’s advantages include integer coefficients preventing floating-point errors, minimal support length (L = 2) reducing boundary effects, and strict orthogonality ensuring energy conservation—aligning well with diffusion models’ probabilistic framework [53]. This suggests prioritizing numerical compatibility over approximation accuracy when selecting wavelets for deep generative models.
For comprehensive evaluation, we conducted visual quality assessments on the real SAR test dataset (Figure 13). Comparing baseline (Experiment 1) with joint prediction (Experiment 2) shows improvements in building contours and strong scatterer fidelity; rural ridge and road boundaries also exhibit greater clarity, verifying the strategy’s effectiveness. Introducing wavelet downsampling (Experiments 3–5) further enhances despeckling quality and edge preservation. Relative to minor edge-rounding or blurring in Daubechies-4 (Experiment 4) and Biorthogonal-2.2 (Experiment 5), Haar (Experiment 3) excels in reconstructing orthogonal buildings, linear features, and subtle terrain boundaries in rural scenes.
In summary, the visual evidence aligns with Table 3’s quantitative analyses, demonstrating our method’s state-of-the-art performance in efficient SAR despeckling with superior structural fidelity.

5.2. Sampling Strategy Analysis

Although iterative sampling is foundational to diffusion models for achieving high-fidelity generation, the typical reverse sampling process involving thousands of steps (T = 1000) incurs substantial inference latency. This section systematically evaluates the denoising performance of our proposed efficient reverse sampling strategy against DDIM across various sampling steps. We conducted rigorous quantitative and qualitative experiments to elucidate the fundamental differences in efficiency, performance, and robustness among different sampling strategies.
Table 4 illustrates a detailed comparison of inference performance across multiple sampling steps (t) for different samplers. Experimental results clearly demonstrate that our proposed method effectively overcomes the inherent trade-off between inference efficiency and generation quality in SAR image denoising tasks. Compared to the DDPM baseline employing the complete 1000-step sampling process (taking 34.69 s), our method dramatically reduces inference time to 0.87 s at just 25 sampling steps (t = 25), achieving nearly a 40-fold theoretical acceleration. Crucially, despite this significant acceleration, performance metrics were not compromised but instead surpassed those of the baseline model.
Moreover, in horizontal comparisons against the widely adopted accelerated sampler DDIM, our method consistently exhibited comprehensive and stable performance advantages across all evaluated sampling steps under comparable computational costs. Furthermore, our proposed method demonstrated exceptional robustness across a broad and practically significant acceleration interval (t = 250 to t = 25), maintaining superior performance metrics in stark contrast to DDIM.
To complement the quantitative analysis, Figure 14 and Figure 15 visually illustrate the despeckling results of the two acceleration strategies applied to synthetic and real SAR images under varying sampling steps.
For synthetic SAR images, the DDIM sampling strategy exhibits inherent instability and pronounced sensitivity to sampling step selection. At longer sampling intervals, DDIM demonstrates inadequate denoising capability, resulting in catastrophic noise residues that completely obscure meaningful scene reconstruction. While shorter sampling intervals (e.g., T 100 ) provide improved speckle suppression and structural preservation, they introduce significant over-smoothing, leading to the loss of high-frequency details and textures despite effective noise reduction. This manifests as blurred boundaries and reduced realism. In contrast, our proposed sampling method displays exceptional robustness across a broad range of sampling intervals, from 750 down to 25 steps. The outputs consistently preserve high fidelity and structural coherence, with precise reconstruction of building contours and fine textures. Even at very short intervals, no discernible degradation in image details occurs.
Visual comparisons on real SAR images further highlight the substantial performance disparity between the two strategies. The DDIM sampler fails to yield satisfactory results across all tested intervals. At higher sampling steps ( T = 750 , 500 ), DDIM’s rigid deterministic trajectory inadequately models the unique statistical properties of SAR speckle noise, producing pervasive Gaussian-like residual artifacts. At intermediate intervals ( T = 250 , 100 , 50 ), although speckle is partially suppressed, catastrophic radiometric distortions emerge, dramatically shifting overall brightness and contrast and thus compromising radiometric fidelity. These distortions intensify at extremely low intervals, resulting in severe structural detail loss and dominant artifacts in highlighted regions.
In stark contrast, our proposed Efficient Stochastic Sampling strategy exhibits exceptional and consistent performance. At higher and intermediate intervals, it effectively suppresses speckle noise while achieving high-fidelity reconstruction of strong scatterers and fine structures, all while maintaining radiometric consistency. Even at low intervals ( T = 25 , 10 ), the results uphold superior structural integrity, with radiometric shifts only appearing under the most extreme condition ( T = 5 ).
In summary, this ablation study systematically demonstrates—from both qualitative visual inspections and quantitative metrics—the fundamental superiority of our proposed sampler over existing approaches. Its success stems from a nuanced understanding and modeling of the reverse sampling process. DDIM’s performance bottleneck arises from its deterministic sampling trajectory, whose rigidity causes unpredictable degradation under varying acceleration ratios, making it unsuitable for complex SAR despeckling tasks. Conversely, our method’s core strength lies in its dynamic, adaptive sampling process. By incorporating learnable variance prediction, the sampler dynamically modulates denoising intensity and direction at each step, thereby averting extremes such as residual noise, over-smoothing, radiometric distortions, and structural detail loss. This ensures stable, high-quality outputs across the full acceleration range.

6. Conclusions

In this paper, we propose an Efficient Conditional Diffusion Model framework. We integrate the wavelet transform into the encoder’s downsampling path. Through lossless frequency-domain decomposition, this converts the spatial detail preservation problem into channel-wise feature discrimination, substantially enhancing feature extraction capabilities. To tackle inference bottlenecks, we propose an efficient stochastic sampling strategy that departs from the deterministic paths of methods like DDIM. Instead, the network jointly predicts the noise component and variance modulation parameter, enabling dynamic interpolation between theoretical variance bounds. This empowers the model to flexibly adapt the recovery trajectory, achieving a 20-fold speedup in inference (from 34.7 s to 1.7 s). Regarding the training paradigm, we adopt a “pre-training on synthetic data followed by fine-tuning on real data” strategy, effectively bridging the domain gap between simulated and real SAR data while ensuring robust generalization. Comprehensive quantitative and qualitative experiments on synthetic datasets and real SAR images validate the superiority of our method. These results underscore the model’s exceptional balance between noise suppression and structural fidelity, while its high inference efficiency provides a solid foundation for high-precision downstream tasks (e.g., object recognition, land cover classification) and resource-constrained edge deployments.
Despite these advancements, our method has limitations. It currently focuses solely on the magnitude domain of SAR images, overlooking phase information, which may constrain performance in complex coherent scenarios. Furthermore, the experimental validation in this study was conducted primarily on C-band Sentinel-1 data. While this provides a robust benchmark, exploring the model’s adaptability and performance on other frequency bands, such as X-band and L-band, and across different polarizations would be a valuable direction for future research. We believe our proposed framework is general enough to be extended to these scenarios, which will be a key focus of our subsequent work. Future work could also incorporate even more advanced ODE/SDE solvers [54] like DPM-Solver++ [55] and explore emerging paradigms like Consistency Models [56] to further optimize denoising performance and efficiency. Additionally, lightweight deployment will be a key focus to advance practical applications.

Author Contributions

Conceptualization, W.H. and Z.G.; methodology, Z.G.; software, Z.G.; validation, Z.G., S.Z. and B.Z.; formal analysis, B.Z. and S.Z.; investigation, Z.G. and J.P.; resources, W.H. and M.Z.; data curation, S.Z. and W.H.; writing—original draft preparation, Z.G.; writing—review and editing, S.Z., B.Z.; visualization, M.F. and Z.G.; supervision, Z.Y. and S.Z.; project administration, W.H. and M.Z.; funding acquisition, W.H. and M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62401314, and in part by the Young Elite Scientists Sponsorship Program by China Association for Science and Technology (CAST) under Grant 2023QNRC001.

Data Availability Statement

The original Sentinel-1 SAR satellite data used in this study are publicly available from the European Space Agency (ESA) Copernicus Open Access Hub (https://esar-ds.eo.esa.int/oads/access/ accessed on 15 July 2025).

Acknowledgments

This research work is Supported by the High Performance Computing Center, Yangtze Delta Region Academy of Beijing Institute of Technology, Jiaxing.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SARSynthetic Aperture Radar
SNRsignal-to-noise ratio
ENLEquivalent Number of Looks
PDFProbability Density Function
PPBthe probabilistic patch-based
SARBM3DSAR block-matching three-dimensional
CNNConvolutional Neural Network
SAR-DRNSAR Dilated Residual Network
SAR-TransformerTransformer based SAR
SAR-ONSAR overcomplete convolutional networks
SAR-CAMSAR Continuous Attention Module
DDPMDenoising Diffusion Probabilistic Model
GANGenerative Adversarial Network
SAR-DDPMSAR Despeckling Using a Denoising Diffusion Probabilistic Model
GNGroup Normalization
SiLUSigmoid Linear Unit
SLCSingle Look Complex
KLKullback–Leibler
MS-SARMulti-Scene SAR Synthesis Dataset

References

  1. Moreira, J. Improved Multi Look Techniques Applied to SAR and Scan SAR Imagery. In Proceedings of the International Geoscience and Remote Sensing Symposium IGARSS 90, College Park, MD, USA, 20–24 May 1990. [Google Scholar]
  2. Singh, P.; Shree, R. Analysis and Effects of Speckle Noise in SAR Images. In Proceedings of the 2016 2nd International Conference on Advances in Computing, Communication, & Automation (ICACCA) (Fall), Bareilly, India, 30 September–1 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–5. [Google Scholar]
  3. Rana, V.K.; Suryanarayana, T.M.V. Evaluation of SAR Speckle Filter Technique for Inundation Mapping. Remote Sens. Appl. Soc. Environ. 2019, 16, 100271. [Google Scholar] [CrossRef]
  4. Palacio, M.G.; Ferrero, S.B.; Frery, A.C. Information Content in SAR Images: A Classification Accuracy Viewpoint. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 823–826. [Google Scholar] [CrossRef]
  5. Aubert, G.; Aujol, J.-F. A variational approach to removing multiplicative noise. SIAM J. Appl. Math. 2008, 68, 925–946. [Google Scholar] [CrossRef]
  6. Lee, J.S. Digital Image Enhancement and Noise Filtering by Use of Local Statistics. IEEE Trans. Pattern Anal. Mach. Intell. 2009, PAMI–2, 165–168. [Google Scholar] [CrossRef] [PubMed]
  7. Frost, V.S.; Stiles, J.A.; Shanmugan, K.S.; Holtzman, J.C. A Model for Radar Images and Its Application to Adaptive Digital Filtering of Multiplicative Noise. IEEE Trans. Pattern Anal. Mach. Intell. 1982, 4, 157–166. [Google Scholar] [CrossRef] [PubMed]
  8. Kuan, D.T.; Sawchuk, A.A.; Strand, T.C.; Chavel, P. Adaptive Noise Smoothing Filter for Images with Signal-Dependent Noise. IEEE Trans. Pattern Anal. Mach. Intell. 1985, 7, 165–177. [Google Scholar] [CrossRef]
  9. Lopes, A.; Nezry, E.; Touzi, R.; Laur, H. Maximum a Posteriori Speckle Filtering and First Order Texture Models in Sar Images. In Proceedings of the International Geoscience & Remote Sensing Symposium, College Park, MD, USA, 20–24 May 1990. [Google Scholar]
  10. Xie, H.; Pierce, L.E.; Ulaby, F.T. SAR speckle reduction using wavelet denoising and Markov random field modeling. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2196–2212. [Google Scholar] [CrossRef]
  11. Argenti, F.; Alparone, L. Speckle removal from SAR images in the undecimated wavelet domain. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2363–2374. [Google Scholar] [CrossRef]
  12. Zhang, J.; Li, W.; Li, Y. SAR Image Despeckling Using Multiconnection Network Incorporating Wavelet Features. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1363–1367. [Google Scholar] [CrossRef]
  13. Gleich, D.; Datcu, M. Gauss–Markov Model for Wavelet-Based SAR Image Despeckling. IEEE Signal Process. Lett. 2006, 13, 365–368. [Google Scholar] [CrossRef]
  14. Jose, N.; Ramesh, R. Patch Ordering Based SAR Image Despeckling via SSC and Wavelet Thresholding. In Proceedings of the 2016 Fifth International Conference on Recent Trends in Information Technology (ICRTIT), Chennai, India, 8–9 April 2016; pp. 1–6. [Google Scholar] [CrossRef]
  15. Solbo, S.; Eltoft, T. A Stationary Wavelet-Domain Wiener Filter for Correlated Speckle. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1219–1230. [Google Scholar] [CrossRef]
  16. Buades, A.; Coll, B.; Morel, J.M. A Non-Local Algorithm for Image Denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 60–65. [Google Scholar] [CrossRef]
  17. Deledalle, C.A.; Denis, L.; Tupin, F. Iterative Weighted Maximum Likelihood Denoising with Probabilistic Patch-Based Weights. IEEE Trans. Image Process. 2009, 18, 2661–2672. [Google Scholar] [CrossRef]
  18. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A Nonlocal SAR Image Denoising Algorithm Based on LLMMSE Wavelet Shrinkage. IEEE Trans. Geosci. Remote Sens. 2012, 50, 606–616. [Google Scholar] [CrossRef]
  19. Cozzolino, D.; Parrilli, S.; Scarpa, G.; Poggi, G. Fast Adaptive Nonlocal SAR Despeckling. IEEE Geoence Remote Sens. Lett. 2013, 11, 524–528. [Google Scholar] [CrossRef]
  20. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep Learning on Image Denoising: An overview. Neural. Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef]
  21. Chierchia, G.; Cozzolino, D.; Poggi, G.; Verdoliva, L. SAR Image Despeckling through Convolutional Neural Networks. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 5438–5441. [Google Scholar] [CrossRef]
  22. Wang, P.; Zhang, H.; Patel, V.M. SAR Image Despeckling Using a Convolutional Neural Network. IEEE Signal Process. Lett. 2017, 24, 1763–1767. [Google Scholar] [CrossRef]
  23. Zhang, Q.; Yuan, Q.; Li, J.; Yang, Z.; Ma, X. Learning a Dilated Residual Network for SAR Image Despeckling. Remote Sens. 2018, 10, 196. [Google Scholar] [CrossRef]
  24. Vitale, S.; Ferraioli, G.; Pascazio, V. Multi-objective CNN-based algorithm for SAR despeckling. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9336–9349. [Google Scholar] [CrossRef]
  25. Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  26. Perera, M.V.; Bandara, W.G.C.; Valanarasu, J.M.J.; Patel, V.M. Transformer-based SAR image despeckling. In Proceedings of the IGARSS 2022–2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 751–754. [Google Scholar]
  27. Ko, J.; Lee, S. SAR image despeckling using continuous attention module. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2022, 15, 3–19. [Google Scholar] [CrossRef]
  28. Dalsasso, E.; Denis, L.; Tupin, F. SAR2SAR: A semi-supervised despeckling algorithm for SAR images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4321–4329. [Google Scholar] [CrossRef]
  29. Molini, A.B.; Valsesia, D.; Fracastoro, G.; Magli, E. Speckle2Void: Deep Self-Supervised SAR Despeckling with Blind-Spot Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5204017. [Google Scholar] [CrossRef]
  30. Vitale, S.; Ferraioli, G.; Pascazio, V. Analysis on the building of training dataset for deep learning SAR despeckling. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  31. Liu, R.; Li, Y.; Jiao, L. SAR Image Specle Reduction based on a Generative Adversarial Network. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020. [Google Scholar]
  32. Perera, M.V.; Bandara, W.G.C.; Valanarasu, J.M.J.; Patel, V.M. SAR Despeckling Using a Denoising Diffusion Probabilistic Model. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  33. Ma, Y.; Ke, P.; Aghababaei, H.; Chang, L.; We, J. Despeckling SAR Images with Log-Yeo–Johnson Transformation and Conditional Diffusion Models. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–17. [Google Scholar] [CrossRef]
  34. Pan, X.; Wang, Z.; Yu, W. SAR Image Despeckling Based on Denoising Diffusion Probabilistic Model and Swin Transformer. Remote Sens. 2024, 16, 3222. [Google Scholar] [CrossRef]
  35. Patel, V.M.; Easley, G.R.; Chellappa, R.; Nasrabadi, N.M. Separated Component-Based Restoration of Speckled SAR Images. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1019–1029. [Google Scholar] [CrossRef]
  36. Di Martino, G.; Poderico, M.; Poggi, G.; Riccio, D.; Verdoliva, L. Benchmarking framework for SAR despeckling. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1596–1615. [Google Scholar] [CrossRef]
  37. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
  38. Dhariwal, P.; Nichol, A. Diffusion models beat GANs on image synthesis. In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021), Virtual Event, 6–14 December 2021; Volume 34, pp. 8780–8794. [Google Scholar]
  39. Nichol, A.; Dhariwal, P. Improved denoising diffusion probabilistic models. In Proceedings of the 38th International Conference on Machine Learning (ICML 2021), Virtual Event, 18–24 July 2021. [Google Scholar]
  40. Song, J.; Meng, C.; Ermon, S. Denoising diffusion implicit models. In Proceedings of the 9th International Conference on Learning Representations (ICLR 2021), Virtual Event, 3–7 May 2021. [Google Scholar]
  41. Brock, A.; Donahue, J.; Simonyan, K. Large scale GAN training for high fidelity natural image synthesis. arXiv 2018, arXiv:1809.11096. [Google Scholar]
  42. Xu, G.; Liao, W.; Zhang, X.; Li, C.; He, X.; Wu, X. Haar wavelet downsampling: A simple but effective downsampling module for semantic segmentation. Pattern Recognit. 2023, 143, 109819. [Google Scholar] [CrossRef]
  43. Zhang, C.; Yue, P.; Tapete, D.; Jiang, L.; Shangguan, B.; Huang, L.; Liu, G. A deeply supervised image fusion network for change detection in high resolution bi-temporal remote sensing images. ISPRS J. Photogramm. Remote Sens. 2020, 166, 183–200. [Google Scholar] [CrossRef]
  44. Tiwari, P. Sentinel-1&2 Image Pairs (SAR & Optical). Kaggle 2021. Available online: https://www.kaggle.com/datasets/requiemonk/sentinel12-image-pairs-segregated-by-terrain (accessed on 20 July 2025).
  45. Pongrac, B.; Gleich, D. Despeckling of SAR Images Using Residual Twin CNN and Multi-Resolution Attention Mechanism. Remote Sens. 2023, 15, 3698. [Google Scholar] [CrossRef]
  46. Vásquez-Salazar, R.D.; Cardona-Mesa, A.A.; Gómez, L.; Travieso-González, C.M.; Garavito-González, A.F.; Vásquez-Cano, E. Labeled dataset for training despeckling filters for SAR imagery. Data Brief 2024, 53, 110065. [Google Scholar] [CrossRef] [PubMed]
  47. Perera, M.V.; Bandara, W.G.C.; Valanarasu, J.M.J.; Patel, V.M. SAR Despeckling Using Overcomplete Convolutional Networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 401–404. [Google Scholar]
  48. Ma, X.S.; Shen, H.F.; Zhao, X.L.; Zhang, L.P. SAR Image Despeckling by the Use of Variational Methods With Adaptive Nonlocal Functionals. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3421–3435. [Google Scholar] [CrossRef]
  49. Gomez, L.; Ospina, R.; Frery, A.C. Unassisted Quantitative Evaluation of Despeckling Filters. Remote Sens. 2017, 9, 389. [Google Scholar] [CrossRef]
  50. Guo, F.; Sun, C.; Sun, N.; Ma, X.; Liu, W. Integrated Quantitative Evaluation Method of SAR Filters. Remote Sens. 2023, 15, 1409. [Google Scholar] [CrossRef]
  51. Feng, H.; Hou, B.; Gong, M. SAR image despeckling based on local homogeneous-region segmentation by using pixel-relativity measurement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 2724–2737. [Google Scholar] [CrossRef]
  52. Deledalle, C.-A.; Denis, L.; Tupin, F.; Reigber, A.; Jäger, M. NL-SAR: A unified nonlocal framework for resolution-preserving (Pol)(In) SAR denoising. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2021–2038. [Google Scholar] [CrossRef]
  53. Phung, H.; Dao, Q.; Tran, A. Wavelet diffusion models are fast and scalable image generators. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 10199–10208. [Google Scholar]
  54. Nie, S.; Guo, H.A.; Lu, C.; Zhou, Y.; Zheng, C.; Li, C. The blessing of randomness: Sde beats ode in general diffusion-based image editing. arXiv 2023, arXiv:2311.01410. [Google Scholar]
  55. Lu, C.; Zhou, Y.; Bao, F.; Chen, J.; Li, C.; Zhu, J. Dpm-solver++: Fast solver for guided sampling of diffusion probabilistic models. J. Mach. Intell. Res. 2025, 22, 730–751. [Google Scholar] [CrossRef]
  56. Song, Y.; Dhariwal, P.; Chen, M.; Sutskever, I. Consistency models. arXiv 2023, arXiv:2303.01469. [Google Scholar]
Figure 1. Overview of the forward and reverse processes in diffusion model.
Figure 1. Overview of the forward and reverse processes in diffusion model.
Remotesensing 17 02970 g001
Figure 2. Visual comparison of the forward diffusion process under two different noise schedules: linear (top row) and cosine (bottom row). Each row displays the latent samples generated at uniformly spaced timesteps from the initial image at t = 0 to pure noise at t = T , illustrating how the noise is incrementally added according to each schedule.
Figure 2. Visual comparison of the forward diffusion process under two different noise schedules: linear (top row) and cosine (bottom row). Each row displays the latent samples generated at uniformly spaced timesteps from the initial image at t = 0 to pure noise at t = T , illustrating how the noise is incrementally added according to each schedule.
Remotesensing 17 02970 g002
Figure 3. The pipeline of the efficient conditional diffusion model for SAR despeckling.
Figure 3. The pipeline of the efficient conditional diffusion model for SAR despeckling.
Remotesensing 17 02970 g003
Figure 4. The architecture of the proposed method and its key modules.
Figure 4. The architecture of the proposed method and its key modules.
Remotesensing 17 02970 g004
Figure 5. The architectures of the core components: (a) residual block, (b) self-attention block, and (c) output block.
Figure 5. The architectures of the core components: (a) residual block, (b) self-attention block, and (c) output block.
Remotesensing 17 02970 g005
Figure 6. The architecture of the Wavelet-based Downsampling module.
Figure 6. The architecture of the Wavelet-based Downsampling module.
Remotesensing 17 02970 g006
Figure 7. Despeckling results of different methods with 1 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Figure 7. Despeckling results of different methods with 1 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Remotesensing 17 02970 g007
Figure 8. Despeckling results of different methods with 2 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Figure 8. Despeckling results of different methods with 2 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Remotesensing 17 02970 g008
Figure 9. Despeckling results of different methods with 4 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Figure 9. Despeckling results of different methods with 4 look speckle noise. The red dashed box highlights the clean reference and the corresponding noisy image, while the blue dashed box showcases the despeckling results from all compared methods. (a) clean reference. (b) Noisy image. (c) PPB. (d) SAR-BM3D. (e) SAR-DRN. (f) SAR-CAM. (g) SAR-Trans. (h) SAR-ON. (i) SAR-DDPM. (j) Proposed.
Remotesensing 17 02970 g009
Figure 10. Visual comparison of despeckling results on a real urban SAR image. Top (blue dashed box): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed; Bottom (red dashed box): corresponding ratio images (b2i2; filtered/original) for structural-preservation assessment.
Figure 10. Visual comparison of despeckling results on a real urban SAR image. Top (blue dashed box): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed; Bottom (red dashed box): corresponding ratio images (b2i2; filtered/original) for structural-preservation assessment.
Remotesensing 17 02970 g010
Figure 11. Visual comparison of despeckling results on a real mountainous SAR image. Upper (blue dashed region): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed (ours). Lower (red dashed region): corresponding ratio images (b2i2); no ratio is shown for the original.
Figure 11. Visual comparison of despeckling results on a real mountainous SAR image. Upper (blue dashed region): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed (ours). Lower (red dashed region): corresponding ratio images (b2i2); no ratio is shown for the original.
Remotesensing 17 02970 g011
Figure 12. Visual comparison of despeckling results on a real farmland (agricultural) SAR image. Upper (blue dashed region): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed (ours). Lower (red dashed region): corresponding ratio images (b2i2); no ratio is shown for the original.
Figure 12. Visual comparison of despeckling results on a real farmland (agricultural) SAR image. Upper (blue dashed region): (a) Original; (b1) PPB; (c1) SAR-BM3D; (d1) SAR-DRN; (e1) SAR-CAM; (f1) SAR-Trans; (g1) SAR-ON; (h1) SAR-DDPM; (i1) Proposed (ours). Lower (red dashed region): corresponding ratio images (b2i2); no ratio is shown for the original.
Remotesensing 17 02970 g012
Figure 13. Ablation study results on real SAR scenes. Top row (a1f1): urban scenes; bottom row (a2f2): rural scenes. Each sequence shows (a) original image and (bf) results of Experiments 1–5. Red boxes indicate ROIs with detailed magnifications.
Figure 13. Ablation study results on real SAR scenes. Top row (a1f1): urban scenes; bottom row (a2f2): rural scenes. Each sequence shows (a) original image and (bf) results of Experiments 1–5. Red boxes indicate ROIs with detailed magnifications.
Remotesensing 17 02970 g013
Figure 14. Visual comparison of synthetic image denoising effects under different samplers at various time steps: Top row (yellow dashed box): DDIM sampling results, where (a1i1) correspond to the original image and denoising results at T = 750, 500, 250, 100, 50, 25, 10, 5 respectively; Bottom row (red dashed box): results of the proposed method, where (a2i2) correspond to the same time step settings.
Figure 14. Visual comparison of synthetic image denoising effects under different samplers at various time steps: Top row (yellow dashed box): DDIM sampling results, where (a1i1) correspond to the original image and denoising results at T = 750, 500, 250, 100, 50, 25, 10, 5 respectively; Bottom row (red dashed box): results of the proposed method, where (a2i2) correspond to the same time step settings.
Remotesensing 17 02970 g014
Figure 15. Visual comparison of SAR image denoising effects under different samplers at various time steps: Top row (yellow dashed box): DDIM sampling results, where (a1i1) correspond to the original image and denoising results at T = 750, 500, 250, 100, 50, 25, 10, 5 respectively; Bottom row (red dashed box): results of the proposed method, where (a2i2) correspond to the same time step settings.
Figure 15. Visual comparison of SAR image denoising effects under different samplers at various time steps: Top row (yellow dashed box): DDIM sampling results, where (a1i1) correspond to the original image and denoising results at T = 750, 500, 250, 100, 50, 25, 10, 5 respectively; Bottom row (red dashed box): results of the proposed method, where (a2i2) correspond to the same time step settings.
Remotesensing 17 02970 g015
Table 1. Quantitative results on the MS-SAR dataset. Bold denotes the best result for each scene within the same look number.
Table 1. Quantitative results on the MS-SAR dataset. Bold denotes the best result for each scene within the same look number.
LookAlgorithmCityFarmMountain
SSIMPSNRSSIMPSNRSSIMPSNR
1-lookPPB0.68320.840.72021.480.67321.03
BM3D0.72521.790.73423.420.73422.97
SAR-DRN0.71321.430.75423.210.73621.93
SAR-Trans0.74122.460.77025.150.77023.78
SAR-ON0.74223.760.79525.460.78724.54
SAR-CAM0.75823.840.78825.410.77624.75
SAR-DDPM0.75023.850.80626.070.78824.31
Proposed0.76424.680.82426.860.79525.89
2-lookPPB0.72821.250.81922.340.80322.48
BM3D0.76323.410.83724.680.82823.89
SAR-DRN0.76622.780.82724.780.82423.61
SAR-Trans0.77523.850.84426.780.80825.12
SAR-ON0.78125.180.85826.950.84625.68
SAR-CAM0.79525.220.86426.840.85525.94
SAR-DDPM0.83725.340.86427.810.86025.34
Proposed0.84226.010.87028.560.86426.12
4-lookPPB0.80824.250.84924.840.82324.48
BM3D0.84325.110.86726.070.85825.23
SAR-DRN0.84624.780.84726.780.82425.61
SAR-Trans0.85025.240.86428.410.87726.46
SAR-ON0.85926.680.89128.890.88527.16
SAR-CAM0.85226.720.88129.050.89027.13
SAR-DDPM0.85726.780.90429.130.88627.34
Proposed0.86327.430.91030.560.89428.06
Table 2. Comparative evaluation of SAR image despeckling algorithms on real-world scenarios. The best and second-best results under different scenarios are highlighted in bold and italics, respectively.
Table 2. Comparative evaluation of SAR image despeckling algorithms on real-world scenarios. The best and second-best results under different scenarios are highlighted in bold and italics, respectively.
SceneAlgorithmENLMOIMOREPD-ROA
Vert
EPD-ROA
Hori
M-Index
CityPPB128.151.08760.90570.90170.87946.865
SAR-BM3D148.350.95431.04560.94320.90184.325
SAR-DRN142.240.92431.07650.94450.89645.765
SAR-CAM206.760.97141.02350.98460.98323.636
SAR-Trans126.650.95041.03450.94380.92172.993
SAR-ON132.280.98131.01430.98490.98463.778
SAR-DDPM127.660.97461.01730.98820.98872.874
Proposed149.370.98321.01040.99230.99182.335
MountainPPB134.190.89731.12930.88170.84578.327
SAR-BM3D76.140.92971.07340.90150.85827.689
SAR-DRN83.540.87521.12930.90150.86315.793
SAR-CAM107.370.94531.05480.90150.87344.038
SAR-Trans154.370.94781.04730.91660.89514.732
SAR-ON137.800.96150.98140.91770.87933.487
SAR-DDPM116.721.04640.99660.92030.88433.019
Proposed128.470.97891.01040.92780.90142.866
FarmPPB317.410.93450.92560.92110.89457.922
SAR-BM3D325.290.95461.05970.95450.92474.472
SAR-DRN164.860.96741.03490.94580.92285.233
SAR-CAM272.990.97281.02880.97660.98743.452
SAR-Trans243.640.97651.02430.95940.94896.768
SAR-ON340.020.98281.01480.98850.98423.019
SAR-DDPM273.670.84071.27930.98840.98874.273
Proposed318.911.01260.98850.99270.99082.986
Table 3. Quantitative evaluation of ablation studies. The best results are highlighted in bold.
Table 3. Quantitative evaluation of ablation studies. The best results are highlighted in bold.
ExperimentWavelet BasisPrediction ModeL = 1 (PSNR/SSIM)L = 2 (PSNR/SSIM)L = 4 (PSNR/SSIM)Single-Step Inference Time (s)
Experiment 1None ϵ θ 25.43/0.76227.58/0.79229.40/0.8340.03473
Experiment 2None ϵ θ + v θ 25.84/0.78827.99/0.81629.71/0.8580.03478
Experiment 3Haar ϵ θ + v θ 26.32/0.81728.47/0.84330.29/0.8830.03468
Experiment 4db4 ϵ θ + v θ 26.23/0.80428.38/0.83430.20/0.8740.03474
Experiment 5bior2.2 ϵ θ + v θ 25.47/0.78227.62/0.81029.44/0.8520.03486
Table 4. Comparison of inference performance across multiple sampling steps for different samplers. The best results are highlighted in bold.
Table 4. Comparison of inference performance across multiple sampling steps for different samplers. The best results are highlighted in bold.
SamplerMetrict = 1000 t = 750t = 500t = 250t = 100t = 50t = 25t = 10t = 5
DDPMSSIM0.748--------
PSNR (dB)25.08--------
Time (s)34.69--------
DDIMSSIM-0.6310.5840.5660.7430.7620.7710.5420.691
PSNR (dB)-21.2119.2120.4722.1723.2923.4518.8222.6
Time (s)-26.6917.588.873.521.780.880.360.19
ProposedSSIM-0.7430.7510.7630.7950.7840.7860.7650.694
PSNR (dB)-25.0225.1225.3425.4625.5325.2925.0219.28
Time (s)-26.1417.498.643.491.730.870.350.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Z.; Hu, W.; Zheng, S.; Zhang, B.; Zhou, M.; Peng, J.; Yao, Z.; Feng, M. Efficient Conditional Diffusion Model for SAR Despeckling. Remote Sens. 2025, 17, 2970. https://doi.org/10.3390/rs17172970

AMA Style

Guo Z, Hu W, Zheng S, Zhang B, Zhou M, Peng J, Yao Z, Feng M. Efficient Conditional Diffusion Model for SAR Despeckling. Remote Sensing. 2025; 17(17):2970. https://doi.org/10.3390/rs17172970

Chicago/Turabian Style

Guo, Zhenyu, Weidong Hu, Shichao Zheng, Binchao Zhang, Ming Zhou, Jincheng Peng, Zhiyu Yao, and Minghao Feng. 2025. "Efficient Conditional Diffusion Model for SAR Despeckling" Remote Sensing 17, no. 17: 2970. https://doi.org/10.3390/rs17172970

APA Style

Guo, Z., Hu, W., Zheng, S., Zhang, B., Zhou, M., Peng, J., Yao, Z., & Feng, M. (2025). Efficient Conditional Diffusion Model for SAR Despeckling. Remote Sensing, 17(17), 2970. https://doi.org/10.3390/rs17172970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop