Next Article in Journal
Obtaining Rotational Stiffness of Wind Turbine Foundation from Acceleration and Wind Speed SCADA Data
Previous Article in Journal
Evaluating Image Quality Metrics as Loss Functions for Image Dehazing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Path Adversarial Denoising Network Based on UNet

1
School of Mathematics and Computer Science, Jilin Normal University, Siping 136000, China
2
School of Computer Science and Technology, Jilin University, Changchun 130012, China
3
Jilin Provincial Key Laboratory for Numerical Simulation, Jilin Normal University, Siping 136000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4751; https://doi.org/10.3390/s25154751 (registering DOI)
Submission received: 30 June 2025 / Revised: 25 July 2025 / Accepted: 29 July 2025 / Published: 1 August 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Digital image quality is crucial for reliable analysis in applications such as medical imaging, satellite remote sensing, and video surveillance. However, traditional denoising methods struggle to balance noise removal with detail preservation and lack adaptability to various types of noise. We propose a novel three-module architecture for image denoising, comprising a generator, a dual-path-UNet-based denoiser, and a discriminator. The generator creates synthetic noise patterns to augment training data, while the dual-path-UNet denoiser uses multiple receptive field modules to preserve fine details and dense feature fusion to maintain global structural integrity. The discriminator provides adversarial feedback to enhance denoising performance. This dual-path adversarial training mechanism addresses the limitations of traditional methods by simultaneously capturing both local details and global structures. Experiments on the SIDD, DND, and PolyU datasets demonstrate superior performance. We compare our architecture with the latest state-of-the-art GAN variants through comprehensive qualitative and quantitative evaluations. These results confirm the effectiveness of noise removal with minimal loss of critical image details. The proposed architecture enhances image denoising capabilities in complex noise scenarios, providing a robust solution for applications that require high image fidelity. By enhancing adaptability to various types of noise while maintaining structural integrity, this method provides a versatile tool for image processing tasks that require preserving detail.

1. Introduction

Images are crucial for the dissemination of information and are widely used in various fields, including scientific research, industry, healthcare, security, and daily life. Separating noise from images while preserving the clean image is a vital preprocessing step in many visual tasks, including image classification, image segmentation, medical imaging, image compression, video denoising, and object detection in remote sensing imaging. Image quality has a direct impact on the accuracy and efficiency of subsequent analysis, recognition, and decision-making tasks. Therefore, it is of great significance to enhance the image quality by restoring the original details and clarity through a model.
Traditional denoising methods, while effective in specific scenarios, have obvious limitations in handling complex noise patterns, preserving details, and adapting to various types of noise [1,2,3,4,5]. For example, wavelet transform-based methods may cause artifacts when processing high-frequency components. Non-local means (NLMs) methods [6], though leveraging image self-similarity, suffer from low computational efficiency and performance degradation in high-noise scenarios. Even recent improvements have not solved these problems [7]. Block-Matching and 3D Filtering (BM3D) [8] enhances denoising performance, but it faces scalability issues with large datasets. Focusing primarily on local information, these methods struggle to capture global image structures and long-range dependencies, leading to unsatisfactory results in complex situations. For instance, some techniques struggle to balance noise removal and detail preservation in scenarios involving Gaussian noise. When removing salt-and-pepper noise, it often causes over-smoothing and texture loss. These shortcomings underscore the need for a more robust and comprehensive denoising framework [9].
Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs) have demonstrated significant potential in image processing tasks. CNNs excel at learning hierarchical feature representations, effectively removing noise and restoring details. GANs enrich denoising models’ training data by generating realistic noisy images through adversarial training. These technologies show significant promise in handling complex noise patterns and preserving key texture information. Researchers have proposed at least dozens of CNN-based denoising methods for general or specific images [5]. For example, DnCNN uses deep convolutional networks for efficient denoising [10]. Noise2Noise expands the applicability of denoising models by utilizing unsupervised training with noisy image pairs [11], thereby eliminating the need for clean reference images. These CNN-based denoising methods consistently outperform traditional denoising techniques. However, CNN-based denoising networks depend heavily on extensive and challenging-to-create training datasets. Additionally, the performance of these models can be significantly affected by discrepancies between the simulated noise and the actual noise. GAN-based methods such as ID-MSE-WGAN [12] and Trident GAN [13] can also lead to specific improvements. The training of GANs is also typically complex as the generator and discriminator struggle to reach an equilibrium, leading to convergence difficulties and generation quality that is difficult to control.
In light of the limitations of the aforementioned denoising methods, we propose a novel image-denoising framework designed to address these issues. The framework combines a UNet-like encoder–decoder architecture with adversarial training. The framework adopts a dual-path adversarial mechanism. The generator synthesizes diverse noisy images to enhance the training data. Meanwhile, the denoiser, with a dual-UNet structure, extracts multi-scale features and fuses dense features to strengthen noise removal and detail restoration. The discriminator assesses the authenticity of the generated images and optimizes the denoising process. This adversarial setup not only enhances the model’s adaptability to various noise types but also boosts its robustness in complex scenarios, offering a comprehensive solution to the challenges faced by traditional methods.
The contributions of this paper can be summarized as follows:
  • We design a dual-path adversarial structure that takes a set of clean and noisy images as input during training. The model can continuously optimize itself through the adversarial structure, thereby enhancing the denoising performance.
  • We modify the UNet architecture and propose a new dual-U-shaped network model. This model can extract image features more finely and effectively reduce image noise.
  • We also verify other multi-U-shaped network designs through experiments, demonstrating the practical reliability of the proposed dual-U-shaped network.

2. Related Work

In the field of image denoising, researchers are constantly exploring. The U-shaped encoder–decoder structure is a typical application of fully convolutional networks in image denoising. It uses downsampling layers to reduce image spatial resolution and mine semantic content and structural information. Through skip connections, it preserves image details and edge information, which is of great significance for maintaining feature continuity and improving denoising accuracy. The UNet network has shown outstanding performance in biomedical image segmentation and other fields [14]. The UNet++ architecture further improves performance by nesting UNet structures and enhancing skip connections [15]. MS-UNet-v2 performs well in medical image segmentation tasks using small training datasets through adaptive denoising methods and other techniques [16]. UNet-like architectures are susceptible to input resolution. When the spatial size of test images deviates significantly from that of the training crops, skip connections may fail due to feature map misalignment, resulting in edge artifacts. Our Dual-UNet mitigates this issue by employing multi-scale random cropping and online re-sampling during training and by inserting a deformable alignment module at every skip connection. This module automatically registers feature maps in both spatial and channel dimensions, ensuring that high-resolution test images fully exploit cross-layer details. But excessive downsampling can lead to blurred image details. Therefore, it is crucial to reduce the number of downsampling layers while maintaining tight connections between layers. Our proposed dual-UNet structure significantly improves denoising performance.
Skip connections were initially used to address the vanishing and exploding gradient problems in neural network training. As network depth increases, training efficiency declines. Adding skip connections enables the propagation of low-level learned weights to higher layers, allowing for efficient feature capture, accelerated training, and improved performance [17]. The DenseUNet model achieves remarkable results in image denoising tasks through dense skip connections and feature fusion [18]. Although dense connectivity alleviates gradient vanishing, it causes the number of feature channels to grow exponentially with depth, leading to a quadratic increase in GPU memory consumption. The dual UNet addresses this by replacing traditional dense connections with a combination of serial and parallel approaches. It retains only lightweight residual blocks within each sub-network, which helps keep the overall memory usage comparable to that of a single UNet. Additionally, skip connections can expand the receptive field of the network. Our designed MRDB blocks reduce the network’s reliance on isolated learned features, better utilizing skip connections. Dense skip connections continuously introduce features from different levels. After undergoing feature fusion, the model can capture multi-level image features, laying the foundation for generating clear images.
Dilated convolution controls the receptive field of the convolution kernel by adjusting the dilation rate [19], enabling the model to capture more contextual information. In image denoising, it can enlarge the receptive field to capture low-frequency structural information better and suppress high-frequency noise. Meanwhile, it maintains the feature map resolution, which is crucial for preserving image details. In image denoising applications, dilated convolution is often combined with other deep learning techniques, such as CNN and ResNets, to build efficient denoising models [20,21]. Chen et al. studied the application of dilated convolution in semantic image segmentation [22]. The multi-scale denoising network based on dilated convolution proposed by Wang et al. improved denoising results [23]. The multi-scale dilated convolutional neural network by Li et al. also effectively enhanced denoising performance [24]. Cascaded dilated convolutions are prone to “gridding artifacts,” manifesting as checkerboard patterns in smooth regions. Moreover, excessively large dilation rates cause the effective sampling grid to become overly sparse, resulting in the loss of fine-grained texture details. The MRDB block we designed uses dilated convolutions with different parameter-adjusted sizes. It implicitly uses downsampling to generate multi-scale features, adaptively suppressing artifacts generated by overly large dilation rates, which is beneficial for learning image contexts. Experiments have shown that denoising models utilizing dilated convolution can achieve effective noise reduction while preserving image details and textures.
Generative Adversarial Networks (GANs) were proposed by Ian Goodfellow and his team in 2014 [25]. These networks generate realistic data samples through adversarial training between a generator and a discriminator. In image denoising, GANs are widely used to learn the mapping between noisy and clean images, thereby removing noise. The generator creates noisy images, and the discriminator determines whether an input image is a real-world noisy image or a generated one. Through adversarial training, the generator learns to produce more realistic noisy images, indirectly learning to recover clean image information from noisy images. Isola et al. explored conditional adversarial networks for image-to-image translation tasks [26]. Lai et al. combined GANs with Laplacian pyramid networks for image super-resolution tasks [27]. Arjovsky et al. proposed a GAN training method based on the Wasserstein distance to alleviate the mode-collapse problem [28]. Thus, GANs show great potential in image denoising. In addition, Tai et al. designed MemNet to remove image blur effectively [29]. Wang et al. proposed the LG-BPN network, which utilized a low-resolution guided bidirectional pyramid structure to enhance image denoising performance [30]. Wu et al. constructed a dual residual attention network, combining residual learning and attention mechanisms to boost image denoising performance [31]. These approaches can effectively remove image noise while preserving details and texture information. However, training GANs is extremely sensitive to hyperparameters; even minor adjustments in learning rates or loss weights can trigger mode collapse or training oscillations. In addition, adversarial losses often sacrifice pixel-level fidelity in favor of perceptual realism, which can introduce structurally unacceptable errors in high-fidelity domains, such as medical imaging. To address these issues, we adopt a composite loss: the adversarial loss is used solely for distribution alignment, the pixel-wise loss enforces strict per-pixel fidelity, and a feature-level loss locks structural consistency. Gradient penalty and early stopping are further introduced to stabilize convergence and simultaneously preserve both perceptual quality and numerical accuracy. Meanwhile, we utilize a specially designed denoising model to ensure compatibility with the generator, stabilize image quality, and employ generated noisy images to assist in denoiser training, thereby enhancing the ability to process noisy images.

3. Network Structure

3.1. Overall Structure

We propose a denoising model that combines a dual-path UNet structure with a GAN. As shown in Figure 1, the overall network processes images by inputting noisy and clean images into the discriminator, a clean image into the generator and a noisy image into the denoiser, for dual-path adversarial training. The generator learns to produce realistic-looking noisy images paired with real, clean ones as fake samples. The denoiser is trained to remove noise from real noisy images, with the processed images paired with the original noisy ones as another set of fake samples, yielding real clean and corresponding noisy images from real samples. These three sets of images are input into the discriminator for joint distribution discrimination. Through continuous adversarial learning, the generator gets better at generating realistic-looking noisy images, thereby indirectly learning to recover clean images from noisy ones. This mechanism enables GANs to excel in image denoising, effectively removing noise while preserving details and textures. Algorithm 1 illustrates the running process of the model we designed.
Algorithm 1 Dual-path adversarial denoising network training procedure
Input:
Clean image set X = { x 1 , , x n } .
Noisy image set Y = { y 1 , , y n } (unpaired).
Hyper-parameters: α , τ D , τ G , λ gp , ksize, batch_size, epochs, early_stop_patience
 Parameters
:
θ D — Denoiser (Dual-UNet) parameters.
θ G — Generator (UNet-G) parameters.
θ P — Discriminator (PatchGAN) parameters.
Output:
Trained denoiser D ( x ; θ D ) with minimal noise and maximum detail preservation
1:
Randomly initialize θ D , θ G , θ P
2:
Pre-compute Gaussian kernel K = get_gausskernel ( ksize , C = 3 )
3:
for epoch = 1  to epochs do
4:
    for mini-batch = 1 to | X | / batch_size  do
5:
        /* 1. Sample data */
6:
         { x } Random batch from X
7:
         { y } Random batch from Y (same size)
8:
         { z } N ( 0 , 1 ) noise map same spatial size as x
9:
        /* 2. Forward passes */
10:
         x ^ y D ( y ; θ D )                                                                      ▹ Denoised image
11:
         y ˜ G ( concat ( x , z ) ; θ G )                                                           ▹ Synthetic noisy image
12:
        /* 3. Update Discriminator P */
13:
         L P α P ( x ^ , y ; θ P ) P ( x , y ; θ P ) + ( 1 α ) P ( x , y ˜ ; θ P ) P ( x , y ; θ P ) + λ gp · gradient_penalty
14:
         θ P θ P η P θ P L P
15:
        if mini-batch mod num_critic = 0  then
16:
           /* 4. Update Denoiser D */
17:
            L D ( 1 α ) P ( x ^ , y ; θ P ) + τ D x ^ x 1
18:
            θ D θ D η D θ D L D
19:
           /* 5. Update Generator G */
20:
            L G α P ( x , y ˜ ; θ P ) + τ G GF ( y ˜ x ) GF ( y x ) 1
21:
            θ G θ G η G θ G L G
22:
        end if
23:
    end for
24:
    /* 6. Early stopping check */
25:
    val_loss ← MAE on validation set
26:
    if val_loss has not improved for early_stop_patience epochs then
27:
        break
28:
    end if
29:
end for
30:
return  D ( x ; θ D )
To overcome the inherent trade-off between global structure preservation and local detail recovery that plagues traditional single-path denoising networks, we propose a dual-path adversarial architecture. In the first path, a generator learns to “forge” noise by mapping clean images to realistic noise distributions that can deceive the discriminator into regarding them as authentic noisy samples. In the second path, a denoiser learns to “restore” the truth by reverting authentic noisy images to their clean counterparts, rendering them indistinguishable from genuine clean images to the discriminator. Both paths converge at the shared discriminator, establishing a dynamic adversarial interplay: the generator progressively refines the accuracy of noise simulation, thereby supplying the denoiser with richer and more challenging training samples, while the denoiser, guided by adversarial feedback, continually enhances its capacity for preserving edges and textures. By sharing a single discriminator, the system achieves closed-loop optimization of “noise synthesis–noise removal,” thereby expanding the model’s adaptability to diverse noise distributions and mitigating the overfitting risks inherent in fixed-prior single-path designs. We adopt the UNet structure because it already has excellent performance. We utilize a dual-UNet to optimize the efficiency of the UNet architecture.

3.2. Denoiser Double-Layer U-Denoise Network

The denoiser we designed, the Double-Layer UNet Denoising Network (DUNetD), uses a dual-UNet structure for image processing tasks. The architecture is shown in Figure 2. This dual UNet is an integrated learning-based serial application, enabling the two networks to extract and process image features at different levels. An input image I R H × W × C (where H, W, and C represent image height, width, and channels, respectively, with C being 3 for RGB images) is fed into both UNet 1 and UNet 2 simultaneously.
The encoder of the UNet 1 network first processes the input image I, extracting features and reducing the spatial dimension through consecutive downsampling and Multi-Receptive Field Perception Denoise Module (MRDB) operations. This process yields feature maps F 1 , F 2 , F 3 , and F 4 , each with dimensions R H i × W i × C , where H i and W i gradually decrease due to downsampling. Finally, the deep characteristics F 5 are derived from F 4 via the MRDB.
Meanwhile, the input image I is also fed into the encoder of UNet 2 , which undergoes similar downsampling and processing by the module MRDB 2 to generate feature maps f 1 , f 2 , and f 3 , with f 4 subsequently derived from f 3 . The decoder of UNet 2 then restores the spatial dimensions of the image through upsampling and module processing, blending with the corresponding skip-connection feature maps to produce f 3 , f 2 , and f 1 .
Four consecutive convolutional layers are employed in the UNet 2 output stage. Each layer has a kernel of 3 × 3 , a stride of 2, and a padding of 1, progressively halving the image size. These layers reduce the size of f 1 step by step, yielding O 2 .
The output O 2 of UNet 2 is added to the deep feature F 5 of UNet 1 , forming the fused feature F end . This F end guides the UNet 1 decoder, which restores the image spatial dimensions via upsampling and MRDB′ module processing. The decoder merges with the corresponding skip-connection feature maps, generating f 4 , f 3 , f 2 , and f 1 . Finally, f 1 passes through the last processing layer (Conv3 × 3) to produce the final output image O 1 , also sized R H × W × 3 .
The dual-UNet structure enables the two networks to extract and process image features at different levels. Feature fusion and skip connections combine their advantages for more efficient image processing. This design enhances the ability of the network to capture image details and improves processing accuracy and efficiency through collaboration between the two networks.

3.2.1. Multi-Receptive Field Perception Denoise Module

We propose a Multi-Receptive Field Perception Denoise Module (MRDB) embedded in the dual-UNet structure of the image denoiser. The architecture is shown in Figure 3. The upper UNet and lower UNet utilize different ASPP modules to enhance multi-scale feature extraction, thereby improving the denoiser’s ability to capture features at various scales, as illustrated in the figure.
The core of the MRDB block is the ASPP submodule, which uses parallel convolutional kernels with different dilation rates to capture image info at multiple scales. During forward propagation, the input data are pre-activated and batch-normalized via a standard convolutional layer and then enter the ASPP submodule. In this module, the data pass through two convolutional layers with dilation rates of 3 and 6 and 4 and 8, respectively, depending on the ASPP version. Another path involves global average pooling, a 1 × 1 convolutional layer, and upsampling to match the output size of the feature extraction path. The outputs of these four paths are connected to form multi-scale features, which are then fused through convolutional layers and subjected to nonlinear transformations through activation functions. After another convolutional layer for further processing, the features undergo batch normalization and activation to produce the final output. Applying different ASPP modules in the dual UNet enhances the network’s feature extraction capability, improving noise removal and image quality restoration in denoising tasks.

3.2.2. Enlarging the Receptive Field Module

In image denoising, obtaining multi-scale information is crucial for feature reconstruction. Although downsampling is common in networks, it can damage image structures and lose information in low-resolution cases, which is bad for feature reconstruction. To enlarge the receptive field, large-size convolutional kernels or large strides in pooling are often used; however, they both have high computational requirements and resolution loss issues.
Inspired by semantic segmentation models, we introduce an enlarged receptive field module into the dual-UNet structure. It can expand the receptive field and capture multi-scale information without increasing computation and sacrificing resolution. The ASPP module (Atrous Spatial Pyramid Pooling) [32], as a spatial pyramid pooling structure, is widely used in semantic segmentation [33] and other tasks. By applying convolutional kernels with different dilation rates in parallel, this module captures multi-scale contextual information, enabling the model to handle objects of varying sizes and improving its adaptability to complex scenes. Typically integrated into the decoder of deep Convolutional Neural Networks (CNNs), ASPP fuses multi-scale contextual information into feature maps, significantly enhancing segmentation accuracy. Importantly, it can achieve a large receptive field without reducing the size of the feature map [34].
Specifically, the ASPP structure employs a pooling layer and a 1 × 1 convolutional layer in the upper UNet, as well as two dilated convolutions with dilation rates of 3 and 6, respectively. The corresponding paddings are 3 and 6, with a stride of 1 and a kernel size 3 × 3 for both, ensuring consistent feature map dimensions for fusion. Subsequently, the LeakyReLU activation function applies a negative slope of 0.2. To obtain richer information on the feature, we first use a 1 × 1 convolution to expand the feature channel. We then concatenate the features obtained from different receptive fields in the fusion section and use a 3 × 3 convolution to compress the feature channels, forming a comprehensive feature representation.
Thus, the denoiser uses a dual-UNet structure. The module enlargement of the receptive field is in the upper UNet, while the lower UNet uses dilated convolutions with dilation rates of 4 and 8, corresponding paddings of 4 and 8, stride of 1, and 3 × 3 kernels, the same as in the upper UNet. The two UNets extract image features to different extents with different ASPP modules. This enables the denoiser to capture more details and multi-scale features, thereby better removing noise and restoring image quality in denoising tasks.

3.3. Generator U-Shaped Imitation Noise Generation Network

This architecture achieves effective simulation of complex data distribution through adversarial learning dynamics between the generator (U-shaped Imitation Noise Generation Network, UING) and the discriminator (Dual-channel Adversarial Discriminator, D2P). In recent years, researchers have proposed various image denoising methods based on GANs [35,36]. These methods have significantly improved noise reduction while preserving the image details. However, these methods may encounter pattern collapse and instability issues during the training process, leading to fluctuations in the quality of the generated images. To address these issues, this paper employs a simplified generator model that is matched to the denoiser to stabilize image quality. Moreover, the generated noisy images can further assist in training the denoiser, enhancing its ability to process noisy images.
In the generator, we adopt the UNet 1 structure of the dual-UNet network used in the denoiser. The architecture is shown in Figure 4. It employs a single-input approach using UNet 1 as the generator. The network takes a clean image I R H × W × 3 as input (where H, W, and C represent the height, width, and channels of the image, respectively, with C being 3 for RGB images). By incorporating noise through network training, it is possible to generate noise simulation images similar to noisy images.
Due to its strong feature extraction capabilities and simple structure, the generator utilizes the upper UNet structure ( UNet 1 ) of the dual-UNet network. This design choice enables the reuse of existing network architectures and parameters, thereby saving development time and computational resources. It also allows the application of established training and optimization techniques for efficient noise image generation. During training, the stability of the UNet architecture ensures quick convergence. Adjusting parameters and training strategies can control the intensity and type of noise generated. Moreover, the flexibility of UNet 1 allows for customization to meet diverse application needs.
m i n G m a x D V ( D , G ) = E x p d a t a ( x ) [ l o g D ( x ) ] + E z p z ( z ) [ l o g ( 1 D ( G ( z ) ) ) ]
First, we define a prior input noise variable p z ( z ) . A differentiable function G ( z ; θ g ) maps this noise to the data space, where θ g represents its parameters. The goal is to learn the data-generating distribution p g . The discriminator D ( x ; θ d ) , also a multi-layer perceptron, outputs a scalar indicating the probability that the input data x come from real data rather than p g .
During training, the discriminator D aims to maximize the probability of correctly distinguishing between real training samples and those generated by G. This means outputting a probability close to 1 for real data and close to 0 for data generated. At the same time, the generator G seeks to minimize l o g ( 1 D ( G ( z ) ) ) in order to fool the discriminator so that it cannot easily tell generated data from real data. This adversarial relationship forms a two-player minimax game, reflected through the value function V ( G , D ) . As training progresses, D continuously improves its recognition ability, and G keeps improving the quality of the generated data. Eventually, this process drives the generator’s data distribution p g to gradually approach the real data distribution, achieving a dynamic equilibrium. The generator’s primary purpose is to create noise simulation images for data augmentation, especially when clean image data are limited. This significantly improves the generalization and robustness of the denoising model.

3.4. Discriminator Dual-Channel Adversarial Discriminator

The discriminator network [37] in this paper is a deep CNN-based structure. Its core consists of multiple consecutive convolutional layers, each of which is followed by a LeakyReLU activation function. The architecture is shown in Figure 5. The network, taking a clean RGB input image I R H × W × C (where H, W, and C, represent the height, width, and channels of the image, respectively, with C being 3 for RGB images), processes the input through a series of downsampling operations, and outputs a scalar value to judge the authenticity of the input image.
The discriminator network begins with a convolutional layer that maps the input image’s channels from I n p u t to n d f using a 4 × 4 kernel, stride of 2, and padding of 1, thereby halving the spatial dimensions. Next, several convolutional layers progressively increase the channel count to n d f × 2 , n d f × 4 , n d f × 8 , n d f × 16 , all with the same kernel size, stride, and padding, followed by LeakyReLU for non-linearity. After these layers, the spatial size of the image is gradually reduced. Finally, the last convolutional layer compresses it to 1 × 1 with n d f × 32 channels.
After convolution, the network flattens the feature maps into a vector and then maps it to a scalar output via a fully connected layer, indicating the discriminator’s judgment on the authenticity of the input image. The weights of all convolutional and fully connected layers are initialized with a normal distribution to ensure stability and optimal performance. This design enables the discriminator to effectively extract multi-scale features and enhance its ability to discern image authenticity through progressive downsampling.

3.5. Loss Function

The adversarial loss function trains the generator and discriminator by integrating the discrimination results of real data pairs, generated data pairs, and real clean images with simulated noisy images. Therefore, the generator can generate more realistic noisy images, and the discriminator can distinguish real from generated data more accurately. By adjusting the hyperparameter α , the impact of different data pairs during training is balanced, thereby optimizing the overall performance of the GAN model.
l o s s g a n = E ( x , y ) [ O ( x , y ) ] α E ( x , y ) [ O ( x , y ) ] ( 1 α ) O ( x , y ) [ D ( x , y ) ]
where the parameters are defined as follows: y is the original noisy image serving as the input to the denoising model; x represents the denoised image, which is the output of the model after processing the noisy image; y is the simulated noisy image generated by adding noise to a clean image, used for model training; x stands for the original clean image, the ideal target and reference for denoising; O ( x , y ) is a real data pair consisting of a clean image x and its corresponding noisy image y, offering authentic data distribution for training; O ( x , y ) is a data pair from the denoiser, including the denoised image x and the original noisy image y, reflecting the denoising effect; O ( x , y ) is a data pair obtained after the generator adds noise, based on the clean image x and the simulated noisy image y , used to train the generator in the GAN; E ( x , y ) denotes the expectation over real data pairs O(x,y), showing the overall distribution of real image data; E ( x , y ) is the expectation over generated data pairs O ( x , y ) , assessing the denoiser’s overall performance on different noisy images; E ( x , y ) is the expectation over data pairs O ( x , y ) , guiding the generator to produce more realistic noisy images. α is a weighting coefficient in the interval [0, 1] that balances the relative emphasis placed on the two adversarially generated data components in the loss: the denoised outputs and the synthetic noisy images. Specifically, as α approaches 1, the loss increasingly penalizes the discriminator’s response to denoised samples, requiring the denoiser to produce cleaner images that are indistinguishable from real clean data. In contrast, as α approaches 0, the loss shifts its focus toward penalizing the discriminator’s response to the generated noise, encouraging the generator to synthesize more realistic noisy images that faithfully approximate the accurate noise distribution. The L G M e a n loss during training epochs is shown in Figure 6.
L G M e a n = E ( x , y ) G F ( G ( x ) G F ( y x ) ) 1
The value of the feature-level loss with Gaussian filtering (GF) is absorbed into the term L G M e a n for the presentation of the output, where G ( x ) denotes the synthetic noise map produced by the generator. The function of the pixel-level difference loss ( L p i x e l ) is to make the denoised image output by denoiser D as close as possible to the actual clean image at the pixel level, where D ( y ) is the output of the denoiser. The L G p i x e l loss during training epochs is shown in Figure 7.
L p i x e l = E ( x , y ) ( y D ( y ) ) x ) ) 1
L o s s = l o s s g a n + γ 1 x x 1 + γ 2 G F ( y x ) G F ( y x ) 1
The total loss function combines adversarial loss, pixel-level difference loss, and feature-level loss with Gaussian filtering (GF). It aims to make the generator’s noisy images closer to authentic images at multiple levels (pixel and feature). Hyperparameters γ 1 and γ 2 balance the impact of different loss terms to achieve better generation results.
γ 1 controls the weight of the loss at the pixel level x x 1 , which enforces the fidelity per pixel between the denoised output x and the clean image of the ground truth x. A larger γ 1 emphasizes sharper pixel-wise reconstruction but may over-smooth fine textures. γ 2 controls the weight of the loss at the pixel level G F ( y x ) G F ( y x ) 1 , which measures the structural similarity of noise patterns after Gaussian filtering. A larger γ 2 encourages the generator to synthesize noise whose low-frequency statistics (edges, smooth regions) closely match those of real noise, potentially at the expense of exact pixel values. According to a “grid search”, it was determined that γ 1 should be 1000 and γ 2 10.
The balance between adversarial loss and pixel-level loss can be framed as a joint optimization of fidelity and perceptual authenticity. Adversarial loss ( L o s s g a n ), driven by the discriminator’s game-theoretic mechanism, encourages the denoised output to align with the distribution of authentic clean images, thus enhancing visual naturalness. However, this mechanism does not offer an explicit constraint on pixel-wise consistency. To compensate, the loss of L1 at the pixel-level intervenes with a dominant weight γ 1 of 1000, rigidly correcting any detail drift that the adversarial loss may induce ( γ 1 x x 1 + γ 2 ). This strategy effectively reformulates the optimization objective into a constrained minimax problem. By ensuring that the discriminator cannot distinguish between authentic and generated images, we further restrict the solution space to those outputs that exhibit strict pixel-wise overlap with the ground-truth image. This approach preserves both reconstruction accuracy and perceptual quality.

4. Experiment and Result Analysis

4.1. Measurement Standards

In our research, the overall quality of the denoised images was evaluated using the Peak Signal-to-Noise Ratio (PSNR) [38] and the Structural Similarity (SSIM) [39].
P S N R = 10 l g { M a x 2 [ G t ( i , j ) ] 1 H W i = 1 H j = 1 W [ G t ( i , j ) N ( i , j ) ] 2 }
S S I M = ( 2 u 1 u 2 + c 1 ) ( 2 σ 1 , 2 + c 2 ) ( u 1 2 + u 2 2 + c 1 ) ( σ 1 2 + σ 2 2 + c 2 )
Here, G t ( i , j ) is the pixel value of the original noise-free image at position (i,j), and N ( i , j ) is the pixel value of the denoised image at position (i,j). H and W represent the height and width of the image. The mean of G t ( i , j ) and N ( i , j ) are represented by u 1 and u 2 . The variances of G t ( i , j ) and N ( i , j ) are indicated by σ 1 and σ 2 . The covariance between G t ( i , j ) and N ( i , j ) is represented by σ 1 , 2 . Constants c 1 = 0.01 and c 2 = 0.02 are introduced to maintain system stability.
Let G e be the ground-truth edge map, D e be the detected edge map, and k be a constant controlling distance influence. Let d be the Euclidean distance between the ground truth and the detected edges. | G e | and | D e | denote the pixel counts of these maps. p indexes each pixel position in the image, iterating through all pixels during FOM computation.
The quality of edge detection in the images was assessed using a metric called the Figure of Merit (FOM). It measures the location and magnitude of detected edges to ensure precision in edge detection.
F O M = 1 M a x ( | G e | | D e | ) p = 1 H W 1 1 + k . d G e 2 ( p )
By calculating the absolute error between the denoised image and the real image in each pixel and taking the average, the quality of the denoising effect can be measured.
M A E = 1 H W i = 1 H j = 1 W | G t ( i , j ) N ( i , j ) |
G t ( i , j ) is the pixel value of the original noise-free image at position (i,j), and N ( i , j ) is the pixel value of the denoised image at position (i,j). The MAE loss during training epochs is shown in Figure 8.

4.2. Experimental Dataset

In this experiment, three datasets with real-noise images were used: SIDD [40], DND [41], and PolyU [42]. They are invaluable for research on denoising algorithms.
The Smartphone Image Denoising Dataset (SIDD) contains 30,000 noisy images in 10 scenes. These images were captured using five representative smartphone cameras under different lighting conditions. Moreover, the SIDD provides ground-truth images corresponding to the noisy ones, laying a solid data foundation for denoising algorithm research.
The Darmstadt Noise Dataset (DND) comprises 50 images captured by four consumer cameras with different sensor sizes. This ensures data diversity and authenticity. For each pair of images, the reference image (ground truth) was shot at the base ISO, while the noisy image was shot at a higher ISO with adjusted exposure. The reference images underwent careful post-processing, including minor camera shift correction, linear intensity scaling, and removal of low-frequency bias to ensure their accuracy as a benchmark. However, since the noise-free images in this dataset are not publicly available, all images were used for testing only.
The PolyU dataset captured 40 scenes and computed the average of these images to create a near-true “ground truth” (a noise-free or low-noise reference). This leveraged the randomness of noise, using multiple samples and averaging to approximate absolute pixel values and reduce noise. Since the PolyU dataset includes noisy images and noise-free or low-noise references in the real world, it was also used as a test set for comparative analysis.
The training images from the SIDD-Medium dataset were used as the training data in this paper’s experiment. Given the large size of the SIDD training images, 300 random 256 × 256 patches were cropped from each original image pair, yielding 96,000 patches. The model was then trained on these smaller patches. For testing, the DND dataset, the testing images from the SIDD dataset, and the PolyU dataset were used, and all tests were performed on RGB images.

4.3. Comparison Method

4.3.1. Comparison of Key Methods

ID-MSE-WGAN is suitable for environments with limited hardware resources [12], which restricts its application in resource-constrained settings. The complexity of its network architecture restricts its capability for feature extraction and fusion. Additionally, its lack of image detail processing results in a gap between the actual denoising effects and theoretical expectations.
In contrast, our approach introduces an early-stopping mechanism. This effectively controls training time, prevents overfitting, and enhances the model’s generalization ability. Our design also includes a specialized module for processing image details. This module maintains high computational efficiency while significantly improving feature extraction and image processing performance.
The Adversarial Gaussian Denoiser Network (AGDN) combines adversarial loss and L1 loss for pixel reconstruction [9], thus optimizing the denoising network. Adversarial loss improves image clarity and sharpness, while L1 loss ensures that the generated images are closely aligned with the target images. However, dynamically adjusting the weights of these two losses across different noise levels and image content for optimal denoising remains a challenge. Moreover, AGDN’s complexity may lead to slow inference speeds in practical applications, making it unsuitable for real-time, demanding scenarios.
Our method uses the MAE loss function. It enhances the quality of the generator’s image generation, supervises the denoiser’s denoising performance, and maintains the overall model’s efficiency and stability. The dual-input design of the dual-UNet network may extend training time. However, the similarity of the top-down architecture does not slow down the inference speed in practical applications. Instead, it further enhances the model’s adaptability to complex scenarios.
CS-PCN employs a three-stage serial–parallel hybrid architecture. The first two stages comprise the cascaded Context Mining Twin Sub-network (CM2S), which is composed of an MLFP dilated convolution, an AED encoder–decoder, and an MCAC multi-head attention controller. These stages are responsible for extracting multi-scale semantic information. The third stage is a parallel Spatial Synthesis Sub-network (3S) that refines local details via cascaded dual-attention blocks. The two feature streams are finally fused by adding elements in a way that is elemental. This staged refinement strategy achieves a compromise between noise suppression and detail preservation. Nevertheless, the fixed serial–parallel topology of CS-PCN limits its adaptability to complex real-world noise; the absence of an adversarial mechanism hinders learning of unknown noise distributions; the effective receptive field is still bounded by network depth; and robustness degrades under training–testing domain shifts.
In contrast, our denoiser generalizes to unseen noise distributions without additional data augmentation. The dual-UNet structure enlarges the receptive field through cascaded-parallel synergy, performing hierarchical feature fusion that jointly ensures statistical fidelity and texture recovery. This design directly enhances robustness under complex and unknown noise scenarios.
Trident GAN introduces the Trident Block for feature extraction [13], which includes the Local Detail Enhancement Block (LDEB), Lightweight Channel Attention Block (LCAB), and Spatial Interaction Attention Block (SIAB). These modules extract local detail features, channel features, and spatial features, respectively, with feature aggregation achieved through a Multi-Feature Fusion (MFF) module [13]. This design improves feature utilization efficiency but has room for improvement in multi-scale feature extraction. Our approach employs the Multi-Receptive Field Perception Denoise Module (MRDB) embedded in a dual-UNet structure. The MRDB uses dilated convolutions with different rates to capture multi-scale features in parallel. This increases feature extraction efficiency and leverages the skip connections in UNet for more effective feature fusion. The dual-UNet network’s similar structure and dual-input design enable better feature fusion for different detail-feature tasks, further enhancing the model’s adaptability to complex scenes.
Trident GAN is suitable for environments with limited hardware resources and offers short testing times, making it ideal for scenarios that demand real-time performance. However, its feature extraction and fusion capabilities are constrained by the complexity of its network structure. In contrast, our proposed design is designed for medium-sized datasets. By incorporating an early-stopping mechanism, our approach effectively controls training time, preventing overfitting and enhancing the model’s generalization ability. This design maintains high computational efficiency and significantly improves feature extraction and image processing performance. The results of comparison are shown in Table 1.

4.3.2. Original Design Proposal

In exploring the concatenation design for the denoiser’s dual-U structure, various methods were initially tried. However, experiments revealed that these initial approaches yielded poor denoising results and underperformed in metric evaluations.
When connecting UNet 1 and UNet 2 sequentially and abandoning the dual-channel input in favor of a single input at the UNet 1 entrance, with processing through both networks leading to an output at the UNet 2 exit, the resulting structure resembled the letter “W” and was termed WNet. However, experimental results indicated that WNet performed poorly in image denoising. This was mainly attributed to insufficient feature fusion and gradient vanishing. In the original dual-UNet structure, feature fusion effectively combined the strengths of both networks, thereby enhancing the denoising effect. In the sequential structure, this fusion was compromised, as UNet 2 could only process UNet 1 ’s final output without access to its intermediate features. Moreover, the increased network depth from sequential connection may intensify gradient decay during backpropagation, negatively impacting training effectiveness and convergence speed.
Connecting the bottoms of UNet 1 and UNet 2 while maintaining dual-channel inputs from both entrances and producing separate outputs before fusing them into a single result created a structure resembling the letter “X”, hence the network was named XNet. Experiments showed that XNet’s denoising performance was even worse than that of WNet. This could be due to gradient conflicts or information interference during training after connecting the two UNet networks at the bottom. Such issues likely arose because the connected bottom layers disrupted information flow, leading to insufficient and interfering feature fusion. Consequently, the networks failed to fully utilize their feature extraction capabilities, negatively impacting the overall denoising effect.
During the dual-U design, a triple-U structure (U3Net) was proposed by adding a UNet 3 connected to the bottom of UNet 2 . U3Net had three inputs and one output. Although performing better than WNet and XNet, U3Net’s training took longer, and its final effect still lagged behind the proposed dual-U structure. While the UNet-based network could effectively utilize multi-level features and accelerate convergence, this was optimal at a certain depth. U3Net, though extracting diverse features, suffered from high complexity, longer training time, and inefficient feature fusion due to significant feature differences at connection points, making it less efficient than the final dual-U structure. The results of different design structures are shown in Table 2.

4.3.3. Computational Efficiency

In terms of parameters and FLOPs, the generator utilizes a 32-channel UNet backbone with approximately 7.8 M parameters and 34.2 GFLOPs per forward pass. The denoiser consists of two parallel UNet branches, which doubles the parameter count to 15.6 M and the computation to 68.9 GFLOPs. The discriminator is a 64-channel PatchGAN, comprising only 4.2 M parameters and 9.1 GFLOPs—the lightest component. In aggregate, the entire framework contains 27.6 M parameters and consumes 112.2 GFLOPs, well within the capability envelope of contemporary edge-side hardware. Training time was measured on 96k 64 × 64 patches from the SIDD dataset using a single V100S GPU (NVIDIA, Santa Clara, CA, USA). One epoch required approximately 9 min for the generator, 18 min for the denoiser, and 5 min for the discriminator; the complete adversarial iteration took a total of 32 min. With early stopping, total training converged within 46 h—comparable to the 40 h of a single-path UNet baseline—demonstrating that the dual-path design yielded substantial performance gains without incurring excessive training overhead. We performed inference delay testing with “batch = 1” on 100 512 × 512 images. Average runtimes were 18 ms for the generator, 38 ms for the denoiser, and 12 ms for the discriminator, resulting in an end-to-end latency of 55 ms, which is below the 100 ms threshold typically required for real-time applications. This confirms an effective balance between performance and efficiency. The calculation results are shown in Table 3.

4.4. Comparative Experiments

In this experiment, we built neural network models using the PyTorch framework. The hardware setup included a Tesla V100S-PCIE-32 GB GPU with 32 GB of memory, an Intel(R) Xeon(R) Gold 5220R CPU, and 32 GB of system RAM. We set the initial number of channels to 32 for the denoiser and generator and 64 for the discriminator. The initial learning rates for the denoiser and generator were 1 × 10 4 , and for the discriminator, it was 2 × 10 4 . All model parameters were updated using the Adam optimizer. To prevent overfitting, we implemented an early-stopping mechanism. Training would stop automatically when the validation loss plateaued, thereby enhancing generalization and avoiding the waste of computational resources. The details of the environment configuration are shown in Table 4.
Our experiment compared the denoising performance of the proposed network using test images from the DND dataset, SIDD Medium dataset, and PolyU dataset. The model was trained for 70 epochs, which took approximately 46 h. Table 5 and Table 6 summarize the quantitative comparisons of different algorithms in terms of PSNR and SSIM metrics. The PSNR for the SIDD-Medium dataset was tested uniformly on the RGB channels, while that of the DND dataset was obtained through the official system. The PolyU dataset contains only noisy images, so this paper provides a visual comparison. Nine denoising algorithms (AINDNet [43], BM3D [8], CBDNet [44], CS-PCN [45], DnCNN [10], FFDNet [46], RIDNet [47], Trident GAN [13], and TWSC [48]) were used for comparison. The results of comparative experiment are shown in Table 5 and Table 6, Figure 9.

4.5. Ablation Studies

To verify the effectiveness of the design, we conducted a series of ablation studies on three key components of the denoiser: the dual U structure, the MRDB module, and the ASPP module (part of MRDB). The results are summarized in Table 7.
  • Experimental settings:
    • Dataset: SIDD, consistent with previous comparative experiments.
    • Parameters: All experimental parameters were strictly set to the same values for accuracy and reproducibility of the results.
  • Experimental operation:
    • We removed the lower U-shaped structure ( UNet 2 ), retaining only the upper U-shaped structure ( UNet 1 ).
    • We removed the MRDB module. When the MRDB module was removed, the ASPP module was also excluded due to its inherent association with it.
    • We removed the ASPP module.
  • Results:
    • Compared to having no modules, the PSNR was improved by 2.83 dB, 2.89 dB, and 2.39 dB.
    • Compared to having no modules, the SSIM was improved 4.14%, 5.4%, and 5.31%.
    • Impact of removing all modules: the PSNR decreased by 2.96 dB and the SSIM by 0.056%, indicating a significant performance drop.
The dual UNet removed the lower UNet ( UNet 2 ) and retained only the upper UNet ( UNet 1 ), resulting in a significant drop in PSNR and SSIM. This is because the dual UNet extracted and fused multi-scale features at different levels through the collaboration of the two networks. With UNet 2 removed, the receptive field shrunk, feature fusion was lost, and the ability to capture complex image details diminished. In addition, the collaborative effect of the dual UNet was weakened, leaving the decoder with insufficient fused features for guidance and thereby reducing the image restoration quality. Moreover, the skip connections in UNet 2 conveyed low-level features, and their removal caused the decoder to lose details and spatial information, thereby impacting the image’s structural similarity (SSIM). The dual UNet was highly adaptable to complex scenes, and the removal of UNet 2 weakened its noise suppression ability, resulting in a decrease in PSNR and SSIM. The dual UNet enhanced image processing results through collaborative work and multi-scale feature fusion. Removing UNet 2 eliminated these advantages, resulting in a decrease in the indicators.
The ASPP module captures multi-scale contextual information through dilated convolutions with varying rates, thereby enhancing the model’s ability to capture features of different scales. This helps to handle the details and noise in images more effectively. Removing ASPP weakened the network’s multi-scale feature extraction, reduced the receptive field, and diminished feature fusion, resulting in a slight drop in denoising performance. ASPP concatenates and fuses features from different receptive fields to form a comprehensive representation, enriching feature information and boosting the model’s adaptability and robustness. ASPP optimizes computational efficiency without significantly increasing computation compared to large-kernel convolutions or large-stride pooling. In experiments, using ASPP led to slight improvements in PSNR and SSIM, indicating its positive role in detail restoration and structure preservation and markedly enhancing the model’s performance in image denoising.
As shown in Table 7, the dual-U structure is crucial to improve the PSNR and SSIM of the network. The MRDB module significantly boosts the network’s SSIM performance. The ASPP module further optimizes the SSIM. These three modules operate in tandem within the network architecture, effectively enhancing denoising performance. Their combined use maximizes denoising results, confirming the rationality and efficiency of our design.

4.6. Superparameter Experiment and Result Analysis

When fine-tuning the denoiser’s hyperparameters, network depth and the number of training epochs are key factors that significantly impact denoising performance. Experimental data suggested that changes in network depth and the number of training epochs affected the PSNR and SSIM, two metrics commonly used for image quality assessment. The results of experiment are shown in Figure 10.
As the number of training epochs increased, the denoiser’s performance initially improved, then stabilized, and could even decline slightly. At 20 epochs, the PSNR was 38.5 dB, and the SSIM was 91.07%, indicating insufficient learning of image features and noise patterns. Performance improved and stabilized at 50, 70, and 100 epochs, with peak PSNR and SSIM at 39.29 dB and 91.5%, respectively, at 70 epochs. This shows that proper training duration optimizes parameters and enhances denoising. However, at 150 epochs, PSNR and SSIM slightly dropped, possibly due to overfitting and reduced generalization. Additionally, excessive training may trap the network in local optima. The results of experiment are shown in Figure 11.
Therefore, both network depth and the number of training epochs impact the denoiser’s performance. The appropriate depth and epochs enable the denoiser to balance the PSNR and SSIM well. However, excessive depth or training time may degrade performance. This phenomenon is attributed to the unique decoder and skip connections of the U-shaped network, which enable rapid convergence, particularly on medium-sized datasets. Therefore, in practical applications, network depth and training epochs should be chosen based on specific task requirements and data characteristics to optimize the denoiser’s overall performance.

5. Conclusions

We introduced an advanced image denoising framework comprising three parts, denoising, generator, and discriminator, which combined the dual-path adversarial mechanism with an enhanced dual-UNet encoder–decoder structure. Adversarial training enhanced the model’s adaptability to various types of noise, while the dual-UNet architecture effectively extracted multi-scale features and fused dense features, resulting in significant improvements in noise removal and detail restoration. The experiment verified the superior performance of the proposed model, effectively overcoming the limitations of traditional methods and providing a robust solution for image denoising.
The proposed dual-path adversarial denoising network demonstrated robust performance across diverse noise types; however, its effectiveness was contingent upon the characteristics of the noise and the parameter configurations. While the model excelled in handling common noise patterns (e.g., Gaussian, salt-and-pepper noise), its stability may degrade with structured or non-stationary noise, such as low-frequency Perlin noise. Perlin noise, which introduces spatial coherence and gradient variations, challenges the model’s reliance on local receptive fields, potentially leading to residual artifacts or over-smoothing in homogeneous regions. Parameter sensitivity further influences performance. Variations in noise intensity or distribution (e.g., altering Gaussian standard deviation or spatial correlation) can disrupt the adversarial equilibrium. For instance, aggressive noise suppression may erode fine textures, while conservative settings might retain noise. The dual UNet with fixed dilation rates in ASPP modules may also limit adaptability to noise scales outside the trained range. Future work could explore dynamic dilation mechanisms or noise-specific attention modules to enhance generalization. These limitations highlight the trade-off between generality and specialization, indicating the need for noise-aware training protocols or hybrid methods to cope with complex real-world scenarios, which is precisely the problem we need to address in the future.

Author Contributions

Conceptualization, J.Y., M.S. and D.W.; methodology, J.Y., M.S. and D.W.; software, J.Y.; validation, J.Y.; formal analysis, J.Y. and Y.Z.; investigation, J.Y.; resources, J.Y.; data curation, J.Y.; writing—original draft preparation, J.Y. and Y.Z.; writing—review and editing, J.Y., M.S. and D.W.; visualization, J.Y. and Y.Z.; supervision, J.Y. and D.W.; project administration, J.Y. and D.W.; funding acquisition, J.Y. and D.W. All authors have read and agreed to the published version of the manuscript.

Funding

Scientific Research Project of Education Department of Jilin Province: JJKH20240573KJ.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in SIDD at DOI:10.1109/CVPR.2018.00182, reference number [40]; DND at DOI:10.1109/CVPR.2017.294, reference number [41]; PolyU at DOI:10.48550/arXiv.1804.02603, reference number [42].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Elad, M.; Aharon, M. Image Denoising Via Sparse and Redundant Representations over Learned Dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef]
  2. Coupé, P.; Hellier, P.; Kervrann, C.; Barillot, C. Nonlocal Means-Based Speckle Filtering for Ultrasound Images. IEEE Trans. Image Process. 2009, 18, 3736–3745. [Google Scholar] [CrossRef]
  3. Sattar, F.; Floreby, L. Image Enhancement Based on a Nonlinear Multi-scale Method. IEEE Trans. Image Process. 1997, 6, 888–895. [Google Scholar] [CrossRef]
  4. Wu, W.; Chen, M.; Xiang, Y.; Zhang, Y.; Yang, Y. Recent Progress in Image Denoising: A Training Strategy Perspective. IET Image Process. 2023, 17, 1627. [Google Scholar] [CrossRef]
  5. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for Image Denoising Using Convolutional Neural Network: A Review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  6. Buades, A.; Coll, B.; Morel, J.-M. Non-Local Means Denoising. Image Process. Line 2011, 1, 208–212. [Google Scholar] [CrossRef]
  7. Kaur, R.; Karmakar, G.; Imran, M. Impact of Traditional and Embedded Image Denoising on CNN-Based Deep Learning. Appl. Sci. 2023, 13, 11560. [Google Scholar] [CrossRef]
  8. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  9. Khan, A.; Jin, W.; Haider, A.; Rahman, M.; Wang, D. Adversarial Gaussian Denoiser for Multiple-Level Image Denoising. Sensors 2021, 21, 2998. [Google Scholar] [CrossRef]
  10. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2016, 26, 3142–3155. [Google Scholar] [CrossRef]
  11. Lehtinen, J.; Munkberg, J.; Hasselgren, J.; Laine, S.; Karras, T.; Aittala, M.; Aila, T. Noise2Noise: Learning Image Restoration without Clean Data. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  12. Chen, S.; Shi, D.; Sadiq, M.; Cheng, X. Image Denoising with Generative Adversarial Networks and Its Application to Cell Image Enhancement. IEEE Access 2020, 8, 82819–82831. [Google Scholar] [CrossRef]
  13. Wang, L.F.; Ren, W.J.; Guo, X.D.; Zhang, R.G.; Hu, L.H. Multi-Feature Generative Adversarial Network for Low-Dose CT Image Denoising. J. Comput. Appl. 2025, 1–12. Available online: https://link.cnki.net/urlid/51.1307.TP.20250327.0948.002 (accessed on 27 March 2025).
  14. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  15. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; DLMIA ML-CDS 2018; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11045, pp. 3–11. [Google Scholar] [CrossRef]
  16. Chen, H.; Han, Y.; Xu, P.; Li, Y.; Li, K.; Yin, J. MS-UNet-v2: Adaptive Denoising Method and Training Strategy for Medical Image Segmentation with Small Training Data. arXiv 2023, arXiv:2309.03686. [Google Scholar]
  17. Huang, G.; Liu, Z.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
  18. Zhao, J.; Li, Q.; Li, H.; Zhang, L. Automated Segmentation of Cervical Nuclei in Pap Smear Images Using Deformable Multi-Path Ensemble Model. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 1514–1518. [Google Scholar] [CrossRef]
  19. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  20. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  22. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  23. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding Convolution for Semantic Segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar] [CrossRef]
  24. Jia, X.; Peng, Y.; Ge, B.; Li, J.Y.; Liu, S.; Wang, W. A Multi-scale Dilated Residual Convolution Network for Image Denoising. Neural Process. Lett. 2022, 55, 1231–1246. [Google Scholar] [CrossRef]
  25. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.C.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 2014, 2, 2672–2680. [Google Scholar]
  26. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar] [CrossRef]
  27. Lai, W.-S.; Huang, J.-B.; Ahuja, N.; Yang, M.-H. Fast and Accurate Image Super-Resolution with Deep Laplacian Pyramid Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 2599–2613. [Google Scholar] [CrossRef] [PubMed]
  28. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. [Google Scholar]
  29. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A Persistent Memory Network for Image Restoration. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar] [CrossRef]
  30. Wang, Z.; Fu, Y.; Liu, J.; Zhang, Y. LG-BPN: Local and Global Blind-Patch Network for Self-Supervised Real-World Denoising. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 18156–18165. [Google Scholar] [CrossRef]
  31. Wu, W.; Liu, S.; Zhou, Y.; Zhang, Y.; Xiang, Y. Dual Residual Attention Network for Image Denoising. Pattern Recognit. 2023, 149, 110291. [Google Scholar] [CrossRef]
  32. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
  33. Eigen, D.; Fergus, R. Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2650–2658. [Google Scholar] [CrossRef]
  34. Takahashi, N.; Mitsufuji, Y. Densely Connected Multidilated Convolutional Networks for Dense Prediction Tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 993–1002. [Google Scholar]
  35. Chen, J.; Chen, J.; Chao, H.; Yang, M. Image Blind Denoising with Generative Adversarial Network Based Noise Modeling. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3155–3164. [Google Scholar] [CrossRef]
  36. Wang, X.; Li, Y.H.; Wu, L.Y.; Zhao, S.B. A CT Image Denoising Method Based on Improved GAN. Digit. Technol. Appl. 2024, 42, 201–203. [Google Scholar]
  37. Toxigon. Dive Deep Into GANs: Generative Adversarial Networks Explained. Toxigon Blog. Available online: https://toxigon.com/deep-dive-into-gans-generative-adversarial-networks (accessed on 1 January 2025).
  38. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 20th International Conference on Pattern Recognition, ICPR 2010, Istanbul, Turkey, 23–26 August 2010. [Google Scholar]
  39. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  40. Abdelhamed, A.; Lin, S.; Brown, M.S. A High-Quality Denoising Dataset for Smartphone Cameras. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1692–1700. Available online: https://ieeexplore.ieee.org/document/8578280 (accessed on 1 January 2025).
  41. Plotz, T.; Roth, S. Benchmarking Denoising Algorithms with Real Photographs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 1586–1595. [Google Scholar] [CrossRef]
  42. Xu, J.; Li, H.; Liang, Z.; Zhang, D.; Zhang, L. Real-world Noisy Image Denoising: A New Benchmark. arXiv 2018, arXiv:1804.02603. [Google Scholar] [CrossRef]
  43. Kim, Y.; Soh, J.W.; Park, G.Y.; Cho, N.I. Transfer Learning From Synthetic to Real-Noise Denoising With Adaptive Instance Normalization. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, WA, USA, 14–19 June 2020; pp. 3479–3489. [Google Scholar] [CrossRef]
  44. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward Convolutional Blind Denoising of Real Photographs. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1712–1722. [Google Scholar] [CrossRef]
  45. Jiang, Y. Image Denoising Algorithm Research Based on CNN-Transformer Multi-Interaction. Master’s Thesis, Beijing Jiaotong University, Beijing, China, 2023. [Google Scholar] [CrossRef]
  46. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN based Image Denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed]
  47. Anwar, S.; Barnes, N. Real Image Denoising With Feature Attention. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3155–3164. [Google Scholar]
  48. Xu, J.; Zhang, L.; Zhang, D. A Trilateral Weighted Sparse Coding Scheme for Real-World Image Denoising. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 20–36. [Google Scholar]
Figure 1. Overall network structure.
Figure 1. Overall network structure.
Sensors 25 04751 g001
Figure 2. Network architecture of the denoiser.
Figure 2. Network architecture of the denoiser.
Sensors 25 04751 g002
Figure 3. Multi-Receptive Field Perception Denoise Module.
Figure 3. Multi-Receptive Field Perception Denoise Module.
Sensors 25 04751 g003
Figure 4. Network architecture of the generator.
Figure 4. Network architecture of the generator.
Sensors 25 04751 g004
Figure 5. Network architecture of the discriminator.
Figure 5. Network architecture of the discriminator.
Sensors 25 04751 g005
Figure 6. L G M e a n loss during training epochs.
Figure 6. L G M e a n loss during training epochs.
Sensors 25 04751 g006
Figure 7. L p i x e l loss during training epochs.
Figure 7. L p i x e l loss during training epochs.
Sensors 25 04751 g007
Figure 8. MAE loss during training epochs.
Figure 8. MAE loss during training epochs.
Sensors 25 04751 g008
Figure 9. Comparison of denoising results on PolyU (left) and DND (right) datasets at noise levels σ of 5, 10, and 25.
Figure 9. Comparison of denoising results on PolyU (left) and DND (right) datasets at noise levels σ of 5, 10, and 25.
Sensors 25 04751 g009aSensors 25 04751 g009b
Figure 10. Comparison of the impact of network depth.
Figure 10. Comparison of the impact of network depth.
Sensors 25 04751 g010
Figure 11. Comparison of the impact of training epochs.
Figure 11. Comparison of the impact of training epochs.
Sensors 25 04751 g011
Table 1. Results of the comparison of key design structures on the SIDD dataset.
Table 1. Results of the comparison of key design structures on the SIDD dataset.
MethodPSNR/dBSSIM/%FOMTT/s
ID-MSE-WGAN29.2837.140.538-
ADGN26.0187.090.72510.064
CS-PCN38.5390.560.91190.059
Trident GAN38.4090.400.91200.047
Ours39.2991.500.91210.051
Table 2. The results of different design structures ion the SIDD dataset.
Table 2. The results of different design structures ion the SIDD dataset.
MethodPSNR/dBSSIM/%
WNet23.6733.38
XNet12.8737.37
U3Net37.7289.28
Table 3. Calculation results.
Table 3. Calculation results.
ModuleParameter (M)FLOPs (G)Single Epoch Training (min)Single Image Reasoning (ms)
Generator (32 channel)≈ 7.834.2918
Denoiser≈15.668.91838
Discriminator≈4.29.1512
Overall27.6112.23255
Table 4. Environment Configuration.
Table 4. Environment Configuration.
ConfigurationParameter
Operating systemUbuntu 20.04.6 LTS
GPUTesla V100S-PCIE-32 GB
CPUIntel(R) Xeon(R) Gold 5220R CPU @ 2.20 GHz
Deep learning frameworkPyTorch 1.3.1
Python version3.7.4
Table 5. The results on the SIDD dataset.
Table 5. The results on the SIDD dataset.
MethodPSNR/dBSSIM/%FOM/%
σ = 5 σ = 10 σ = 25 σ = 5 σ = 10 σ = 25 σ = 5 σ = 10 σ = 25
DnCNN23.6624.6724.5258.3057.3762.5762.5762.5862.77
BM3D25.6525.7125.6768.5068.6168.7168.5068.6168.71
TWSC26.0227.9527.8262.0261.9563.8267.0266.9567.82
FFDNet33.0733.2833.4862.2562.3062.5072.6772.7572.52
CBDNet30.7735.7138.0276.3786.3086.8077.2577.1677.34
RIDNet32.2532.3132.3681.1680.2880.9779.2779.6579.85
AINDNet32.2232.3232.4180.2180.1581.0379.1180.0180.21
Trident GAN38.1038.4038.6990.1090.4090.1989.1091.2091.19
CS-PCN38.3438.5338.7390.2290.5690.8989.9891.1991.21
Ours38.9139.2939.1091.1391.5092.0191.0391.2191.30
Table 6. The results on the DND dataset.
Table 6. The results on the DND dataset.
MethodPSNR/dBSSIM/%FOM/%
σ = 5 σ = 10 σ = 25 σ = 5 σ = 10 σ = 25 σ = 5 σ = 10 σ = 25
DnCNN32.4732.4332.7579.1379.1179.0768.5768.5167.77
BM3D34.5133.2034.8985.1585.1085.2469.7069.6169.68
TWSC34.1233.1534.8284.0284.9584.8270.0171.3271.82
FFDNet34.2534.4534.5084.2584.6084.5077.5577.9978.50
CBDNet38.0537.7138.0284.0784.2184.0278.7778.7178.72
RIDNet38.2538.2138.2685.1685.2885.0579.7680.0280.15
AINDNet38.3738.4138.3985.2185.0585.1580.2180.1780.35
Trident GAN38.9039.1039.1993.1093.4093.1993.1093.1993.20
CS-PCN39.0139.0239.0993.1193.3093.8993.1193.3093.89
Ours39.2339.4839.3194.0394.4894.5294.0194.2294.36
Table 7. The results of ablation studies on the SIDD dataset.
Table 7. The results of ablation studies on the SIDD dataset.
MethodPSNR/dBSSIM/%
None36.33 ± 0.1085.91 ± 0.11
Dual UNet39.16 ± 0.1590.05 ± 0.17
Dual UNet without ASPP39.22 ± 0.1291.31 ± 0.15
MRDB38.72 ± 0.1191.22 ± 0.14
Ours39.29 ± 0.0591.50 ± 0.09
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, J.; Zhou, Y.; Sun, M.; Wang, D. Dual-Path Adversarial Denoising Network Based on UNet. Sensors 2025, 25, 4751. https://doi.org/10.3390/s25154751

AMA Style

Yu J, Zhou Y, Sun M, Wang D. Dual-Path Adversarial Denoising Network Based on UNet. Sensors. 2025; 25(15):4751. https://doi.org/10.3390/s25154751

Chicago/Turabian Style

Yu, Jinchi, Yu Zhou, Mingchen Sun, and Dadong Wang. 2025. "Dual-Path Adversarial Denoising Network Based on UNet" Sensors 25, no. 15: 4751. https://doi.org/10.3390/s25154751

APA Style

Yu, J., Zhou, Y., Sun, M., & Wang, D. (2025). Dual-Path Adversarial Denoising Network Based on UNet. Sensors, 25(15), 4751. https://doi.org/10.3390/s25154751

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop