You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

15 December 2023

NGDCNet: Noise Gating Dynamic Convolutional Network for Image Denoising

and
Computer School, Beijing Information Science and Technology University, Beijing 100101, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Recent Advances in Object Detection and Image Processing

Abstract

Deep convolution neural networks (CNNs) have become popular for image denoising due to their robust learning capabilities. However, many methods tend to increase the receptive field to improve performance, which leads to over-smoothed results and loss of critical high-frequency information such as edges and texture. In this research, we introduce an innovative end-to-end denoising network named the noise gating dynamic convolutional network (NGDCNet). By integrating dynamic convolution and noise gating mechanisms, our approach effectively reduces noise while retaining finer image details. Through a series of experiments, we conduct a comprehensive evaluation of NGDCNet by comparing it quantitatively and visually against state-of-the-art denoising methods. Additionally, we provide an ablation study to analyze the contributions of dynamic convolutional blocks and noise gating blocks. Our experimental findings demonstrate that NGDCNet excels in noise reduction while preserving essential texture information.

1. Introduction

Due to the intricate acquisition processes involving capturing, transmitting, and compressing images through digital devices, the acquired images are susceptible to variations in lighting conditions and sensor noise [1]. Image noise, which manifests as random fluctuations in brightness or color, poses a challenge to image observation and accurate information extraction in image processing. The detrimental effects of noise can impact a range of image processing tasks, such as video processing, image analysis, and segmentation, potentially leading to erroneous judgments [2]. In response to this challenge, image denoising techniques have been devised.
Image denoising techniques are specifically designed to mitigate the impact of noise. They aim to remove unwanted noise, allowing for the accurate representation and storage of a clean image based on a degenerate model y = c + n , where c represents a potentially clean image, n is the noise, and y corresponds to the noisy image [3]. Building on this understanding, filter methods have gained significant traction for effectively suppressing noise in noisy images [4]. Spatial domain filters, for example, employ pixel averaging within specific regions to enhance the quality of predicted images [5]. Furthermore, combining compensation functions and bilateral filtering can augment edges to extract more detailed information, thus bolstering denoising efficiency [6]. Additionally, based on image characteristics, transforming the image to the transform domain has proven effective in noise removal [7]. Block Matching 3D (BM3D) leverages block comparability assessments to perform transform domain filtering in image denoising [8]. This process involves grouping similar patches, condensing the 3D transform domain for collaborative filtering, and then reconstructing the original position by aggregating these acquired patches. Given BM3D’s outstanding performance, various variants have been developed for denoising tasks [9]. Moreover, perceptual concepts have been employed to extract noisy information from a given noisy image, contributing to denoising techniques [10]. Total variation, an approach involving minimizing the total variation of an image, was introduced in image denoising and has demonstrated effectiveness [11]. Additionally, the partial differential equation (PDE) algorithm treats the denoising process as solving an equation through multiple iterations, approaching a solution close to a potentially clean image [12]. Due to PDE’s properties, its variants are advantageous in preserving edge and texture information during image denoising [13]. Another approach, weighted kernel paradigm minimization (WNNM), utilizes low-rank clustering correlation concepts to filter noise and recover high-quality images [14]. Other popular denoising methods encompass Markov random fields [15] and gradient methods [16].
With the rapid advancement of deep learning, particularly the widespread adoption of convolutional neural networks (CNNs) known for their high speed of execution and strong learning capabilities, the domain of image denoising has witnessed a significant transformation [17]. Over the past decade, numerous CNN-based denoising methods have surfaced, including IRCNN [18], DnCNN [19], and complex-valued deep CNN [20]. However, a common challenge faced by these methods is the depth of the network which makes training arduous. To address the training issue, Tian et al. [21] introduced an enhanced convolutional neural denoising network (ECNDNet), incorporating residual learning [22], dilation convolution, and batch normalization [23] to expedite network convergence. Additionally, combining CNNs with other deep learning techniques has shown promise in achieving improved denoising results. For instance, Zhao et al. [24] proposed a hybrid denoising model that integrates a transformer encoder and a convolutional decoder network, effectively leveraging the strengths of both networks for the successful denoising of real images. Kumwilaisak et al. [25] devised a novel approach based on deep CNNs and multidirectional long short-term memory networks to eliminate pine tree noise from images. Moreover, the fusion of unsupervised learning with CNNs has paved the way for even more versatile applications. Pang et al. [26] introduced an unsupervised deep noise reducer to tackle the issue of overfitting due to the lack of real images.
Despite the success of various deep CNN-based methods in image denoising, many recent approaches [27] aim to achieve enhanced performance by increasing the receptive field. However, this potentially leads to over-smoothed results and loss of critical high-frequency information such as edges and textures [28,29]. Therefore, this paper introduces a novel denoising method called NGDCNet to achieve effective image denoising while being able to better preserve the texture information of the image. The method is based on deep CNN and replaces the traditional static convolution with dynamic convolution [30]. This transformation can aggregate dynamic kernels during the convolution operation to fully utilize the correlation within the deep features [30], giving the network the ability to recover complex high-frequency information. However, dynamic convolution is sensitive to noise [31]. Therefore, this paper proposes a residue-like structure that combines a noise gating mechanism with skip connections. This residual-like structure can effectively control the propagation of low-level features to higher levels through the learning process, thus reducing the interference of low-level redundant features on dynamic convolution.
The main contributions of this study can be summarized as follows:
  • We propose a residual-like structure that combines the noise gating mechanism with skip connections. The residual-like structure effectively refines the ordinary skip connection.
  • We combine dynamic convolution and noise gating mechanisms to propose a simple and effective network architecture that can achieve excellent denoising performance.
This paper follows a structured organization. Section 2 provides a comprehensive review of the related work. Section 3 presents our denoising framework in detail. In Section 4, we conduct a series of comparative experiments to showcase the effectiveness of the proposed denoising methods. Section 5 concludes the work presented in this paper.

3. Proposed Method

In this subsection, we present our proposed denoising network model, comprising noise gating dynamic convolutional block (NGDCB) and residual block (RB). NGDCB is composed of two crucial components: dynamic convolution block (DCB) [30] and noise gating block (NGB). Our main focus for improving the image denoising effect revolves around these two blocks: DCB and NGB.
To elaborate, DCB plays a pivotal role in enabling the network to recover intricate high-frequency information effectively. It achieves this by incorporating five stacked dynamic convolutions. On the other hand, NGB effectively controls the flow of lower features to higher levels, optimizing skip connections. In the subsequent sections, we provide a detailed description of the network model, its constituent blocks, and the pertinent functions involved.

3.1. Network Model

This paper introduces a network model named NGDCNet, which is composed of NGDCB and RB as depicted in Figure 1. To provide a clearer representation of the operational process, we present it in the following equation:
I c = N G D C N e t ( I n ) = R B ( N G D C B ( N G D C B ( N G D C B ( I n ) ) ) )
where I c denotes the predicted clean image, N G D C N e t denotes the function of NGDCNet, and I n denotes the given noisy image. Also, R B and N G D C B are functions of RB and NGDCB, respectively.
Figure 1. Network model of the proposed NGDCNet.

3.2. Loss Function

To enhance the credibility of our denoising network, we chose the L 1 as the loss function for training the model parameters. This process can be expressed using the following formula:
L 1 ( θ ) = 1 N i = 1 N | y i N G D C N e t ( x i ) |
where x i and y i denote the i-th noisy images and given clean images, respectively, and θ denotes the parameters in training the denoising model. In addition, N stands for the total of noisy images.

3.3. Noise Gating Dynamic Convolutional Block

3.3.1. Noise Gating Block

In general, residual learning is usually implemented using simple skip connections. This can be used in many tasks [22]. However, during the denoising process, skip connections may be counterproductive by passing lower redundant features to higher levels. Therefore, this study utilizes a noise gating mechanism [47,48], which controls the transmission of trivial features by learning spatial feature correlations. The noise gating mechanism is obtained as follows:
G a t e i , j = i = 1 H j = 1 W W g · I
F e a t u r e s i , j = i = 1 H j = 1 W W f · I
O u t p u t i , j = δ ( G a t e i , j θ ( F e a t u r e s i , j ) )
where δ is the LeakyReLU activation function, and θ is the sigmoid function.
In this paper, we combine the noise gating mechanism with a skip connection to form a residual-like structure named NGB. The NGB comprises two branches: one with a standard convolutional layer denoted as Conv and the other with Conv + Sigmoid. These branches individually process the input features, and their results are combined through element-wise multiplication (represented as ⊗ in Figure 1. Subsequently, the combined features undergo processing by LeakyReLU and are then summed with the output from D C B . To provide a clearer representation of this process, we use the following equation to describe it:
O N G B = N G B ( I ) = L R ( C ( I ) C S ( I ) )
where L R denotes LeakyReLU, C S denotes 3 × 3 kernel size of Conv + Sigmoid, and N G B denotes noise gating convolutional block. For the first N G B , I denotes the image disturbed by noise. For the middle stacked N G B , I is the result of summing the outputs of the previous N G B and the D C B . O N G B is the output of N G B .

3.3.2. Dynamic Convolutional Block

The main component consists of three stacked DCBs, with each DCB comprising five dynamic convolutional layers [30]. These are primarily aimed at extracting finer high-frequency information. Specifically, the first layer consists of a standard convolutional layer and ReLU to extract lower-level features. It has a kernel size of 3 × 3 , with input and output channels of 1 and 64, respectively (or 3 for color image denoising). ReLU is responsible for transforming linear features into nonlinear features. Each DCB consists of five DynamicConv + BN + ReLU layers. Each DCB has both input and output channels set at 64 where the end of each DCB is connected to a skip connection improved by NGB. To provide a clearer representation of this process, we use the following equation to describe it:
O D C B = D C B ( I ) = D C B R ( D C B R ( D C B R ( D C B R ( D C B R ( C R ( I ) ) ) ) ) )
where C R denotes the 3 × 3 kernel size of Conv + ReLU and D C B R denotes the 3 × 3 kernel size of DynamicConv + BN + ReLU. For the first D C B , I denotes the image disturbed by noise. For the middle stacked D C B , I is the result of summing the outputs of the previous N G B and the D C B . O D C B is the output of D C B .

3.4. Residual Block

The residual block is employed to construct a latent clean image. It encompasses a single convolutional layer, denoted as Conv in Figure 1. The residual block employs a residual learning operation to eliminate noise from the provided image in the following manner:
I c = R B ( C ( O N G D C B ) ) = I n ( C ( O D C B + O N G B ) )
where − is the ⊖ in Figure 1. C means that the kernel size of the convolution is 3 × 3 . O N G D C B is the output of the last N G D C B .

4. Experiments

4.1. Datasets and Implementation Details

For grayscale Gaussian image denoising, the train set was established by converting the 432 training images sourced from the BSD68 [49] dataset into grayscale images. For the subsequent evaluation, we utilized two distinct test sets. The first test set comprised 68 test images from BSD68, while the second test set was derived from the Set12 [50] dataset. We compared the peak signal-to-noise ratio (PSNR) [51] and structural similarity (SSIM) [52] under varying noise conditions, benchmarking these results against the outcomes obtained from other denoising methods.
For color Gaussian image denoising, we utilized the same 432 training images from the BSD68 dataset. We constructed two test sets: the first set consisted of 68 test images from BSD68, and the second set was derived from the Set5 [53] dataset. Within this context, we performed a comparative analysis by comparing the PSNR and SSIM under diverse noise conditions, analyzing them in relation to various denoising methods.
The hardware platform for this experiment is a Tesla V100 with 16 GB of RAM; the software is Ubuntu18.04.5, CUDA10.1, python3.7.0; the deep learning framework is Pytorch1.5.0; the batch processing is 4; the Adam algorithm is used to update the gradient; the initial learning rate is 0.001 , and the learning rate decreases with the number of training times.

4.2. Evaluation Metrics

After image denoising, image quality evaluation criteria are usually introduced to evaluate the merits of denoising models. In this paper, PSNR and SSIM are used for evaluation and analysis.

4.2.1. PSNR

PSNR is a metric that quantifies the ratio of the maximum possible power of a signal to the power of the noise in an image. A higher PSNR value implies that the image has a greater proportion of signal compared to noise, resulting in better image quality. It is measured in decibels (dB), and in general, if the PSNR exceeds 40 dB, the image quality is very good. If M A X i denotes the maximum gray value and M S E [54] denotes the mean square error, the P S N R is expressed by the following formula:
P S N R = 10 log 10 ( M A X i 2 M S E )
M S E = 1 M × N i = 0 M 1 j = 0 N 1 [ c ( i , j ) n ( i , j ) ] 2
where M × N is the total number of pixel values of the two images, respectively, c ( i , j ) is the real clean image, and n ( i , j ) is the denoised image. A smaller M S E value means that the difference between the predicted value and the real value is smaller, which implies that the denoising effect is better and the performance of the model is better.

4.2.2. SSIM

SSIM measures the similarity of two images and is a common metric for image quality assessment [55]. Assuming input images X and Y, the S S I M of images X and Y is computed as follows:
S S I M ( X , Y ) = ( 2 μ X μ Y + a 1 ) ( 2 σ X Y + a 2 ) ( μ X 2 + μ Y 2 + a 3 ) ( σ X 2 + σ Y 2 + a 4 )
where μ X and μ Y are the standard deviation of X and Y, σ X 2 and σ Y 2 are the variance of X and Y, σ X Y is the covariance of X and Y, and a 1 , a 2 , a 3 and a 4 are constants. Generally, the SSIM value is in the range of [ 1 , 1 ] , and the larger the SSIM value, the better the image denoising.

4.3. Experimental Result and Analysis

4.3.1. Comparison with State-of-the-Art Denoising Methods

  • Gray Gaussian Image Denoising
To evaluate the effectiveness of the proposed NGDCNet for denoising gray Gaussian noisy images, we compared it with five existing methods: DnCNN [19], FFDNet [39], IRCNN [18], DRUNet [56], and NIFBGDNet [57]. The image blind denoising experiments, presented in Table 1, involved testing these methods across three noise levels (15, 25, and 50). The quantitative results demonstrate that NGDCNet achieved notably high PSNR and SSIM values on both datasets, indicating its superior denoising performance.
Table 1. Average PSNR (dB) and SSIM of different methods on BSD68 with noise levels of 15, 25, and 50.
The results presented in Table 1 demonstrate the superior performance of NGDCNet in terms of PSNR and SSIM values when compared to the five other methods across the three distinct noise levels. Particularly at a noise level of σ = 15 , our model exhibited substantially higher PSNR and SSIM scores than the other methods. Impressively, even at a higher noise level ( σ = 50 ), our model consistently maintained outstanding performance. It is worth noting that our model achieved these impressive results without needing an additional training set. This robust performance across varying noise levels underscores the NGDCNet’s formidable capability to recover corrupted images effectively.
Table 2 and Table 3 display the test results of NGDCNet in comparison to the other five methods on the Set12 dataset. The results in Table 2 show that NGDCNet consistently achieves higher average PSNR at noise levels of 15, 25, and 50 compared to the other models. Particularly noteworthy is the performance at a noise level of 50, where NGDCNet’s PSNR value surpasses the second-ranked method by 1.04 dB. These results underscore the exceptional denoising effect of NGDCNet at low and medium noise levels, demonstrating its superior ability to maintain image quality in such scenarios. Even at higher noise levels, NGDCNet maintains excellent performance, showcasing its effectiveness across various noise conditions.
Table 2. Average PSNR (dB) of different methods on Set12 with noise levels of 15, 25, and 50.
Table 3. Average SSIM of different methods on Set12 with noise levels of 15, 25, and 50.
In Table 3, it is apparent that NGDCNet achieves higher SSIM values than the other methods at noise levels of 15 and 25. While at a noise level of 50, the SSIM value of NGDCNet is slightly lower than that of DRUNet, indicating NGDCNet’s ability to maintain image structure effectively in most situations.
  • Color Gaussian Image Denoising
Table 4 displays the test results of NGDCNet in comparison with other methods on the CBSD68 [49] dataset. It is evident that NGDCNet exhibits remarkable performance in color Gaussian image denoising. NGDCNet consistently outperforms other methods, as indicated by its higher PSNR and SSIM values across various noise levels. Notably, NGDCNet achieves an impressive SSIM of 0.8717 at a noise level of 50, showcasing its ability to preserve image structure even under very high noise conditions.
Table 4. Average PSNR (dB) and SSIM of different methods on CBSD68 with noise levels of 15, 25, and 50.
Table 5 demonstrates the test results of NGDCNet compared to other methods on the Set5 dataset. These results reveal that NGDCNet consistently achieves higher average PSNR and SSIM values compared to other methods at noise levels of 15 and 25. This underscores the exceptional performance of NGDCNet, particularly in situations involving low to medium noise levels. Impressively, at the noise level of 50, NGDCNet achieves a PSNR value of 30.52 dB, surpassing DRUNet by 0.65 dB.
Table 5. Average PSNR (dB) of different methods on Set5 with noise levels of 15, 25, and 50.
Table 6 further supports the superiority of NGDCNet, with its SSIM values outperforming other methods across all three noise levels. Even at the noise level of 50, the SSIM value of NGDCNet exceeds the second-ranked method by 0.0336 , indicating NGDCNet’s ability to effectively preserve image structure under varying noise conditions.
Table 6. Average SSIM of different methods on Set5 with noise levels of 15, 25, and 50.
Table 7 compares the parameter-based denoising methods. It is worth noting that DRUNet has the largest number of parameters and IRCNN has the smallest number of parameters. The number of parameters of FFDNet, DnCNN, NIFBGDNet, and NGDCNet are in the middle of the range. From the experimental results, DRUNet has the best denoising performance compared to FFDNet, DnCNN, NIFBGDNet, and IRCNN, while DRUNet also has the highest model complexity. NGDCNet surpasses these methods in denoising performance and achieves a desirable balance between model complexity and denoising effect.
Table 7. The number of parameters of different denoising methods.

4.3.2. Comparison of Visual Effects

To assess the denoising performance of NGDCNet, we conducted a visual comparison with several competing methods, including FFDNet [39], DnCNN [19], DRUNet [56], IRCNN [18], and NIFBGDNet [57]. Our method was evaluated by selecting specific images from the test set and zooming into a region of the predicted image for visual observation. A clearer observation region indicated superior denoising performance for the corresponding method.
For grayscale Gaussian image denoising, a visual comparison was conducted using selected images from the BSD68 [49] and Set12 [50] datasets. As shown in the results presented in Figure 2, at a noise level of 25, most of the other denoising methods were effective in reducing noise. However, these methods tended to overly smooth textured areas, compromising the texture details. When the noise level was increased to 50, as depicted in Figure 3, the limitations of other methods became more apparent. The enlarged portion of the images revealed that the results from these methods were unsatisfactory, with significant corruption of texture information. While DRUNet achieved clarity in the overall view, a closer examination revealed distortion in some of the smaller textures. In contrast, NGDCNet, while appearing slightly blurred overall, effectively preserved the original texture information in the image, as evident from the enlarged portion.
Figure 2. Denoising results of a gray noisy image from BSD68 [49] with σ = 25. (a) Clean image, (b) Noisy image (20.43 dB/0.3973), (c) FFDNet (29.24 dB/0.8028), (d) DnCNN (29.38 dB/0.8040), (e) IRCNN (29.32 dB/0.8015), (f) NIFBGDNet (29.09 dB/0.7930), (g) DRUNet (29.66 dB/0.8150), (h) Ours (31.24 dB/0.8674).
Figure 3. Denoising results of a gray noisy image from Set12 with σ = 50. (a) Clean image, (b) Noisy image (14.90 dB/0.2869), (c) FFDNet (25.49 dB/0.7715), (d) DnCNN (25.60 dB/0.7715), (e) IRCNN (25.48 dB/0.7675), (f) NIFBGDNet (25.45 dB/0.7640), (g) DRUNet (26.49 dB/0.8013), (h) Ours (28.49 dB/0.8603).
For color Gaussian image denoising, the comparative denoising results are presented in Figure 4 and Figure 5. Upon closer examination, it was evident that our model effectively removed the majority of noise while preserving more detailed texture information, a feat not achieved by other methods. These visual comparisons further underscore the superior denoising performance of NGDCNet.
Figure 4. Denoising results of a color noisy image from Set5 with σ = 25. (a) Clean image, (b) Noisy image (20.47 dB/0.6689), (c) FFDNet (31.04 dB/0.9379), (d) DnCNN (31.08 dB/0.9360), (e) IRCNN (30.95 dB/0.9342), (f) NIFBGDNet (30.84 dB/0.9348), (g) DRUNet (32.31 dB/0.9501), (h) Ours (32.87 dB/0.9614).
Figure 5. Denoising results of a color noisy image from CBSD68 with σ = 50. (a) Clean image, (b) Noisy image (14.68 dB/0.3197), (c) FFDNet (28.29 dB/0.7577), (d) DnCNN (28.24 dB/0.7603), (e) IRCNN (28.19 dB/0.7614), (f) NIFBGDNet (27.96 dB/0.7507), (g) DRUNet (28.81 dB/0.7814), (h) Ours (30.27 dB/0.8541).

4.3.3. Ablation Experiments

In this section, the effects of DCB and NGB on NGDCNet’s performance were studied. Experiments were conducted on grayscale and color Gaussian images with noise levels of 15, 25, and 50. In order to evaluate the contributions of DCB and NGB, these components were individually removed from NGDCNet’s architecture for comparison. Specifically, for DCB, × means replacing the dynamic convolution in DCB with the standard convolution using the 3 × 3 kernel. For NGB, × means to remove NGB and just do a normal skip connection.
Table 8 and Table 9 present the results of these ablation experiments for grayscale and color Gaussian images, respectively. The tables demonstrate that using either DCB or NGB alone results in improved performance for NGDCNet in both tasks. This suggests that DCB’s capacity to efficiently extract high-frequency information from features enhances the model’s performance in image denoising. NGB is effective in controlling the propagation of low-level feature information to higher levels, contributing to performance improvements in the denoising task.
Table 8. Study of ablation experiments of different components in gray scale image denoising on BSD68 [49] test set.
Table 9. Study of ablation experiments of different components in color scale image denoising on CBSD68 test set.
From the results in these tables, it can be observed that NGDCNet using DCB alone provides limited improvements in denoising performance because dynamic convolution is sensitive to noise [31]. In contrast, NGB can effectively filter information in the lower layers. When both DCB and NGB are combined, NGDCNet’s denoising performance is significantly enhanced. The last row in these tables shows the results of NGDCNet with combined DCB and NGB, demonstrating the effectiveness of this combined approach.

5. Conclusions

In this paper, we introduced a novel image denoising method, NGDCNet, which leverages a noise gating mechanism and dynamic convolution. NGDCNet is constructed through a collaboration of dynamic convolution blocks, noise gating blocks, and residual blocks. The dynamic convolution blocks employ stacked dynamic convolutions to capture high-frequency information within the features. The noise gating blocks effectively manage the flow of low-level feature information to higher layers, enhancing the model’s ability to process noise. Additionally, the residual blocks leverage residual operations to generate a denoised image.
Our proposed NGDCNet demonstrates impressive denoising performance, particularly in preserving image structure and fine details, even in high noise conditions. In future work, we aim to further evaluate NGDCNet in real image denoising tasks, extending its application to various practical scenarios. We believe that NGDCNet holds great promise for enhancing image quality and addressing real-world denoising challenges.

Author Contributions

Conceptualization, M.Z. and Z.L.; methodology, M.Z. and Z.L.; software, M.Z. and Z.L.; validation, M.Z. and Z.L.; formal analysis, M.Z. and Z.L.; investigation, M.Z. and Z.L.; resources, M.Z.; data curation, M.Z. and Z.L.; writing—original draft preparation, M.Z. and Z.L.; writing—review and editing, M.Z. and Z.L.; visualization, M.Z. and Z.L.; supervision, M.Z.; project administration, M.Z.; funding acquisition, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Qiyuan Innovation Foundation (S20210201067) and sub-themes (No. 9072323404).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Training and test sets and experimental results in the paper: https://github.com/OceanL008/NGDCNet.git (accessed on 3 November 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Healey, G.E.; Kondepudy, R. Radiometric CCD camera calibration and noise estimation. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 267–276. [Google Scholar] [CrossRef]
  2. Diwakar, M.; Kumar, M. A review on CT image noise and its denoising. Biomed. Signal Process. Control 2018, 42, 73–88. [Google Scholar] [CrossRef]
  3. Gurrola-Ramos, J.; Dalmau, O.; Alarcón, T.E. A residual dense u-net neural network for image denoising. IEEE Access 2021, 9, 31742–31754. [Google Scholar] [CrossRef]
  4. Lee, J.S. Digital image enhancement and noise filtering by use of local statistics. IEEE Trans. Pattern Anal. Mach. Intell. 1980, PAMI-2, 165–168. [Google Scholar] [CrossRef] [PubMed]
  5. Kuan, D.T.; Sawchuk, A.A.; Strand, T.C.; Chavel, P. Adaptive noise smoothing filter for images with signal-dependent noise. IEEE Trans. Pattern Anal. Mach. Intell. 1985, PAMI-7, 165–177. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, B.; Allebach, J.P. Adaptive bilateral filter for sharpness enhancement and noise removal. IEEE Trans. Image Process. 2008, 17, 664–678. [Google Scholar] [CrossRef]
  7. Alkinani, M.H.; El-Sakka, M.R. Patch-based models and algorithms for image denoising: A comparative review between patch-based images denoising methods for additive noise reduction. EURASIP J. Image Video Process. 2017, 2017, 58. [Google Scholar] [CrossRef]
  8. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  9. Yang, D.; Sun, J. BM3D-Net: A convolutional neural network for transform-domain collaborative filtering. IEEE Signal Process. Lett. 2017, 25, 55–59. [Google Scholar] [CrossRef]
  10. Alrudaini, J.K.; Hayder, D.M.; Hamzah, A.K.; Ruhaina, A.A. Visual Perception Method for Medical Image De-noising. Malay. J. Med. Health Sci. 2022, 18, 40–44. [Google Scholar]
  11. Chambolle, A. An algorithm for total variation minimization and applications. J. Math. Imaging Vis. 2004, 20, 89–97. [Google Scholar]
  12. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef]
  13. Li, Q.; Jiang, L. A novel variable-separation method based on sparse and low rank representation for stochastic partial differential equations. SIAM J. Sci. Comput. 2017, 39, A2879–A2910. [Google Scholar] [CrossRef]
  14. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  15. Lan, X.; Roth, S.; Huttenlocher, D.; Black, M.J. Efficient belief propagation with learned higher-order markov random fields. In Computer Vision, Proceedings of the ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; Proceedings, Part II 9; Springer: Berlin/Heidelberg, Germany, 2006; pp. 269–282. [Google Scholar]
  16. Li, P.; Liang, J.; Zhang, M.; Fan, W.; Yu, G. Joint image denoising with gradient direction and edge-preserving regularization. Pattern Recognit. 2022, 125, 108506. [Google Scholar] [CrossRef]
  17. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef]
  18. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning deep CNN denoiser prior for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3929–3938. [Google Scholar]
  19. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  20. Quan, Y.; Chen, Y.; Shao, Y.; Teng, H.; Xu, Y.; Ji, H. Image denoising using complex-valued deep CNN. Pattern Recognit. 2021, 111, 107639. [Google Scholar] [CrossRef]
  21. Tian, C.; Xu, Y.; Fei, L.; Wang, J.; Wen, J.; Luo, N. Enhanced CNN for image denoising. CAAI Trans. Intell. Technol. 2019, 4, 17–23. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  23. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  24. Zhao, M.; Cao, G.; Huang, X.; Yang, L. Hybrid transformer-CNN for real image denoising. IEEE Signal Process. Lett. 2022, 29, 1252–1256. [Google Scholar] [CrossRef]
  25. Kumwilaisak, W.; Piriyatharawet, T.; Lasang, P.; Thatphithakkul, N. Image denoising with deep convolutional neural and multi-directional long short-term memory networks under Poisson noise environments. IEEE Access 2020, 8, 86998–87010. [Google Scholar] [CrossRef]
  26. Pang, T.; Zheng, H.; Quan, Y.; Ji, H. Recorrupted-to-recorrupted: Unsupervised deep learning for image denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2043–2052. [Google Scholar]
  27. Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
  28. Zou, X.; Xiao, F.; Yu, Z.; Li, Y.; Lee, Y.J. Delving deeper into anti-aliasing in convnets. Int. J. Comput. Vis. 2023, 131, 67–81. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Li, K.; Li, K.; Sun, G.; Kong, Y.; Fu, Y. Accurate and fast image denoising via attention guided scaling. IEEE Trans. Image Process. 2021, 30, 6255–6265. [Google Scholar] [CrossRef] [PubMed]
  30. Chen, Y.; Dai, X.; Liu, M.; Chen, D.; Yuan, L.; Liu, Z. Dynamic convolution: Attention over convolution kernels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11030–11039. [Google Scholar]
  31. Shen, H.; Zhao, Z.Q.; Zhang, W. Adaptive dynamic filtering network for image denoising. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 2227–2235. [Google Scholar]
  32. Ma, R.; Li, S.; Zhang, B.; Fang, L.; Li, Z. Flexible and generalized real photograph denoising exploiting dual meta attention. IEEE Trans. Cybern. 2022, 53, 6395–6407. [Google Scholar] [CrossRef] [PubMed]
  33. Ma, R.; Li, S.; Zhang, B.; Hu, H. Meta PID attention network for flexible and efficient real-world noisy image denoising. IEEE Trans. Image Process. 2022, 31, 2053–2066. [Google Scholar] [CrossRef] [PubMed]
  34. Ma, R.; Li, S.; Zhang, B.; Li, Z. Towards fast and robust real image denoising with attentive neural network and PID controller. IEEE Trans. Multimed. 2021, 24, 2366–2377. [Google Scholar] [CrossRef]
  35. Ma, R.; Hu, H.; Xing, S.; Li, Z. Efficient and fast real-world noisy image denoising by combining pyramid neural network and two-pathway unscented Kalman filter. IEEE Trans. Image Process. 2020, 29, 3927–3940. [Google Scholar] [CrossRef]
  36. Ma, R.; Li, S.; Zhang, B.; Li, Z. Generative adaptive convolutions for real-world noisy image denoising. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2022; Volume 36, pp. 1935–1943. [Google Scholar]
  37. El Helou, M.; Süsstrunk, S. Blind universal Bayesian image denoising with Gaussian noise level learning. IEEE Trans. Image Process. 2020, 29, 4885–4897. [Google Scholar] [CrossRef]
  38. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  39. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
  40. Lan, R.; Zou, H.; Pang, C.; Zhong, Y.; Liu, Z.; Luo, X. Image denoising via deep residual convolutional neural networks. Signal Image Video Process. 2021, 15, 1–8. [Google Scholar] [CrossRef]
  41. Xu, M.; Xie, X. NFCNN: Toward a noise fusion convolutional neural network for image denoising. Signal Image and Video Process. 2022, 16, 175–183. [Google Scholar] [CrossRef]
  42. Ren, C.; He, X.; Wang, C.; Zhao, Z. Adaptive consistency prior based deep network for image denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8596–8606. [Google Scholar]
  43. Li, B.; Liu, X.; Hu, P.; Wu, Z.; Lv, J.; Peng, X. All-in-one image restoration for unknown corruption. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17452–17462. [Google Scholar]
  44. Zhang, K.; Li, Y.; Zuo, W.; Zhang, L.; Van Gool, L.; Timofte, R. Plug-and-play image restoration with deep denoiser prior. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 6360–6376. [Google Scholar] [CrossRef] [PubMed]
  45. Zhang, Q.; Xiao, J.; Tian, C.; Chun-Wei Lin, J.; Zhang, S. A robust deformed convolutional neural network (CNN) for image denoising. CAAI Trans. Intell. Technol. 2023, 8, 331–342. [Google Scholar] [CrossRef]
  46. Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, D. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
  47. Yu, J.; Lin, Z.; Yang, J.; Shen, X.; Lu, X.; Huang, T.S. Free-form image inpainting with gated convolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 4471–4480. [Google Scholar]
  48. Chang, Y.L.; Liu, Z.Y.; Lee, K.Y.; Hsu, W. Free-form video inpainting with 3d gated convolution and temporal patchgan. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 9066–9075. [Google Scholar]
  49. Roth, S.; Black, M.J. Fields of experts: A framework for learning image priors. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 860–867. [Google Scholar]
  50. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  51. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  52. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  53. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the 23rd British Machine Vision Conference (BMVC), Surrey, UK, 3–7 September 2012. [Google Scholar]
  54. Ephraim, Y.; Malah, D. Speech enhancement using a minimum-mean square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1984, 32, 1109–1121. [Google Scholar] [CrossRef]
  55. Ding, K.; Ma, K.; Wang, S.; Simoncelli, E.P. Image quality assessment: Unifying structure and texture similarity. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2567–2581. [Google Scholar] [CrossRef]
  56. Devalla, S.K.; Renukanand, P.K.; Sreedhar, B.K.; Subramanian, G.; Zhang, L.; Perera, S.; Mari, J.M.; Chin, K.S.; Tun, T.A.; Strouthidis, N.G.; et al. DRUNET: A dilated-residual U-Net deep learning network to segment optic nerve head tissues in optical coherence tomography images. Biomed. Opt. Express 2018, 9, 3244–3265. [Google Scholar] [CrossRef]
  57. Thakur, R.K.; Maji, S.K. Multi scale pixel attention and feature extraction based neural network for image denoising. Pattern Recognit. 2023, 141, 109603. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.