You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

2 March 2023

Deep Signal-Dependent Denoising Noise Algorithm

,
and
1
The Higher Educational Key Laboratory for Measuring & Control Technology and Instrumentations of Heilongjiang Province, Harbin University of Science and Technology, Harbin 150080, China
2
School of Information Engineering, Quzhou College of Technology, Quzhou 324000, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Deep Learning in Image Processing and Pattern Recognition

Abstract

Although many existing noise parameter estimations of image signal-dependent noise have certain denoising effects, most methods are not ideal. There are some problems with these methods, such as poor noise suppression effects, smooth details, lack of flexible denoising ability, etc. To solve these problems, in this study, we propose a deep signal-dependent denoising noise algorithm. The algorithm combines the model method with a convolutional neural network. We use the noise level of the noise image and the noise image together as the input of the convolutional neural network to obtain a wider range of noise levels than the single noise image as the input. In the convolutional neural network, the deep features of the image are extracted by multi-layer residuals, which solves the difficult problem of training. Extensive experiments demonstrate that our noise parameter estimation has good denoising performance.

1. Introduction

Image denoising is a classic and indispensable research topic in low-level vision tasks, which is the premise of high-level vision tasks []. The clear and high-quality images obtained by denoising serve the high-level vision and make completing the tasks better. This paper mainly introduces the denoising of signal-dependent noisy images.
With the continuous updating of digital image acquisition technology, equipment for acquiring images and the number of acquired images are also increasing, and people’s requirements for image quality are also increasing. However, due to the imperfections of imaging systems, imaging equipment, and transmission media, many different types of noise are introduced during the formation of images in different devices, which can affect the imaging effect, and thus the quality of the image. Noise-contaminated images will have a considerable impact on subsequent image processing. Noise pollution will affect the quality of subsequent image processing, such as image recognition, segmentation, classification, and so on [].
In this paper, we model the signal-independent noise in the image, and then input it into the denoising network to solve the problem. During training and testing, networks that take the noise level of the input image and the noise image as input have a wider range of noise levels than networks with a single noise image as the input. The success of CNN denoisers is significantly dependent on whether the distributions of synthetic and real noises are well matched. Therefore, realistic noise models are the foremost issue for blind denoising of real photographs. According to [], signal-dependent noise can be modeled by signal-dependent Gaussian noise distribution, and the signal-dependent Gaussian model has been considered as a more appropriate alternative than other models for real raw noise modeling. In addition, it is necessary to further optimize the graph after noise modeling. Because there are two unavoidable errors in modeling, the first one is the error generated in the estimation of noise parameters. The second is when pixel values are obtained by the local statistical feature method, which also has errors. Thus, after acquiring the preliminary picture, we use the preliminary image and the noise image together as the input into the denoising network. The two errors are reduced through the network.
At present, image denoising relies on the image degradation model, but most of the image degradation model algorithms are simple and do not match the complex noise in the real image. It leads to the less ideal denoising effect. Therefore, it is necessary to design a more reasonable model or an algorithm with better performance. This paper proposes a method combining the model method and the convolutional neural network. The preliminary image obtained by the model method is used as a part of the network’s input, and the final image is output through the continuous iteration of the network.
Most image denoising is Gaussian noise denoising. The variance of the noise is constant and does not change with the position and intensity of the pixel, the noise level is the only parameter required for modeling. The model of this degradation process is generally defined as:
y = x + n
where  x  is the degraded image pixel value,  y  is the original image pixel value,  n  is Additive White Gaussian Noise (AWGN) with a standard deviation of  σ . Since Gaussian white noise has only one unknown parameter, the variance of the noise, DnCNN enumerates the variance of different values in the process of training the network model [], meaning that the network can remove AWGN in the case of unknown noise variance. However, because there are other noises in the real image, the Gaussian modeling of the image cannot achieve a good denoising effect.
Alessandro Foi [] modeled the real noise as a signal-dependent Poisson–Gaussian noise model and proposed an estimation method for Poisson–Gaussian noise parameters. However, he trimmed the underexposure and overexposure data when experimenting. Despite achieving good performance, the method is limited to Poisson–Gaussian noise due to modeling problems. The model needs to know the Gaussian and Poisson components a and b, the model is as follows:
z ( v ) = k ( v ) + σ ( k ( v ) ) ζ ( v )
σ 2 ( y ( v ) ) = a y ( v ) + b
where  v  is the pixel position in the domain.  z ( v )  is the observed signal, and  k ( v )  is the original signal.  ζ ( v )  is zero-mean independent random noise with a standard deviation equal to 1.  σ  is a function that gives the standard deviation of the overall noise component. The variance is an expression for the Poisson Gaussian noise component from which estimates of  a  and  b  are obtained.
Liu et al. [] proposed a generalized signal-dependent noise model.
g = f + f γ u + w
σ ( f ) = f 2 γ σ u 2 + σ w 2
where  g  is the noisy pixel value,  f  is the noise-free pixel value, and  u  and  w  are zero-mean Gaussian variables.  γ  is an exponential parameter that controls signal dependence by changing the three NLF parameters  γ σ u 2 , and  σ w 2 , and the noise model can represent various types of noise by changing these three parameters. Existing models usually assume that one parameter is known, and only estimate the remaining two parameters to simplify the problem. This method requires three parameters; thus, the amount of calculation is large.
The model we propose in this paper is a noise parameter estimation for real camera noise, which can be modeled with a signal-dependent Gaussian distribution [].
x p ~ N ( y p , σ r 2 + σ s y p )
where  x p  is the degraded image pixel value,  y p  is the original image pixel value, and the noise parameters  σ s  and  σ r  are fixed and only change with the gain of the sensor. The main reasons for adopting this model are as follows: similar to the generalized noise model shown in [], the exponential coefficient will make the calculation of the numerical solution difficult, and the number of unknown parameters in the Gaussian correlated noise model is less than that of the generalized signal-dependent noise model, thus the estimation difficulty will be much less. The model we use is more accurate in parameter estimation than the Poisson–Gaussian noise model, and the denoising effect is better.
In this paper, we propose a deep signal-dependent denoising noise algorithm. The algorithm further improves the noise suppression ability and improves the visual effect of the restored image. The denoising algorithm in this paper can effectively improve image quality, which plays an important role in many fields. First, we find a more reasonable model for the signal-related noise image, which has a wider range of application scenarios and a higher accuracy of parameter estimation. Second, we combine the model with the convolutional neural network to further improve the image-denoising effect. The algorithm further improves the noise suppression ability and improves the visual effect of the restored image. The denoising algorithm in this paper can effectively improve the image quality, which has a wide range of applications for the multimedia, military, medical, and other industries.

3. Proposed Methods

Synthetic noise images are used in this paper. The most important noises in real images are Poisson noise (shot noise) and Gaussian noise. Some scholars have proposed a simple and practical noise model based on the original data of digital imaging sensors []. The signal-dependent noise model gives a function consisting of the Poisson part of the modeled photon sensing and the Gaussian part of the remaining stationary perturbations in the output data. Two parameters of a given noisy image model can be estimated by the image restoration algorithm based on local statistical features. Overall architecture of the algorithm is shown in Figure 1.
Figure 1. Overall architecture of the algorithm.

3.1. The Image Restoration Algorithm on Local Statistical Features

In this paper, the Gaussian correlation noise model shown in Equation (6) is used to model the image degradation process. The estimated values of the two parameters are obtained by estimating the parameters of the noise model, and then the preliminary image is obtained through local statistical features.

3.1.1. Noise Model Parameter Estimation

First, two important assumptions are given: first, according to [], the statistical features of a pixel in a degraded image with signal-dependent noise can be derived from a local region centered at that pixel. Second, there are a certain number of locally homomorphic sub-blocks in the original image, and the variance of the pixel intensity in the sub-blocks is approximately 0. Assuming that the pixel blocks of  y p  and its adjacent areas are homomorphic sub-blocks, the brightness of the original pixels in the homomorphic sub-blocks are  y p  and  E ( y p ) = y p . The equation represents the local mean of the homomorphic subblock image without noise as equal to the local mean of noisy image. According to the modeling formula of  x p , we can acquire:
D ( x p ) = σ r 2 + σ s 2 y p
E ( x p ) = y p
where  E  is local mean,  D  is variance of noise,  x p  is the degraded image pixel value,  y p  is the original image pixel value, and  σ s  and  σ r  are the noise parameters. The two formulas can be simplified to acquire:
D ( x p ) = σ r 2 + σ s 2 E ( x p )
Equation (9) is derived from Equations (7) and (8). Equation (7) is a quadratic equation of  σ s  and  σ r , which has an infinite number of solutions about the variables  σ s  and  σ r . To obtain the unique solution of  σ s  and  σ r , it is necessary to select multiple homomorphic sub-blocks, then a set of linear equations composed of binary linear equations in the form of Equation (9) can be obtained. The number of sub-equations depends on the homomorphic sub-blocks quantity. Typically, the number of image homomorphic sub-blocks are more than two, thus the system of equations is an overdetermined system of equations. The overdetermined systems of equations are expressed.
p H [ σ r 2 + σ s 2 E ( x p ) D ( x p ) ] 2 = 0
In this paper, Equation (10) is solved by the least squares method. By taking the partial derivative of  σ s  and  σ r  in Equation (10), we can acquire both of the equations in Equation (11). The solutions of  σ s  and  σ r  are transformed into solving the system of linear equations shown below. We polluted the images of the Set12 dataset with noise  σ s  = 2.0 × 10−2 σ r  = 10−1, and then estimated the average value of noise parameters obtained by Equation (11) to be  σ s  = 1.978 × 10−2 σ r  = 1.15 × 10−1, with an error within 0.02.
{ p H [ σ r 3 + σ s 2 σ r E ( x p ) σ r D ( x p ) ] = 0 p H [ σ s σ r 2 E ( x p ) + σ s 3 E 2 ( x p ) σ s D ( x p ) E ( x p ) ] = 0
At this point, the unknown parameters  σ s  and  σ r  of the model have unique solutions that satisfy the least squares method.

3.1.2. The Image Restoration Algorithm on Local Statistical Features

It can be seen from the above reasoning that the variance corresponding to the observations of the Gaussian signal-dependent noise model is as follows:
σ 2 ( x p ) = σ r 2 + σ s 2 y p
Arranging the above formula will result in  y p = σ 2 ( x p ) σ r 2 σ s 2 . Since the parameters  σ s  and  σ r  can be estimated by solving the above equations, just solve for  σ 2 ( x p )  to acquire the preliminary image. According to the assumption proposed by Lee [], the mean  μ ( x p )  and variance  σ 2 ( x p )  of  x p  are determined by the mean and variance of the observations in the neighborhood pixels point of the center pixel p. Thus,  σ 2 ( x p )  can be obtained as follows:
σ 2 ( x p ) = 1 M t q W p ( x q μ ( x p ) ) 2
where  M t  is the number of pixels in the neighborhood. Substituting the expression of  σ 2 ( y p )  into the above formula will result in:
y p = 1 σ s 2 M t q W p ( x q μ ( x p ) ) 2 σ r 2 σ s 2

3.2. Convolutional Neural Network

The preliminary image obtained by the above image restoration algorithm based on the statistical characteristics of the local regions inevitably has two errors. The first error is caused when estimating the parameters  σ s  and  σ r . The second error is brought about by the local statistical features of the selected pixels. Therefore, the denoising effects of the preliminary image are not ideal. The following two errors caused by the above algorithm are reduced by the convolutional neural network.

3.2.1. Network Design

The network in this paper employed a simple design, which can reduce the amount of calculation and speed up the training while completing the task. The function of the designed network is mainly to reduce the error of the preliminary image. The convolutional network structure is illustrated in Figure 2. The network takes the preliminary image and noisy image as input, and then extracts the shallow features of the image by a convolutional layer, and then extracts the deep features of the image through the residual network. The residual network is composed of a set of 6 residual blocks. Deeper networks are used to extract deep features. However, after the network layer is deepened, the gradient disappearance and gradient explosion make the training of the network meaningless. Therefore, the residual network is used to avoid the problem, which not only ensures the effect but also controls speed. After the residual network extracts the deep-level features, the data are up_sampled using a up_convolutional layer and the image is returned to the original size. The last convolution layer fuses the features into a single-channel image. The size of the convolution kernel used in the first convolutional layer and the last convolutional layer are both 9 × 9, and the size of the convolution kernel used by all the remaining convolutional layers is 3 × 3. The convolution layers in the residuals block are all 128 convolution kernels with a stride of 1. Only in this way can each layer in the residual network output the same size as the feature map. It helps to fuse features. The LeakyReLU activation function with a slope of 0.2 is employed in the paper, which fixes the problem that the non-positive gradient of ReLU is 0, causing some parameters never to be updated. This activation function has a small positive slope in the negative area, thus it can backpropagate even with negative input values. Furthermore, after experiments, it is found that batch normalization is hardly helpful for the denoising of real images. Part of the reason is that the real noise distribution is fundamentally different from the Gaussian distribution. The architecture of the proposed network is shown in Figure 3a, and the residual connection is shown in Figure 3b. Table 1 shows the parameter settings of Convolutional Neural Networks.
Figure 2. Convolutional network diagram.
Figure 3. The designed network details of each convolutional layer. (a) Convolutional neural network structure. (b) Residual connection.
Table 1. The parameter settings of Convolutional Neural Network.

3.2.2. Loss Function Design

The mean squared error (MSE) below is used as the objective function (the loss function) to measure the difference between the predicted image  R ( Y j )  and the corresponding ground truth value  Y j X j X j  represents the jth clean image.  θ  represents the parameters of the trained model in the network.  { ( Y j , X j ) } j = 1 N  represents there are N image pairs of predicted images and real images. Potentially clean images are recovered by the Adam optimizer using a loss function.
L ( θ ) = 1 2 N j = 1 N | | R ( Y j , θ ) ( Y j X j ) | | 2 2

4. Experiment

In this section, we assess the performance of the proposed method and compare it with other methods to show the improvement of the method relative to other methods.
The experimental environment comprised an Intel Core i7-9700k CPU, 128 GB RAM, and a NVIDIA GeForce RTX 2070 GPU. In addition, the deep learning framework used was Pytorch.
The paper uses the Berkeley Segmentation Dataset (Berkeley segmentation dataset, BSD) (BSD500 for short) as the network training dataset. The BSD500 dataset contains 500 grayscale images, and all images are of size 481 × 321. In the experiment, 200 sets of noise are added to each image to construct an image with real noise. Therefore, the training dataset consists of 105 training samples in total. The noise parameters  σ s  and  σ r  are randomly sampled from [10−3, 10−2] to [10−2.5, 10−1], respectively.
In the selection of test data sets, we use Kodak, McMaster, and Set12 as test data sets, and compare the more advanced methods proposed on these data sets. Although KPN [] is designed for multi-frame image input, it can be adjusted to a single-frame image by changing the network input to compare with our method.
For comparison with other methods, we evaluated the noise levels corresponding to three fixed sets of  σ s  and  σ r . Similar to [], gamma correction was performed first, then we added signal-dependent Gaussian noise, took the generated noise image as the input:  N ( 0 , σ r 2 + σ s y p ) , where  y p  was the intensity of image pixels and the noise parameters  σ s  and  σ r  were randomly sampled from [10−3, 10−2] to [10−2.5, 10−1]. Furthermore, similar to [], the noise level was estimated as:  σ r 2 + σ s y p . Note that the noise Equation (6) assigned three single sets of parameter values, i.e.,  σ 1 ( σ s  = 2.0 × 10−2 σ r  = 10−1),  σ 2 ( σ s  = 6.0 × 10−2 σ r  = 2 × 10−1), and  σ 3 ( σ s  = 9.0 × 10−2 σ r  = 3 × 10−1), to generate clean-noisy image pairs for image denoising. We claim that harnessing noise Equation (6) with combinations of different parameter settings may lead to better results. In spite of that, it was found that the parameter setting as above is a more robust candidate than several other choices.
Similar to most CNNs used for image denoising, the optimizer chooses the adaptive moment estimation (Adam) method in setting network parameters, and the loss function is MSE. The whole training process contains 180 epochs, and the batch size is set to 20, thus the training data set can be divided into 5000 batches. However, for each epoch, we only sample the first 3000 batches for training. Before the next epoch training, we shuffle the order of all batches and select the top 3000 batches for training. The training adopts a linear piecewise learning rate, the learning rate of the top 90 epochs is 10−3, the next 60 epochs are 10−4, and the last 30 epochs are 10−5.
To test the effectiveness of each block for our network on image denoising, we used ablation experiments to analyze the performance. The images in the Set12 dataset were selected for the experiment, and the noise level  σ 1 ( σ s  = 2.0 × 10−2 σ r  = 10−1) was added. Then, we denoised the images separately in groups. The denoising results are shown in Table 2, which shows the PSNR. The results show that residuals made the denoising effect better. Additionally, the addition of preliminary image made the denoising effect better.
Table 2. The average PSNR(DB) results of different methods on three datasets. (a) Remove three residual blocks. (b) Remove all residual blocks. (c) Remove the preliminary image from the input. (d) Original network.
The paper selects four methods: KPN [], CBDNet [], SRNet [], and WCNN []. The link to the code is posted in references [,,,]. Five methods are compared on all the datasets used in the paper. KPN generates a stack of per-pixel filter kernels that jointly aligns, averages, and denoises a burst to produce a clean version of the reference frame. The establishment of the CBD model takes into account both the Poisson–Gaussian model and the in-camera processing pipeline. However, the Poisson–Gaussian model does not have a wider range of scenarios than the signal-dependent Gaussian model in the paper. Each subnetwork of WCNN has its input and output and is supervised by its loss function to capture image features with a specific direction and scale, allowing the WCNN to have sufficient convolutional layers to capture image features while avoiding vanishing/exploding gradients. SRNet embeds the CNN into the sparse representation framework. In each phase, two subnetworks are designed with MSR block to model the updating of the sparse coefficient and image, respectively. However, it suffers from performance bottlenecks and large time consumption.
Figure 4a–d is a diagram of each stage of the paper. To see the contrast, we enlarge the green part of the image into the red part. It can be seen that the images have more detail loss and less clarity even if the preliminary image can see the denoising effect. The denoising effect of the restored image is better. The picture detail loss is less, the recovery effect is better.
Figure 4. The denoising process of our algorithm. (a) Original image. (b) Noise image ( σ 3 ( σ s  = 9.0 × 10−2 σ r  = 3 × 10−1)). (c) The preliminary image. (d) The final image.
Figure 5a–e shows a visual comparison of different algorithms in the Set12 dataset. The KPN image effect is normal, the details are lost, and the denoising effect of the shadow area is not very ideal. CBDNet is better than KPN in detail recovery. It loses less detail and restores the shadow part well, but the overall image is not clear. CBDNet is better at processing low-noise images. While SRNet and WCNN performed well in terms of image sharpness and detail, there is still room for improvement. The algorithm proposed in this paper can recover image details more clearly and obtain a better visual effect. It still does not effectively remove all real noise, because the noise model does not perfectly represent real noise.
Figure 5. Denoising results of the images on the Set12 dataset by different methods. (a) KPN []. (b) CBDNet []. (c) SRNet []. (d) WCNN []. (e) Ours.
Table 3 shows the average PSNR in Set12 data sets. Table 4 shows the average SSIM of different algorithms in Set12 data sets. The average PSNR of the algorithm in the paper is higher than the other algorithms. Except for low-noise images, the denoising effect of CBDNet is slightly better than that of the method in this paper. The performance of SRNet and WCNN is similar to that of the proposed algorithm. However, in terms of SSIM, our algorithm is superior to other algorithms. It shows that the proposed algorithm is superior to the previous algorithms in denoising performance. The average SSIM index of the restored image is close to that of the original image, indicating that the quality of the restored image is close to that of the original noiseless image.
Table 3. The average PSNR(DB) results of different methods on three datasets.
Table 4. The average SSIM results of different methods on three datasets.
Figure 6 shows the average PSNR of our algorithm on the Set12 test set at the noise level of  σ 3 ( σ s  = 9.0 × 10−2 σ r  = 3 × 10−1). It shows the line graph of PSNR changing with the number of iterations. As the number of iterations increases, PSNR gradually flattens out.
Figure 6. The relationship between the number of iterations Epoch and PSNR.
Figure 7 shows the loss of our algorithm on the Set12 test set at the noise level of  σ 3 ( σ s  = 9.0 × 10−2 σ r  = 3 × 10−1). It shows the line graph of loss changing with the number of iterations. As the number of iterations increases, loss decreases and gradually flattens out.
Figure 7. The relationship between the number of iterations of Epoch and Loss.
To verify the denoising effect of our algorithm under high noise, we added a set of denoising experiments  σ ( σ s  = 9.0 × 10−1 σ r  = 3.0 × 1) for verification. Figure 8a,b shows a comparison of the denoising effect of the proposed method in the case of high noise.
Figure 8. Image denoising comparison under high noise  σ ( σ s  = 9.0 × 10−1 σ r  = 3.0 × 1). (a) High noise image. (b) Denoising image.
To verify the performance of the denoising algorithm in real scenes, we selected the α57 digital cameras launched by Sony to collect 30 images. To evaluate the performance of our method in terms of visual quality, we performed a subjective evaluation as a qualitative analysis experiment. The evaluation was carried out as a rating experiment where 40 users experienced in image processing viewed, in random order, a set of images. The users were asked to provide ratings for each image according to the following attributes: noise and detail. The final result of the ratings averaged over the observers is illustrated in Figure 9 and Figure 10.
Figure 9. Qualitative analysis result of noise, showing the average ratings from the conducted experiment, with error bars for the standard errors. (a) KPN []. (b) CBDNet []. (c) SRNet []. (d) WCNN []. (e) Ours.
Figure 10. Qualitative analysis result of detail, showing the average ratings from the conducted experiment, with error bars for the standard errors. (a) KPN []. (b) CBDNet []. (c) SRNet []. (d) WCNN []. (e) Ours.
Figure 11a–f shows two images we randomly selected. From the first picture, the details of the houses in the shaded part of the image restored by KPN are missing, and the overall effect of the image is not ideal. The local detail of the CBDNet algorithm is stronger than KPN, but the overall visual effect is a bit fuzzy, and the image quality is not good. WCNN and SRNet are better than the first two algorithms in terms of sharpness and image detail, but there is still room for improvement. In the marked areas in the image, we can see that other algorithms are rough in detail recovery in the shadow area, and the denoising effect is not ideal. Our algorithm performs well in this respect. The algorithm in this paper has the strongest local detail retention ability, and the restored image has fine details and a clear texture. In the second picture, KPN is missing some details in the recovered image. The text in the book is obscured, and the overall picture is poor in definition. In the image recovered by CBDNet, the text of the book can be seen, but it is not clear, and the overall image is a little fuzzy. The images recovered by WCNN and SRNet are good, but lower than the algorithm in this paper. In the marked areas in the images, we can see that the labels in the images recovered by other algorithms are fuzzy, and the denoising effect around the labels is not ideal, while the label recovered by our algorithm is clear and the denoising effect is ideal. The algorithm in this paper has a good performance in detail recovery and overall quality.
Figure 11. Image denoising comparison in real scenes. (a1,a2) The original image. (b1,b2) KPN []. (c1,c2) CBDNet []. (d1,d2) SRNet []. (e1,e2) WCNN []. (f1,f2) Ours.
Figure 12a,b shows the comparison before and after denoising of our algorithm. It can be intuitively seen from comparison figures that the noise in the images is greatly reduced after denoising by our algorithm. The image recovery effect is good, and the details are perfectly restored.
Figure 12. Image denoising comparison. (a1,a2) The original image. (b1,b2) Ours.
In general, our algorithm can reconstruct the detailed features more successfully. Although it does not achieve the best denoising effect in low noise images, the overall denoising effect is better and the restored images are clear, which is better than other algorithms.

5. Conclusions

The paper presents a deep signal-dependent denoising noise algorithm. There are two main stages to the work. Firstly, the noise parameters of the image are estimated by the noise parameters, and then the preliminary image is obtained according to the local statistical characteristics. Secondly, using the initial recovery map as input to the network along with the original map, take the initial restored image and the original image together as the input of the network, and the convolutional neural network can be used to reduce the error of the preliminary image and make the recovered image closer to the original image. We use the noise level of the noise image and the noise image together as the input of the convolutional neural network to obtain a wider range of noise levels than the single noise image as the input. It makes the denoising algorithm more widely used and has a better effect on detail recovery. We compared and analyzed denoising algorithms based on subjective vision. It was verified that the algorithm in this paper has a relatively evident inhibitory effect on signal-dependent noise, and the recovered images are of higher quality. This paper deals with grayscale images. For the future work, we hope to denoise color noise images. We believe that a combination of traditional image denoising and deep learning will become a great way to deal with noisy images.

Author Contributions

Conceptualization, L.Z.; methodology, L.Z. and J.W.; software, S.L.; validation, L.Z., S.L. and J.W.; formal analysis, S.L.; investigation, S.L.; resources, J.W.; data curation, L.Z.; writing—original draft preparation, S.L.; writing—review and editing, J.W.; visualization, J.W.; supervision, L.Z.; project administration, L.Z. and J.W.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Quzhou Science and Technology Plan Project, grant number 2022K108, and the Heilongjiang Provincial Natural Science Foundation of China, grant number YQ2022F014; and Basic Scientific Research Foundation Project of Provincial Colleges and Universities in Hei longjiang Province, grant number 2022KYYWF-FC05.

Data Availability Statement

As the algorithm in this paper relates to the project currently in progress, the project is private for some reasons, so the code is not convenient to be made public. If any reader wants to obtain the code of this article, you can send us an email. After receiving the email, we will determine whether to give the code to the reader after evaluation.

Acknowledgments

The authors acknowledge the Quzhou Science and Technology Plan Project (grant number 2022K108) and the Heilongjiang Provincial Natural Science Foundation of China (grant number YQ2022F014).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
Short NameFull Name
CNNsConvolutional Neural Networks
AWGNAdditive White Gaussian Noise
DnCNNFeed-forward Denoising Convolutional Neural Networks
NLFNoise Level Function
BDEBayesian Deep Ensemble
DRNEDeep Residual Noise Estimator
KPNKernel Prediction Network
MKPNMulti-Kernel Prediction Networks
REDRegularization by Denoising
WCNNWide CNN
MSR-blockMultiscale Residual Block
CBDNetConvolutional Blind Denoising Network
NERNetNoise Estimation module and Removal Network
SDNetSoftware Defined Network
SRNetSparse Representation-based Network
MUS-ADDMulti Scaling Aided Double Decker
ADMMAlternating Direction Method of Multipliers
MS-WGSCMulti-Scale Weighted Group Sparse Coding
AWTDAdaptive Weighted Tensor
CRCNNCascaded and Recursive Convolutional Neural Network
KCRDKronecker Component with Low-Rank Dictionary
GRDNGrouped Residual Dense Network
MSEMean Squared Error
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity

References

  1. Radlak, K.; Malinski, L.; Smolka, B. Deep learning based switching filter for impulsive noise removal in color images. Sensors 2020, 20, 2782. [Google Scholar] [CrossRef]
  2. Zhang, W. Digital image processing method for estimating leaf length and width tested using kiwifruit leaves (Actinidia chinensis Planch). PLoS ONE 2020, 15, e0235499. [Google Scholar] [CrossRef]
  3. Mildenhall, B.; Barron, J.T.; Chen, J.; Sharlet, D. Burst denoising with kernel prediction networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; Volume 13, pp. 2502–2510. [Google Scholar]
  4. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image de-noising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  5. Foi, A.; Trimeche, M.; Katkovnik, V.; Egiazarian, K. Practical Poissonian-Gaussian Noise Modeling and Fitting for Single-Image Raw-Data. IEEE Trans. Image Process. 2008, 17, 1737–1754. [Google Scholar] [CrossRef]
  6. Liu, X.; Tanaka, M.; Okutomi, M. Practical Signal-Dependent Noise Parameter Estimation From a Single Noisy Image. IEEE Trans. Image Process. 2014, 23, 4361–4371. [Google Scholar] [CrossRef]
  7. Liu, X.; Tanaka, M.; Okutomi, M. Signal dependent noise removal from a single image. IEEE Int. Conf. Image Process. IEEE 2014, 20, 2679–2683. [Google Scholar]
  8. Tan, H.; Xiao, H.; Lai, S.; Liu, Y.; Zhang, M. Pixelwise Estimation of Signal-Dependent Image Noise Using Deep Residual Learning. Comput. Intell. Neurosci. 2019, 30, 45–46. [Google Scholar] [CrossRef]
  9. Marinc, T.; Srinivasan, V.; Gul, S.; Hellge, C.; Samek, W. Multi-Kernel Prediction Networks for Denoising of Burst Images. In Proceedings of the 2019 IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; Volume 15, pp. 2404–2408. [Google Scholar] [CrossRef]
  10. Guo, B.; Song, K.; Dong, H.; Yan, Y.; Tu, Z.; Zhu, L. NERNet: Noise estimation and removal network for image denoising. J. Vis. Commun. Image Represent. 2020, 71, 102851. [Google Scholar] [CrossRef]
  11. Qi, G.; Hu, G.; Mazur, N. A novel multi-modality image simultaneous denoising and fusion method based on sparse representation. Computers 2021, 10, 129. [Google Scholar] [CrossRef]
  12. Xie, Z.; Liu, L.; Wang, C. Model-guided boosting for image denoising. Signal Process. 2022, 201, 108721. [Google Scholar] [CrossRef]
  13. Liu, G.; Dang, M.; Liu, J.; Xiang, R.; Tian, Y.; Luo, N. True wide convolutional neural network for image denoising. Inf. Sci. 2022, 610, 171–184. [Google Scholar] [CrossRef]
  14. Sheng, J.; Lv, G.; Wang, Z.; Feng, Q. SRNet: Sparse representation-based network for image denoising. Digit. Signal Process. 2022, 130, 103702. [Google Scholar] [CrossRef]
  15. Tojo, L.; Maik, V.; Devi, M. Image Denoising Using Multi Scaling Aided Double Decker Convolutional Neural Network. Optik 2022, 13, 170350. [Google Scholar] [CrossRef]
  16. Xu, S.; Zhang, J.; Wang, J.; Sun, K.; Zhang, C.; Liu, J.; Hu, J. A model-driven network for guided image denoising. Inf. Fusion 2022, 85, 60–71. [Google Scholar] [CrossRef]
  17. Ou, Y.; Swamy, M.; Luo, J.; Li, B. Single image denoising via multi-scale weighted group sparse coding. Signal Process. 2022, 200, 108650. [Google Scholar] [CrossRef]
  18. Zhang, L.; Liu, C. Robust Low-Rank Analysis with Adaptive Weighted Tensor for Image Denoising. Displays 2022, 73, 102200. [Google Scholar] [CrossRef]
  19. Li, X.; Meng, X.; Xiong, B. A fractional variational image denoising model with two-component regularization terms. Appl. Math. Comput. 2022, 427, 127178. [Google Scholar] [CrossRef]
  20. Khowaja, S.A.; Yahya, B.N.; Lee, S.L. Cascaded and Recursive ConvNets (CRCNN): An effective and flexible approach for image denoising. Signal Process. Image Commun. 2021, 99, 116420. [Google Scholar] [CrossRef]
  21. Zhang, L.; Liu, C. Kronecker component with robust low-rank dictionary for image denoising. Displays 2022, 74, 102194. [Google Scholar] [CrossRef]
  22. Phan, T.D.K. A weighted total variation based image denoising model using mean curvature. Optik 2020, 217, 164940. [Google Scholar] [CrossRef]
  23. Yao, H.; Zou, M.; Qin, C.; Zhang, X. Signal-Dependent Noise Estimation for a Real-Camera Model via Weight and Shape Constraints. IEEE Trans. Multimed. 2021, 24, 640–654. [Google Scholar] [CrossRef]
  24. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward Convolutional Blind Denoising of Real Photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; Volume 16, pp. 1712–1722. [Google Scholar] [CrossRef]
  25. Zhao, H.; Shao, W.; Bao, B.; Li, H. A simple and robust deep convolutional approach to blind image denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019; Volume 29. [Google Scholar]
  26. Kim, D.W.; Ryun Chung, J.; Jung, S.W. Grdn: Grouped residual dense network for real image denoising and gan-based real-world noise modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019; Volume 14. [Google Scholar]
  27. Akyüz, A.O.; Reinhard, E. Noise reduction in high dynamic range imaging. J. Vis. Commun. Image Represent. 2007, 18, 366–376. [Google Scholar] [CrossRef]
  28. Lee, J.-S. Digital Image Enhancement and Noise Filtering by Use of Local Statistics. IEEE Trans. Pattern Anal. Mach. Intell. 1980, 2, 165–168. [Google Scholar] [CrossRef]
  29. Plotz, T.; Roth, S. Benchmarking denoising algorithms with real photographs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 17, pp. 1586–1595. [Google Scholar]
  30. Available online: https://paperswithcode.com/paper/burst-denoising-with-kernel-prediction (accessed on 5 November 2022).
  31. Available online: https://paperswithcode.com/paper/toward-convolutional-blind-denoising-of-real (accessed on 8 November 2022).
  32. Available online: https://github.com/JiechaoSheng/SRNet (accessed on 9 November 2022).
  33. Available online: https://paperswithcode.com/paper/dc-wcnn-a-deep-cascade-of-wavelet-based (accessed on 11 November 2022).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.