You are currently viewing a new version of our website. To view the old version click .
Computers
  • Article
  • Open Access

24 May 2023

Image Denoising by Deep Convolution Based on Sparse Representation

,
,
and
School of Automation, University of Science and Technology Beijing, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.

Abstract

Noise filtering is a crucial task in digital image processing, performing the function of preprocessing. In this paper, we propose an algorithm that employs deep convolution and soft thresholding iterative algorithms to extract and learn the features of noisy images. The extracted features are acquired through prior and sparse representation theory for image reconstruction. Effective separation of the image and noise is achieved using an end-to-end network of dilated convolution and fully connected layers. Several experiments were performed on public images subject to various levels of Gaussian noise, in order to evaluate the effectiveness of the proposed approach. The results indicated that our algorithm achieved a high peak signal-to-noise ratio (PSNR) and significantly improved the visual effects of the images. Our study supports the effectiveness of our approach and substantiates its potential to be applied to a broad spectrum of image processing tasks.

1. Introduction

Digital image devices have a diverse range of applications, such as human recognition [1] and remote sensing [2]. During transmission and compression processes, noise is generated from latent observations that are unknown.
The focus of this paper is the classic image denoising problem. Specifically, the goal is to denoise the measurement of an ideal image X, which has been corrupted by the addition zero-mean white and homogeneous Gaussian noise, V, with a standard deviation of δ . The measured image Y can be represented mathematically by the following equation:
Y = X + V .
and our objective is to recover X from Y by knowing the parameter δ .
Image denoising is a crucial preprocessing step in low-level vision tasks, since noise corruption is often unavoidable during image acquisition. Demonstrably, this is a subtask within resolving an assortment of image reconstruction problems [3], such as super-resolution [4] and image restoration [5] problems, among others. VS Unni et al. [6] introduced a novel framework for image restoration that integrates linearized alternating direction method of multipliers (ADMM) and fast nonlocal denoising algorithms. The main objective of this method is to offer a computationally efficient solution for plug-and-play image restoration, which refers to utilizing pretrained denoisers within a general optimization framework for image restoration. Therefore, an effective image denoising algorithm could be applied to various image reconstruction problems. This explains why numerous researchers have proposed and improved denoising algorithms in the field of denoising in recent years.
Recently, there has been a plethora of sparse representation-based approaches suggested for image denoising. The K-SVD algorithm, introduced by Elad and Aharon in 2006 [7], was among the earliest sparse representation-based approaches. The K-SVD algorithm is a dictionary learning method that yields a dictionary of atoms that grasps the structural properties of natural images. By representing the noisy image as a linear combination of atoms learned from the dictionary, the K-SVD algorithm denoises the image by thresholding the coefficients. In 2007, another popular sparse representation-based approach named the BM3D algorithm was proposed by Dabov et al. [8]. The BM3D algorithm, which is a block-matching and 3D-filtering technique, makes use of the 3D redundancy existing in similar image patches. Initially, the BM3D algorithm groups identical patches together and employs 3D filtering on each group, to acquire an initial estimate of the denoised image. This estimate is then enhanced iteratively using a nonlocal means filter.
In recent years, various approaches based on sparse representation have been suggested for image denoising. Sun et al. [9] introduced a novel image denoising algorithm based on the sparse representation model that uses linear Bayesian maximum a posteriori (MAP) estimation. In order to solve inverse generalized image problems, an a prior probability distribution is constructed in the representation vector. This approach enables the construction of a linear Bayesian MAP estimator that adapts to the observations and acquires the most probable solution. In addition, Xu et al. [10] recommended a technique that learns a nonlocal similarity prior from images and utilizes this for denoising. A widely used approach to image denoising was suggested in [11], where the method uses the concept of linear Bayesian maximum a posteriori (MAP) estimation.
The main contribution of this paper is proposing an image denoising algorithm based on a combination of hollow convolution and the traditional KSVD algorithm. By combining the KSVD dictionary learning algorithm with deep learning, the accuracy and robustness of the denoising algorithm are improved. The experimental results show that this algorithm performed much better than BM3D, WNNM, and other algorithms in objective indicators and in subjective visual evaluation. Moreover, it can compete with DnCNN, a deep learning method with a good comprehensive performance, providing a new concept for image denoising.
The paper is structured as follows: Section 2 provides a literature review on image denoising, Section 3 outlines the proposed denoising framework, Section 4 details the simulation design and comparative analysis. The conclusion section summarizes the main contributions.

3. Proposed Framework

This paper proposes an end-to-end network model that utilizes deep dilated convolution and a multilayer perceptron as the primary architecture. Additionally, it employs a sparse representation technique to extract deep image features. The sparse representation technique in this model is the standard in the area of compressed sensing. The adoption of sparse representation to extract features in this model deals with deep networks that tend to over-smooth image textures during image denoising.
Our design takes into consideration dilated convolution networks to effectively capture the intrinsic features of images. Dilated convolution can provide a larger receptive field than regular convolution, thereby enhancing the ability and flexibility of using image features as the network depth increases. Moreover, convolution networks have definite advantages in terms of training regularization parameters and dictionaries, which include the use of rectified linear units and pooling operations. These advantages make it highly appropriate for GPU computing, accelerating the training process and improving the denoising performance. The convolutional layer is a filter-based image processor that can fit a filter with excellent denoising performance on noisy images. This is achievable through training on a vast amount of data, consequently diminishing edge artifacts. Figure 1 shows that the denoising framework comprises image decomposition, patch denoiser, and patch averaging components.
Figure 1. Denoising framework.
As referenced in [33], mapping between the noisy image and the denoised image is performed on local image patches. The primary concept of this paper is to broaden local image denoising to global image denoising. This involves designing a local image Bayesian estimator along with a global image Bayesian estimator.
Since the overcomplete dictionary used here is tailored to the characteristics of the noisy input image, it is appropriate to apply dictionary learning to smaller image blocks. An image X of dimensions M × N is segmented into several fully overlapping blocks of size n × n. This approach accelerates dictionary learning, while enhancing the data processing efficiency. Moreover, the sparse nature of each block permits its local prior to be easily transformed into a global prior consistent with the Bayesian framework.

3.1. Local Prior for Image Blocks

Consider an image patch x with dimensions n × n , which is reshaped by element as a column vector of length n 2 . To effectively separate the noise from the essential features of the image, we assume the sparsity of each image block. The sparse representation of image block x can be obtained as x = D α , where D R n 2 × m , m > > n 2 , is the DCT dictionary set tailored to x, and α R m is a nonzero sparse vector with sparsity k, i.e., x 0 = k . Given y, a noisy version of x corrupted by an additive zero-mean white Gaussian noise with standard deviation δ , a maximum a posteriori (MAP) estimator can be obtained using Bayesian theory by minimizing α 0 subject to y = D α , as shown in Equation (2).
α ^ = arg min α α 0 s . t y = D α .
The sparse coefficients α are obtained from the sparse representation of the image patch x through the dictionary D. To simplify the estimator and facilitate its solution, we replace the constraint term in Equation (2) with the difference between the noisy and the denoised image blocks. Therefore, the resulting equation is
α ^ = arg min α α 0 s . t D α y 2 2 n 2 δ 2 .
An accurate reconstruction of an ideal image can be obtained by utilizing sparse representation through an overcomplete dictionary D. The coefficient values obtained from the original image can be reconstructed exactly from the dictionary’s atoms. However, the presence of noise in the image decreases the quality of the reconstruction using the coefficients obtained from sparse representation. For controlling the sparse vector α and removing errors due to the standard deviation δ and noise block y, a regularization parameter λ is utilized. The estimator problem is simplified by transforming the constraint into a penalty term to solve α ^ , and subsequently transforming the equation into its Lagrangian form.
α ^ = arg min α λ α 0 + 1 2 D α y 2 2 .

3.2. Global Prior for Image Blocks

Sparse representation theory emphasizes the significance of sparsity in image denoising, as it automatically eliminates features that do not contain image information for noise removal. Additionally, utilizing the similarity of information between distinct image blocks in an image has been found to be advantageous for denoising, as affirmed in prior studies [37]. To highlight the synergistic nature between image blocks, it is assumed that
X = [ x 1 , x 2 , , x N ] R M × N
The natural generalization of the above MAP estimator is the replacement of (4) with
α i ^ , X ^ = arg min α i , X φ 2 X Y 2 2 + i λ i α i 0 + 1 2 D α i A i X 2 2
The notation A k represents the operator extracted from the image. As mentioned previously, obtaining the local prior of the image block enables a simple derivation of the global prior for the entire image X in the Bayesian framework.
The first term calculates the global logarithmic likelihood function that measures the similarity between the original image X and its noisy counterpart Y. φ represents the Lagrange multiplier, which serves as a constraint to limit the range of the first term in Equation (6). The second term is the image prior that makes sure that, in the constructed image, X, every patch x i = A i X of size n × n in every location has a sparse representation with bounded error.
Given all α i ^ , we can now fix these and turn to updating X. Returning to (6), we need to solve
X ^ = arg min X φ 2 X Y 2 2 + 1 2 i D α i A i X 2 2

3.3. Patch Denoiser

According to Equation (7), obtaining an optimal sparse coefficient is crucial for reconstructing a high-quality, denoised image. This section is devoted to solving the image’s optimal sparse coefficient. Three steps are involved in image denoising: sparse coding, parameter estimation, and image reconstruction. Figure 2 illustrates these steps.
Figure 2. Patch denoiser.

3.3.1. Sparse Coding

Equation (4) provides the objective equation for solving the sparse code. The complete dictionary D is pre-initialized as a DCT matrix based on the characteristics of the noise block. The sparsity rule operator, L1 parameterization, simplifies the problem by allowing for as many zeros in the model parameter vector as possible, while excluding features that have little impact on the predicted values. To fully reflect the sparsity, we use L1 parameterization instead of L0 parameterization to solve the sparse coding. The objective equation is as follows, where the complete dictionary D has been pre-initialized as a DCT matrix based on the characteristics of the noise block:
α ^ = arg min α 1 2 D α y 2 2 + λ α 1
Equation (8) is a convex optimization problem, which is typically addressed using gradient-based methods. Employing this approach allows primarily focusing on the product of matrix D and vector α , reducing the associated computational burden. The algorithm used in this research is characterized by low computational complexity and a simple structure. Among different gradient-based methods, the iterative shrinkage threshold algorithm (ISTA) [47] holds a prominent position. The ISTA updates α by performing a soft thresholding operation in every iteration. This approach ensures that the sparse solution converges to the global optimum.
α ^ i + 1 = Soft λ t α ^ i 2 t D T D α ^ i y ; α ^ 0 = 0
Soft τ ( ε ) = sign ε i ε i τ + .
We define t as the reciprocal of the soft threshold iteration step and S o f t λ t as the corresponding soft threshold iteration function.

3.3.2. λ Estimation

The sparse coding α is obtained by addressing a regularization parameter λ k for each noise patch y k , to regulate the sparsity. The regularization parameter is fitted based on the noise level and signal characteristics. The optimal sparsity coefficient α ^ is evaluated and utilized with the sparse reconstruction theory formula x ^ = D α ^ to reconstruct the denoised image patch x ^ with high precision. This accomplishes the mapping of the noisy image to the denoised image. In this study, a data-driven approach to realize this mapping is proposed, through learning a regression function.
x k = f τ y k
This paper proposes a network that performs various tasks, such as estimating λ k , solving sparse coding, updating parameters, and mapping noisy and denoised images. The network, as illustrated in Figure 3, estimates λ by utilizing seven layers of hybrid dilated convolution paired with batch normalization (BN). The network also uses four layers of multilayer perceptron, where every hidden layer is connected to a Relu activation function. The pooling operation in some cases can extract valuable features, but it can also result in the loss of significant image details, which is not desirable for the denoising task at hand. Therefore, a pooling operation is not employed, since the model’s input and output size remain unchanged.
Figure 3. λ estimation network.
To conclude, the proposed network f is a parameterized function of τ (the parameters of the network computing λ ), t (the step-size in the ISTA algorithm), D (the dictionary), and w (the weights for the patch-averaging).

3.3.3. Patch Reconstruction

This stage reconstructs the cleaned version x ^ of the patch y using D and the sparse code α ^ , which is given by x ^ = D α ^ . D and α ^ are both updated in the aforementioned network.

3.4. Patch Averaging

Now we can discuss the complete system architecture. First, we will decompose the input image into fully overlapping patches, and then process each damaged patch through the aforementioned patch denoising stage. Finally, we will average these damaged patches, to reconstruct the image. In the final stage, we realize the learned weighted combination of patches. The weight of each patch is represented as w R n × n , so the retrieval of the reconstructed image is obtained as
X ^ = i A i T w x ^ i i A i T w
Obtaining a denoised image X is feasible, with ⊙ representing the Schur product and A i signifying the ith operator extracted from the image.

3.5. Loss Function

Considering a noisy image Y, we use a denoising network to produce a denoised image X given that X = F ( Y ) , where F represents the end-to-end denoising network. We derive the loss function from this specific mapping:
L = i X i F Y i 2 2
The network is trained using data to minimize the loss function. Here, X i i and Y i i denote the training images and their noisy counterparts, respectively. The noise added is zero-mean and white Gaussian iid.

4. Simulation Results

Here, we present the simulation experiments of our proposed model, aiming to demonstrate the superiority of the fusion of traditional sparse representation theory and deep learning over conventional state-of-the-art denoising algorithms such as BM3D and WNNM. Moreover, our model was expected to outperform DnCNN in terms of objective metrics and image detail preservation.

4.1. Simulation Experiments

Training settings: We trained our model using the BSD400 dataset as the training dataset, comprising 400 grayscale images of natural and face scenes with dimensions 481 × 321 and 321 × 481 pixels. The images during training were randomly cropped to 200 × 200 patches with a stride of 40. The input noise during training was independent and identically distributed Gaussian noise, with a mean of zero, and the noise level δ was specified. We trained several models for each noise level: δ = 15, 25, and 50. Our model used the mean squared error (MSE) as the loss function, and the optimization algorithm used to minimize this loss was Adam. Our model’s initial learning rate was set to 1 × 10 3 and decayed to 1 × 10 5 over 9 epochs. The learning rate was set to 1 × 10 3 for the first three epochs, 1 × 10 4 for the next three epochs (i.e., the fourth to sixth epochs), and 1 × 10 5 for the last three epochs (i.e., seventh to ninth epochs). The DCT matrix was overcomplete and initialized using the K-singular value decomposition (KSVD) [7] method. The parameter t was set using the squared spectral norm of the DCT matrix. The Kaiming uniform method was used to initialize the other network parameters, including Conv, BN, and MLP.
Validation settings: The trained models were evaluated using two datasets, BSD68 and Set12. The objective evaluation metric used in this study was the peak signal-to-noise ratio, which was calculated based on the following formula:
PSNR = 10 · log 10 MAX I 2 MSE
where M A X I represents the maximum possible pixel value (usually 255 for 8-bit grayscale images), and M S E is the mean squared error.
In this paper, we utilized PyTorch 1.2.0 and Python 3.7 to train and test our proposed model. We conducted all experiments in the Windows 10 operating system with a 3.60 GHz clocked AMD Ryzen 7 3700X CPU, 32 GB of RAM, and an Nvidia GeForce RTX 2080 Ti GPU. To speed up the GPU training process, we used Nvidia CUDA 10.0 and cuDNN 7.4. It took approximately 20 h to train the grayscale Gaussian denoising model.

4.2. Analysis of Simulation Results

Evaluating the effectiveness of an algorithm involves considering both objective and subjective criteria. The objective criterion usually measures the average peak signal-to-noise ratio (PSNR) of the algorithm across all images in a given dataset, where higher values suggest a better performance. In contrast, an improved subjective evaluation results from preserving more image detail and minimizing edge artifacts.
Table 1 shows that average PSNR (dB) of different methods on BSD68 with different noise levels of 15, 25, and 50. The red and blue colors mark the highest and second-best denoising performance in Table 1, respectively. Our proposed method outperformed BM3D [8], KSVD [7], WNNM [18], TNRD [21], DnCNN [35], and MLP [33] at various noise levels. Our method demonstrated a 0.02 dB improvement over DnCNN, a 0.28 dB improvement over TNRD, a 0.67 dB improvement over BM3D, and a 0.37 dB improvement over WNNM at a noise level of 15. Similar results were observed for noise levels of 25 and 50, demonstrating that the combination of sparse representation theory and deep learning is a powerful approach, which leverages the accurate prior knowledge provided by sparse representation to complement deep learning techniques. These results confirm the effectiveness of our proposed method and validate the feasibility of our approach.
Table 1. Average PSNR (dB) of the different methods on BSD68 with different noise levels of 15, 25, and 50.
Table 2 shows the average SSIM (dB) of the different methods on BSD68 with different noise levels of 15, 25, and 50. SSIM (structural similarity index) is a widely used image quality assessment metric that measures the structural similarity between two images. It takes into account the luminance, contrast, and structural information of images and produces a score between 0 and 1, where 1 indicates that the two images are identical. It can be observed that the SSIM and PSNR rankings of the denoising algorithms were consistent. Therefore, in our evaluation experiments, we used PSNR as the objective evaluation metric and image detail preservation as the subjective evaluation metric.
Table 2. Average SSIM (dB) of the different methods on BSD68 with different noise levels of 15, 25, and 50.
Table 3 depicts the average peak signal-to-noise ratio (PSNR) (dB) of the various image denoising methods on Set12 with varying noise levels of 15, 25, and 50. The highest and second-best performances in Table 2 are marked with red and blue lines, correspondingly. Our proposed method exhibited a superior performance to the other methods tested, as revealed in Table 2. Notably, our method surpassed the DnCNN and WNNM techniques, with a PSNR improvement of 0.03 dB and 0.19 dB, respectively, for a noise level of 15. Remarkably, we observed that the performance of WNNM on Set12 was nearly comparable to that of DnCNN and our proposed method, whereas it significantly outperformed KSVD. We conjecture that this could be attributed to the fact that both WNNM and BM3D exploit nonlocal self-similarity, which can facilitate the acquisition of accurate structural information and lead to better noise reduction by discovering similar pixel blocks across the entire image. This characteristic is also leveraged in our algorithm.
Table 3. Average PSNR (dB) of the different methods on Set12 with different noise levels of 15, 25, and 50.
Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 present the visual performance and corresponding peak signal-to-noise ratio (PSNR) values in decibels (dB) of the different methods on the five images in the BSD68 dataset at a noise level of 25. The red box in the Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8 indicates the enlargement of a detail. Overall, our proposed method and the DnCNN method achieved the leading denoising quality among the five methods compared, in terms of both noise reduction and detail preservation. Specifically, as shown in Figure 4, the detail enlargement of the bird’s eye reveals that only our method and the MLP method fully preserved the details of the eyeball, and our method achieved the clearest result. Additionally, Figure 6 shows that the detail enlargement of the animal nose indicates that our method preserved the texture details of the nose most clearly and closest to the original image. Furthermore, Figure 8 indicates that the detail enlargement of the penguin’s abdomen reveals that our method preserved the texture details of the abdomen the best, followed by the DnCNN method. Combining the results in Table 1 and Table 2, and the performance shown in Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8, we believe that our proposed method achieved the predetermined goal. Our method performed significantly better than traditional methods such as the BM3D and WNNM methods on the public datasets, while being able to compete with DnCNN methods in terms of objective metrics and image quality.
Figure 4. Denoising results of the different methods on an image from BSD68 with δ = 25. Original image, WNNM/30.42 dB, DnCNN/30.75 dB, BM3D/30.33 dB, MLP/30.47 dB, Ours/30.78 dB.
Figure 5. Denoising results of the different methods on an image from BSD68 with δ = 25. Original image, WNNM/29.76 dB, DnCNN/30.19 dB, BM3D/29.53 dB, MLP/29.72 dB, Ours/30.17 dB.
Figure 6. Denoising results of the different methods on an image from BSD68 with δ = 25.Original image, WNNM/27.30 dB, DnCNN/27.68 dB, BM3D/27.11 dB, MLP/27.51 dB, Ours/27.72 dB.
Figure 7. Denoising results of the different methods on an image from BSD68 with δ = 25. Original image, WNNM/37.20 dB, DnCNN/38.41 dB, BM3D/36.94 dB, MLP/37.53 dB, Ours/38.36 dB.
Figure 8. Denoising results of the different methods on an image from BSD68 with δ = 25. Original image, WNNM/32.56 dB, DnCNN/33.05 dB, BM3D/32.41 dB, MLP/29.72 dB, Ours/33.09 dB.

5. Conclusions

In this paper, we propose a deep convolutional denoiser based on sparse representation that combines the strengths of sparse representation and deep learning. Sparse representation provides accurate prior knowledge, while deep learning networks can use large amounts of data to automatically train parameters, overcoming the limitations of traditional denoising methods. The ideal combination of sparse coding and deep learning networks is reflected in the simulation results, where our proposed method outperformed both the traditional denoising algorithms, such as BM3D and WNNM, and DnCNN in terms of objective metrics and detail preservation. This demonstrates that our proposed approach is effective and highly competitive.

Author Contributions

Conceptualization, Z.X. and S.B.; methodology, X.H.; software, X.H.; validation, X.H.; formal analysis, X.H.; investigation, X.H.; resources, L.Z.; data curation, S.B.; writing—original draft preparation, S.B.; writing—review and editing, S.B.; visualization, S.B.; supervision, S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

In this section, the datasets generated and analyzed during the current study are available at https://paperswithcode.com/dataset/bsds500 (accessed on 16 April 2023). The validation set is available at https://research.repository.duke.edu/concern/datasets/4b29b657c (accessed on 16 April 2023).

Conflicts of Interest

The authors do not have any relevant financial or non-financial interests to disclose. The authors did not receive support from any organization for the submitted work. All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or nonfinancial interest in the subject matter or materials discussed in this manuscript.

Abbreviations

The following abbreviations are used in this manuscript:
AWGNadditive white Gaussian noise;
ISTAiterative shrinkage thresholding algorithm
PSNRsignal-to-noise ratio
BM3Dblock-matching and 3D filtering
MAPmaximum a posteriori
WNNMweighted nuclear norm minimization
TNRDlinear reaction diffusion
BNbatch normalization
DCTdiscrete cosine transform

References

  1. Lei, Y.; Yuan, W.; Wang, H.; Wenhu, Y.; Bo, W. A skin segmentation algorithm based on stacked autoencoders. IEEE Trans. Multimed. 2016, 19, 740–749. [Google Scholar] [CrossRef]
  2. Fonseca, L.M.G.; Namikawa, L.M.; Castejon, E.F. Digital image processing in remote sensing. In Proceedings of the 2009 Tutorials of the XXII Brazilian Symposium on Computer Graphics and Image Processing, Rio de Janeiro, Brazil, 11–14 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 59–71. [Google Scholar]
  3. Wang, G.; Ye, J.C.; De Man, B. Deep learning for tomographic image reconstruction. Nat. Mach. Intell. 2020, 2, 737–748. [Google Scholar] [CrossRef]
  4. Schermelleh, L.; Ferrand, A.; Huser, T.; Eggeling, C.; Sauer, M.; Biehlmaier, O.; Drummen, G.P. Super-resolution microscopy demystified. Nat. Cell Biol. 2019, 21, 72–84. [Google Scholar] [CrossRef] [PubMed]
  5. Ali, A.M.; Benjdira, B.; Koubaa, A.; El-Shafai, W.; Khan, Z.; Boulila, W. Vision Transformers in Image Restoration: A Survey. Sensors 2023, 23, 2385. [Google Scholar] [CrossRef] [PubMed]
  6. Unni, V.; Ghosh, S.; Chaudhury, K.N. Linearized ADMM and fast nonlocal denoising for efficient plug-and-play restoration. In Proceedings of the 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Anaheim, CA, USA, 26–29 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 11–15. [Google Scholar]
  7. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  8. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar]
  9. Sun, D.; Gao, Q.; Lu, Y.; Huang, Z.; Li, T. A novel image denoising algorithm using linear Bayesian MAP estimation based on sparse representation. Signal Process. 2014, 100, 132–145. [Google Scholar] [CrossRef]
  10. Xu, H.; Jia, X.; Cheng, L.; Huang, H. Affine non-local Bayesian image denoising algorithm. Vis. Comput. 2023, 39, 99–118. [Google Scholar] [CrossRef]
  11. Glynn, C.; Tokdar, S.T.; Howard, B.; Banks, D.L. Bayesian analysis of dynamic linear topic models. Bayesian Anal. 2019, 14, 53–80. [Google Scholar] [CrossRef]
  12. Kollem, S.; Reddy, K.R.L.; Rao, D.S. A review of image denoising and segmentation methods based on medical images. Int. J. Mach. Learn. Comput. 2019, 9, 288–295. [Google Scholar] [CrossRef]
  13. de Ridder, D.; Duin, R.P.; Verbeek, P.W.; Van Vliet, L. The applicability of neural networks to non-linear image processing. Pattern Anal. Appl. 1999, 2, 111–128. [Google Scholar] [CrossRef]
  14. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  15. Osher, S.; Burger, M.; Goldfarb, D.; Xu, J.; Yin, W. An iterative regularization method for total variation-based image restoration. Multiscale Model. Simul. 2005, 4, 460–489. [Google Scholar] [CrossRef]
  16. Ren, D.; Zuo, W.; Zhang, D.; Zhang, L.; Yang, M.H. Simultaneous fidelity and regularization learning for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 284–299. [Google Scholar] [CrossRef]
  17. Cross, G.R.; Jain, A.K. Markov random field texture models. IEEE Trans. Pattern Anal. Mach. Intell. 1983, PAMI-5, 25–39. [Google Scholar] [CrossRef]
  18. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  19. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, St. Louis, MO, USA, 4–7 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 2272–2279. [Google Scholar]
  20. Schmidt, U.; Roth, S. Shrinkage fields for effective image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2774–2781. [Google Scholar]
  21. Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1256–1272. [Google Scholar] [CrossRef]
  22. Zuo, W.; Zhang, L.; Song, C.; Zhang, D.; Gao, H. Gradient histogram estimation and preservation for texture enhanced image denoising. IEEE Trans. Image Process. 2014, 23, 2459–2472. [Google Scholar]
  23. Khmag, A.; Ramli, A.R.; Kamarudin, N. Clustering-based natural image denoising using dictionary learning approach in wavelet domain. Soft Comput. 2019, 23, 8013–8027. [Google Scholar] [CrossRef]
  24. Tallapragada, V.; Manga, N.A.; Kumar, G.; Naresh, M.V. Mixed image denoising using weighted coding and non-local similarity. SN Appl. Sci. 2020, 2, 997. [Google Scholar] [CrossRef]
  25. Mahdaoui, A.E.; Ouahabi, A.; Moulay, M.S. Image denoising using a compressive sensing approach based on regularization constraints. Sensors 2022, 22, 2199. [Google Scholar] [CrossRef]
  26. Fukushima, K.; Miyake, S. Neocognitron: A self-organizing neural network model for a mechanism of visual pattern recognition. In Competition and Cooperation in Neural Nets; Springer: Berlin/Heidelberg, Germany, 1982; pp. 267–285. [Google Scholar]
  27. Zhou, Y.; Chellappa, R.; Jenkins, B. A novel approach to image restoration based on a neural network. In Proceedings of the International Conference on Neural Networks, San Diego, CA, USA, 21 June 1987. [Google Scholar]
  28. Tamura, S. An analysis of a noise reduction neural network. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Glasgow, UK, 23–26 May 1989; IEEE: Piscataway, NJ, USA, 1989; pp. 2001–2004. [Google Scholar]
  29. Liang, J.; Liu, R. Stacked denoising autoencoder and dropout together to prevent overfitting in deep neural network. In Proceedings of the 2015 8th International Congress on Image and Signal Processing (CISP), Shenyang, China, 14–16 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 697–701. [Google Scholar]
  30. Zhang, Z.; Wang, L.; Kai, A.; Yamada, T.; Li, W.; Iwahashi, M. Deep neural network-based bottleneck feature and denoising autoencoder-based dereverberation for distant-talking speaker identification. EURASIP J. Audio, Speech, Music. Process. 2015, 2015, 1–13. [Google Scholar] [CrossRef]
  31. Yuan, D.; Fan, N.; He, Z. Learning target-focusing convolutional regression model for visual object tracking. Knowl.-Based Syst. 2020, 194, 105526. [Google Scholar] [CrossRef]
  32. Tian, C.; Xu, Y.; Zuo, W.; Zhang, B.; Fei, L.; Lin, C.W. Coarse-to-fine CNN for image super-resolution. IEEE Trans. Multimed. 2020, 23, 1489–1502. [Google Scholar] [CrossRef]
  33. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image denoising: Can plain neural networks compete with BM3D? In Proceedings of the 2012 IEEE conference on computer vision and pattern recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2392–2399. [Google Scholar]
  34. Mao, X.; Shen, C.; Yang, Y.B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. Adv. Neural Inf. Process. Syst. 2016, 29, 1–9. [Google Scholar]
  35. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning (PMLR), Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  37. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Phoenix, AZ, USA, 25–28 September 2016; pp. 770–778. [Google Scholar]
  39. Ghose, S.; Singh, N.; Singh, P. Image denoising using deep learning: Convolutional neural network. In Proceedings of the 2020 10th International Conference on Cloud Computing, Data Science and Engineering (Confluence), Noida, India, 29–31 January 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 511–517. [Google Scholar]
  40. Gai, S.; Bao, Z. New image denoising algorithm via improved deep convolutional neural network with perceptive loss. Expert Syst. Appl. 2019, 138, 112815. [Google Scholar] [CrossRef]
  41. Zhang, K.; Zuo, W.; Zhang, L. Deep plug-and-play super-resolution for arbitrary blur kernels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 1671–1681. [Google Scholar]
  42. Cheng, S.; Wang, Y.; Huang, H.; Liu, D.; Fan, H.; Liu, S. Nbnet: Noise basis learning for image denoising with subspace projection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4896–4906. [Google Scholar]
  43. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
  44. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 1712–1722. [Google Scholar]
  45. Gurrola-Ramos, J.; Dalmau, O.; Alarcón, T.E. A residual dense U-Net neural network for image denoising. IEEE Access 2021, 9, 31742–31754. [Google Scholar] [CrossRef]
  46. Jia, F.; Wong, W.H.; Zeng, T. DDUNet: Dense dense U-net with applications in image denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 354–364. [Google Scholar]
  47. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.