Next Article in Journal
Study on Aberration Correction of Adaptive Optics Based on Convolutional Neural Network
Next Article in Special Issue
Aerial Projection 3D Display Based on Integral Imaging
Previous Article in Journal
Design and Simulation of Terahertz Perfect Absorber with Tunable Absorption Characteristic Using Fractal-Shaped Graphene Layers
Previous Article in Special Issue
Resolution Enhancement in Coherent Diffraction Imaging Using High Dynamic Range Image
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Restoration Based on End-to-End Unrolled Network

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
Key Laboratory of Optical System Advanced Manufacturing Technology, Chinese Academy of Sciences, Changchun 130033, China
3
State Key Laboratory of Modern Optical Instrumentation, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
Photonics 2021, 8(9), 376; https://doi.org/10.3390/photonics8090376
Submission received: 22 June 2021 / Revised: 30 August 2021 / Accepted: 31 August 2021 / Published: 8 September 2021
(This article belongs to the Special Issue Smart Pixels and Imaging)

Abstract

:
Recent studies on image restoration (IR) methods under unrolled optimization frameworks have shown that deep convolutional neural networks (DCNNs) can be implicitly used as priors to solve inverse problems. Due to the ill-conditioned nature of the inverse problem, the selection of prior knowledge is crucial for the process of IR. However, the existing methods use a fixed DCNN in each iteration, and so they cannot fully adapt to the image characteristics at each iteration stage. In this paper, we combine deep learning with traditional optimization and propose an end-to-end unrolled network based on deep priors. The entire network contains several iterations, and each iteration is composed of analytic solution updates and a small multiscale deep denoiser network. In particular, we use different denoiser networks at different stages to improve adaptability. Compared with a fixed DCNN, it greatly reduces the number of computations when the total parameters are equal and the number of iterations is the same, but the gains from a practical runtime are not as significant as indicated in the FLOP count. The experimental results of our method of three IR tasks, including denoising, deblurring, and lensless imaging, demonstrate that our proposed method achieves state-of-the-art performances in terms of both visual effects and quantitative evaluations.

1. Introduction

Image restoration (IR) is a classical topic in the field of low-level image processing. Digital images are always degraded during the acquisition process, with issues such as electronic noise caused by the thermal vibration of atoms and blur caused by camera shake [1,2]. Therefore, image restoration is of great significance and is widely used in a variety of applications, e.g., smartphone imaging, medical imaging, and remote sensing. The purpose of image restoration is to recover an unknown latent image from a corrupted observation. In general, IR is an ill-posed inverse problem. The mathematical degradation model can be written as y = Ax + n, where y and x are the degraded measurement and the clean image, respectively. n denotes the additive noise, which is generally assumed to be additive white Gaussian noise (AWGN), and A denotes the system degradation matrix. Although IR problems have been extensively studied, they are still full of challenges due to the large amount of natural image contents [3] and the diversity of degradations.
Over the past few decades, many methods have been proposed to tackle IR problems, including denoising [4,5,6,7,8,9,10], deblurring [11,12,13,14,15], image super-resolution [16,17,18,19,20], and lensless imaging [21,22,23,24]. Recently, the rapid development of deep learning technology has injected new vitality into IR research. Many works based on deep convolutional neural networks (DCNNs) have achieved excellent results [10,23,25]. From the point of view of linear algebra, the reason leads to ill condition of IR is that the null space of A is nonzero. Because of this, the prior knowledge of the image is crucial, and we need to select a good estimation of the latent image from the solution space according to this prior knowledge. A common approach is to establish a cost function by maximizing the posterior probability P(x|y):
x ^ = a r g   max x   l o g   P ( y | x ) + l o g   P x
where l o g   P x represents the prior information of the latent image, and l o g   P ( y | x ) denotes the log-likelihood of the measurement. l o g   P ( y | x ) is derived from statistical model of noise, such as 2   n o r m to Gaussian noise and 1   n o r m to Laplacian noise. Under the AWGN model, the cost function can be reformulated as:
x ^ = a r g   min x 1 2 || y A x || 2 2 + λ · ϕ x
where ϕ x = l o g   P x is a regularization term. There are two paths for obtaining the solution of Equation (2). One involves model-based methods [26,27,28], and the other involves learning-based methods [29,30,31,32]. The former methods reduce the value of the cost function gradually through the optimization principles, whereas the latter methods mainly rely on DCNNs and datasets. These two types of methods are described in more detail in the following four paragraphs.
Among model-based methods, a large number of models based on various priors have been proposed, including the frequently used total variation (TV) prior [33]; sparse representation prior [34] and dictionary learning [7,35,36]; nonlocal means prior [37] and nonlocal self-similarity (NLSS) [6,38]; the low rank approximation prior [39,40]; and the Markov random field (MRF) [41,42]. The characteristics of various priors are as follows.
The well-known TV prior works well with images with simple textures, but it can lead to distinct blurring in complex areas with rich details. The sparse prior model can represent local image patches as a few atoms in some domains, such as the DCT basis and the discrete wavelet transform basis [43]. Compared with the analytical dictionaries, learned dictionaries have a stronger adaptive ability to represent image patches, and they can deal with various tasks more flexibly [3]. In the past two decades, sparse models have made outstanding contributions to IR, and a large number of IR algorithms based on them have been proposed [4,6,7,16,26,34]. Using the redundant information in an image, the nonlocal means can eliminate Gaussian noise well [44]. A more efficient and robust solution is to apply the NLSS to the sparse prior model [26]. Another powerful prior is the low-rank approximation because a matrix with many nonlocal similar patches is essentially of low rank [39]. The soft threshold function can be used to solve this problem easily and quickly [45]. WNNM [40] improves the flexibility of nuclear norm and achieves good results in terms of both visual effects and quantitative evaluations. Combined with the MRF model, the Bayesian optimization framework is applied to low-level vision [41,46,47]. Although the MRF can learn a generic prior that can represent the statistics of natural scenes, its complexity is high. It is also difficult to carry out a physical interpretation. In summary, model-based methods can deal with all kinds of visual problems flexibly, but they usually incur high time and computational costs, and their effects are not as good as those of the popular DCNNs.
With the rapid development of deep learning, DCNNs have blossomed in the field of low-level vision. Many learning-based methods [8,9,10,29,30,48,49] have been applied to denoising tasks. Burger et al. [29] proposed using a plain multilayer perceptron for denoising. This work showed the great potential of neural networks because a simple perceptron was shown to be able to achieve the effect of a traditional well-known BM3D denoiser [4]. With batch normalization, the DnCNN [30] was established to provide an end-to-end residual learning network for predicting residuals to indirectly eliminate noise. CBDNet [8] separated noise estimation from nonblind denoising and used two sub-networks to complete these two functions separately. RIDNet [9] exploited channel dependencies by using feature attention and built a blind real image denoising network under a modular architecture. SADNet [10] introduced the deformable convolution to implement spatial adaptive denoising, which can achieve reconstruction with a high signal-to-noise ratio while effectively maintaining spatial texture and edges of the image. Zhang et al. [49] provides a novel and efficient RDN that achieves superiority over comparative methods for several image restoration tasks.
In addition to noise removal, DCNNs are also widely used in the fields of super-resolution and lensless imaging. The SRCNN [18] mapped low-resolution patches to high-resolution patches by three layers of convolutions. After that, many deep networks were proposed for super-resolution, such as the ESPCN [19], DRRN [20], VDSR [32], and SRGAN [25]. DCNNs have also performed well in the field of lensless imaging. Nguyen et al. [22] used a DCNN to restore lensless images and protected privacy. Khan et al. [23] first carried out model fitting and then used a DCNN to improve image quality.
Although learning-based DCNNs can quickly complete high-quality image reconstruction on GPUs after training, they usually lose the flexibility inherent in model-based methods. Additionally, the improvement in reconstruction quality is only due to the strong fitting ability of pure DCNN. Therefore, hybrid IR methods under unrolled framework were proposed [50,51]. Section 2.2 describes more details about IR methods under unrolled framework. These kinds of methods combine traditional methods with DCNNs, and the advantages of both can be exploited. When solving the inverse problem, they can incorporate the physical models of systems into the networks. First, this structure can make full use of the prior knowledge of the systems, such as the observation matrix in compressed sensing and the point spread function in deconvolution. In addition, the main function of the network is to learn a prior, rather than the whole inverse operation. Thus, unrolled structure has lower requirements on the network. Additionally, the functions are easier to achieve. Furthermore, they can improve the reconstruction quality compared with pure DCNNs. In spite of their wide applications in low-level vision tasks, there is still room for improvement in terms of optimization and the networks. For example, the existing methods usually use a fixed DCNN in each iteration, and so they cannot fully adapt to the image characteristics at each iteration stage. The gradient descent is slower for the convex problem.
In this paper, we propose a deep denoiser-based unrolled network that combines DCNNs with optimization to exploit the advantages of both. The entire end-to-end network can be unfolded into several analytic solution blocks and the subsequent small deep denoiser networks. All the parameters are learned through training. On the one hand, we solve the convex problem in the form of an analytic solution that is faster than gradient descent which usually requires multiple iterations because the solution is not accurate. On the other hand, each small deep denoiser network adopts a structure with an encoder and a decoder to capture multiscale information from the image. The small deep denoiser networks are also different at different stages so that they can better adapt to the image characteristics at each stage. Compared with a fixed DCNN, it greatly reduces the number of computations when the total parameters are the same and the number of iterations is the same. The experimental results of several IR tasks, including denoising, deblurring, and lensless imaging, demonstrate that our approach is effective and computationally efficient. The visual effects and the objective evaluations indicate that our network achieves excellent performances in high-quality image reconstruction.
The remainder of this paper is organized as follows. Section 2 reviews related works. Section 3 introduces our proposed method. Section 4 shows the numerical results of several IR tasks, and Section 5 concludes this paper.

2. Related Work

The unrolled network we proposed is mainly derived from two aspects: deep learning and denoiser-based IR methods under unrolled optimization. In this section, we briefly review these two aspects.

2.1. Deep Learning

With the rapid development of computing power, deep learning technology has led to many breakthroughs in the field of vision, including low-level denoising [10,30,48], deblurring [15,31], super-resolution [32,52] and high-level recognition [53], segmentation [54]. The image textures are richer under the generative adversarial networks (GANs) [55]. The perceptual loss takes advantage of high-level abstract features that enhance the details of super-resolution images [56]. Additionally, the training method for networks has been developed, such as batch normalization [57], gradient clipping [58], and Xavier initialization [59].

2.2. IR Methods under Unrolled Optimization

By decoupling the deconvolution and denoising, the original complex regularization term can be transferred. Many denoiser based IR methods have been proposed [11,50,60,61,62] that can integrate the strengths of model-based methods and DCNNs. Under the framework of half-quadratic splitting (HQS), a new auxiliary variable z is introduced to Equation (2). Therefore, the IR problem can be written in the following form:
φ μ x , z = 1 2   || y A x || 2 2 + μ 2   || z x || 2 2 + λ · ϕ z
where μ is the penalty parameter. The above equation can be solved alternatively by:
x ^ = a r g   min x || y A x || 2 2 + μ   || x z ^ || 2 2 z ^ = a r g   min z μ 2   || x ^ z || 2 2 + λ · ϕ z
As we can see, the former equation is a convex equation while the latter is a proximal operator with special regularization parameters. In practice, we usually treat it as a denoising problem that can avoid the explicit expression of priors and there are many successful solutions at present. The decoupling can also be achieved through the ADMM [63]; the principle is similar and will not be repeated here. IR methods under unrolled optimization frameworks can be divided into two categories: deep unfolding networks and plug-and-play. The former type is an overall end-to-end network, while the latter is not.
Plug-and-play methods can be flexibly applied to various tasks [64,65] by using one well-trained denoiser. In [11], the well-known BM3D denoiser was used for deblurring based on the generalized Nash equilibrium. The CBM3D denoiser was used for single image super-resolution (SISR) [17]. In [66], Brifman et al. realized SISR by using the NCSR denoiser [26] which combines the traditional sparse prior with the NLSS. Additionally, the results were better than those of original NCSR. The TV prior and BM3D prior were used for Fourier ptychographic microscopy [67]. In [68], the denoising-based IR method with the ADMM was used for electron microscope imaging. Additionally, the state-of-the-art DCNN denoiser prior was used in IR tasks [69]. Zhang et al. [61] trained 25 CNN denoisers at different noise levels. There are some theoretical analyses on this topic. Sreehari et al. [68] analyzed the convergence of the plug-and-play approach when the denoiser is a symmetric smoothing filter. Chan’s algorithm with a bounded denoiser [70] was proven to be convergent. Ryu’s work theoretically established the convergence of the PnP-ADMM algorithm when the denoiser has a certain Lipschitz condition [71]. Although the plug-and-play technique has achieved start-of-the-art results, it usually requires multiple iterations. The proposed SISR solver in [66] iterates 35 times, and the IRCNN [61] takes 30 iterations to deblur.
In response to this problem, the end-to-end deep unfolding network consisting of only a few iterations was proposed for IR [51,62]. Zhang et al. [72] proposed a deep network for compressed sensing reconstruction. Dong et al. [51] proposed an end-to-end approach named as the DPDNN. In [51], the whole iterative process was carried out six times, where the same denoising network was called each time. Despite the small number of iterations, its effect was still outstanding. Jeon et al. [73] achieved high-spectral reconstruction through a deep unfolding network.

3. Proposed Algorithm for Image Restoration

In this section, we introduce the principle and process of our method in detail. The general form of the analytic solution is given. Additionally, its application and variation to the three IR tasks are discussed in detail.

3.1. Our End-to-End Unrolled Network

Generally, the goal of the IR task is to obtain an output with a lower cost. In our method, the cost function is described in Equation (3). The HQS method converts our cost function into two sub-problems, as described in Equation (4). In this way, the two sub-problems are easy to solve. The first convex equation in Equation (4) can be solved by gradient descent or in the form of an analytic solution. The gradient descent algorithm is a simple and generalized method. The first-order method is often used in various inverse problems because it can obtain a good result after many iterations, and the second-order method is too computationally expensive. As we stated before, the gradient descent algorithm usually requires multiple iterations because the solution is not accurate. This leads to high time costs in traditional methods. The number of iterations is limited by time and space costs in DCNN-based IR methods. In this paper, we solve the first equation in (4) in the form of an analytic solution. By deriving x and setting the derivative equal to zero, the following formula can be obtained:
A T A + μ I x ^ = A T y + μ z ^
It is evident that matrix inversion is a stumbling block on the road to an analytic solution because of its computational complexity. We use the singular value decomposition of A to reduce the computational complexity of matrix inversion [74] because the cost of the inversion a diagonal matrix is small. Through SVD on degenerate matrix A, the analytic solution can be written in the following form:
x ^ = V A S A T S A + μ I 1 V A T V A S A T U A T y + μ z ^
where A = UASAVAT. As we can see, the updates of the analytic solution can be calculated quickly and efficiently. The overall framework of our proposed approach is shown in Figure 1a. Since the first equation in Equation (4) is convex, we can obtain the optimal solution for each iteration. This lays the foundation for our end-to-end network consisting of only a few iterations to achieve excellent results. The solution of the latter equation in Equation (4) is a proximal operator, which is as follows:
p r o x τ ϕ x ^ = a r g   min z 1 2   p r o x τ ϕ || x ^ z || 2 2 + τ · ϕ z
where τ = λ / μ . In this step, the prior information is important. The research on IR methods under unrolled optimization shows that the DCNNs can express image priors implicitly. Combined with the DCNN’s strong fitting ability, we use the deep denoiser network to solve the latter problem in Equation (4). We use different deep prior networks at different stages to better adapt to the image characteristics of each stage. The proposed end-to-end deep analytic network based on deep priors is summarized in Algorithm 1. In our method, the number of iterations is set to six. Our deep denoiser networks at each stage are small. Benefiting from the above settings, our overall model is not too large, which allows it to avoid overfitting.
Algorithm 1. DCNNs based end-to-end unrolled network for IR
Input: A ,   y , μ 0 > 0 ,   k = 0
Initialization:
   ( 1 )   Initial   estimation :   x 0 ^ = A T y   or   the   least   squares   estimate ,   z 0 ^ = x 0 ^
   ( 2 )   SVD :   A = U A · S A · V A T
For i t e r = 1 : k  do
  (1) Analytic updates:
          x k ^ = V A S A T S A + μ k I 1 S A T U A T y + μ k V A T z k 1 ^
   ( 2 )   Deep   prior   net :   z k ^ = D k x k ^
end
Output z k ^

3.2. Structure of the Deep Denoiser Network

The well-known U-Net [54] was proposed for medical image segmentation originally. Additionally, it has been widely used in visual tasks due to its excellent effects. Inspired by those works, our proposed DCNNs is a residual learning network with a U-Net structure. The architecture of our proposed deep network is illustrated in Figure 1b. The network is a four-scale U-Net with a soft-threshold function. The first half is a multiscale encoder for feature extraction, and the second half is a multiscale decoder for image reconstruction based on these features. In each scale of the encoder, we use two convolutional layers to encode spatial features and a max-pooling layer to increase the receptive field. The number of channels in the two convolutional layers in the first two scales of the encoder is 32. In the third scale, the number of channels is 64. After three feature extractions, there are two 64-channel convolutional layers at the top of the DCNNs. The kernel size of each convolutional layer is 3 × 3. In each scale of the decoder, there is a trans-convolutional layer, a skip layer, and two convolutional layers. The number of channels in the two convolutional layers in the first scale of the decoder is 32. In other two scales of the decoder, the number of channels is 64. The skip layer combines feature maps of the same size to compensate for the loss of spatial details caused by multiple extraction operations. After the decoder, a soft-threshold function is used to shrink the multichannel image. Then, a convolutional layer restores the image to the original color space. Finally, we establish a long residual connection between the input and output, because residual learning is easier to optimize [30] and more robust [51]. Different from the original U-Net, we adopt the leaky ReLU [75] as the activation function.

3.3. Variation in Three Applications

We have introduced the principles and process of our method and gave the general form of analytic solutions. In this section, we discuss its specific form and variations in three visual problems (denoising, deblurring, and lensless imaging). In the denoising problem, the system degradation matrix A is the identity matrix. In this case, the analytic solution degenerates into the following form:
x ^ = 1 1 + μ y + μ z
As we can see, the update of the analytic solution is a basic matrix operation, which can be completed quickly and efficiently. For deblurring with a uniform kernel, the former equation in Equation (4) is usually written in a convolutional form:
x ^ = a r g   min x || y p s f x || 2 2 + μ || x z ^ || 2 2
where ∗ denotes a two-dimensional convolution operation. In this situation, the system degradation matrix is a large sparse blurring matrix A. It is not wise to solve the equation in matrix form. Hence, we obtain an analytic solution in the frequency domain based on energy equality, as shown below:
x ^ = F 1 μ k F z ^ + F p s f ¯ F y F p s f ¯ F p s f + μ
where F and F 1 represent the fast Fourier transform (FFT) and the inverse FFT, respectively. M ¯ represents the complex conjugate of matrix M. We use Equation (10) instead of analytic updates in Algorithm 1 for deblurring. The third scenario is a lensless imaging problem named FlatCam [21]. In FlatCam, the system model is:
y = Φ L x Φ R T + n
where Φ L , Φ R are system transfer matrixes. n denotes noise. Therefore, the former equation in (4) becomes the following:
x ^ = a r g   min x || y Φ L x Φ R T || 2 2 + μ || x z ^ || 2 2
The corresponding analytic solution is as follows:
x ^ = V L U L T y U R σ L σ R T + μ V L T z ^ V R σ L 2 σ R T 2 + μ · || o n e s || V R T
where U L , S L , V L T = SVD Φ L ,   U R , S R , V R T = SVD Φ R , and the vectors σ L ,   σ R are the diagonal entries of S L ,   S R . || o n e s || denotes a matrix in which all elements are ones. is the Hadamard product. SVD is carried out in advance, and then the above formula can be calculated efficiently. In the lensless image restoration experiment, Equation (13) is used instead of the analytic update in Algorithm 1.
As shown in [76], the analytic updates of the lensless model y = Φ L x Φ R T also do not introduce singularity and gradient explosion. This lays a theoretical foundation for our network to successfully complete the training process. In the actual training process, we record the PSNR values of the test image for different epochs, as shown in Figure 2. It can be seen that the losses converge gently to a straight line, which confirms the above analysis from the side.

4. Experiments

In this section, we perform experiments on three IR tasks: image denoising, image deblurring, and lensless imaging. We fix and train a model for each specific problem. All models are implemented in TensorFlow [77] and trained on a Linux server with an Intel E5-2678 CPU at 2.5 GHz with 64 GB of memory and four graphic cards (NVIDIA GTX1080Ti) with 11 GB of memory. We train our models through the ADAM optimizer [78] by setting β 1 = 0.9 , β 2 = 0.999 , and ε = 10 8 . 2 l o s s is used as the loss function in all experiments, which means that l o s s = z 6 G T || 2 2 .
We train each model for 50 epochs. In addition, the PSNR and SSIM [79] metrics are used for objective evaluation.

4.1. Ablation Study

Dong et al. [51] undertook some research into the deep unfolding IR method. Additionally, he performed a comparative experiment of deep unfolding net and pure DCNNs. His results show that the performance of deep unfolding net which combining traditional with deep learning is better than pure DCNNs. On this basis, we conducted five groups of deblurring experiments to show the superiority of the analytical solution under the deep unfolding framework. The datasets and training settings in each group used were the same. More details are described in Section 4.3. The commonly used 10 images are used for deblurring tests are shown in Figure 3. The results are summarized in Table 1.
DPDNN is a deep unfolding net that uses gradient descent in the first step and a net in the second step. The DPDNN-AS method replaces the gradient descent in the original DPDNN with the analytic solution and keeps the rest unchanged. From Table 1, the average PSNR is increased over 0.2 dB by using the analytic solution, which demonstrates the advantage of the analytical solution. In addition, the increased PSNR between our network and the DPDNN-AS method shows the power of six small deep prior networks that have stronger adaptabilities at different stages.
We also performed the ablation study results on different iteration numbers of our method. Two groups of deblurring experiments on the Kodak24 dataset with different kernel size were conducted as examples. The results are summarized in Table 2. The comparisons of calculation amount are shown in Table 3. Ours-1, Ours-2, Ours-4, Ours-6 (Ours), and Ours-8 indicate the number of iterations in our method are 1, 2, 4, 6, and 8, respectively. Considering the effectiveness and computation cost, the number of iterations is set to six in our method. Compared with a fixed DCNN (DPDNN), our method greatly reduces the number of FLOPs when the total parameters are the same and the number of iterations is the same.

4.2. Image Denoising

In image denoising, the analytic updates are as in Equation (8) and x 0 ^ = y . To train our network, we built a training dataset from the DIV2K dataset [81]. First, we cut out 27,594 small patches form DIV2K, each of which was   256 × 256 in size. Then, we added zero-mean Gaussian noise with the variance of σ n to these small patches. Finally, we saved the values of these patches as integers from 0 to 255. In this way, the original patch and the patch after adding noise formed a training pair. We performed three groups of experiments for image denoising, and σ n was set as 15, 25, and 50. In these three experiments, the batch size was 32, the initial value of μ 0 was 0.9, and learning rate was 0.0005. The learning rate was halved after every five epochs.
To illustrate the excellent effect of our network, we compare it with the existing model-based methods—i.e., the BM3D method [4], EPLL method [5], and WNNM [40]—and learning based methods—i.e., the TNRD [60], IRCNN [61], DnCNN-S [30], and FFDNet-cl [48]. The BSD68 dataset and the Kodak24 dataset are used for testing. The values of noisy inputs in the test are also clipped to integers between 0 and 255. The images of the above three test sets are tested in grayscale. Table 4 records the average PSNR (dB) and SSIM values of the compared methods on the BSD68 and the Kodak24 datasets. The highlighted results show that our network is better than the compared methods. Among them, the data of TNRD and IRCNN are derived from a published paper [61], and the data of other methods are obtained according to public codes.
Figure 4 shows the denoising result of the well-known image Lena. As we can see, our result is more delicate and maintains more thin lines in the complex area than other results. When the noise level is σ n = 25 , our results are also better than the compared results. As shown in the green box in Figure 5, our method works well in the high-frequency region and restores more details and textures than other methods.

4.3. Image Deblurring

In order to verify the deblurring ability of our network, we performed five groups of experiments. In these experiments, we used the clear image convolution blur kernel to obtain the data set. Three blur kernels were selected, including a 25 × 25 Gaussian blur kernel with a standard deviation of 1.6 and two motion blur kernels from [80], one of which was 17 × 17, and the other was 19 × 19. For image deblurring, the analytic update in Algorithm 1 is shown in Equation (10), and   x 0 ^ = A T y . To train our deblurring network, we first convoluted the images in the DIV2K dataset with a blur kernel and added additive Gaussian noise with a standard deviation of σ n . In particular, we used zeros to fill the border of the image during the convolution. The convolution results were saved as integers between 0 and 255. Next, we cut the edge of each blurred image at half the length of the blur kernel and extracted 27,468 patches with size of 256 × 256 pixels from it. We trained five models for five experiments, and the blur settings are shown in Table 4. In these deblurring experiments, the batch size was 32, the initial value of μ 0 was 0.9, and the learning rate was 0.0005. Similar to the denoising networks, our deblurring networks also reduced the learning rate by half every five epochs. During the test phase, we employed the 10 commonly used images seen in Figure 3 and the Kodak24 dataset as test images. Similarly, the above convolution was used to generate the blurred inputs for the test images. All images were operated at grayscale. With the purpose of demonstrating the excellent performance of our network, we compare it with the classical model-based methods (IDD-BM3D [11], EPLL [5], and NCSR [26]), the denoising-based IR method under framework of plug and play (IRCNN [61]), and the end-to-end deep unfolding network (DPDNN [51]). The PSNR values of the deblurring results of 10 commonly used images (see Figure 3) are shown in Table 5. The results of the IDD-BM3D, EPLL, NCSR, and IRCNN methods are obtained by restoring the test images with open codes. If there is a ringing effect in the test process, we use the egdetaper function of MATLAB to perform an edge-preserving processing on the input.
We retrain the DPDNN method with our training set according to open codes. It can be seen from Table 5 that our method is obviously superior to the compared methods. Our results are 0.36 dB higher than those of the DPDNN, on average. We selected three groups of images to visually show the effect of the deblurring, as shown in Figure 6, Figure 7 and Figure 8. Among them, Figure 6 shows that our result recovers the most object information in the region close to the background. Figure 7 shows the effect of processing high-frequency areas, and the enlarged green boxes show that we have restored more hairs and details.
In addition to the good performance of our method on the motion blur kernel, our method also works well on the Gaussian kernel. As shown in Figure 8, our result has high contrast and sharp edges. We also test our network on the Kodak24 dataset. The PSNR and SSIM values are summarized in Table 6. Our results are superior to the other results.

4.4. Lensless Imaging

In this section, we apply our network to the lensless FlatCam. The imaging model is as in Equation (11). The corresponding analytic solution is shown in Equation (13). x 0 ^ is the least squares estimation with the Tikhonov regularization term. We use the training set from the MLS (maximum length sequence) mask in [76]. There are 10,000 pairs of data in this training set. The degraded measurement is 2048 × 2048 in size. The ground truth is 512 × 512 in size. More experimental details are described in [76]. The test images come from two places: ImageNet [82] and the valid set of DIV2K [81]. As with the training set, we displayed the test images on the screen and then record the values on the CMOS. For the restoration of FlatCam images, the batch size was 8, the initial value of μ 0 was 0.001, and the learning rate was 0.0005. Similarly, the learning rate was halved every five epochs.
After the model was trained, we compared it with the existing methods, such as Tikhonov [21], FISTA [83], and Khan’s method FlatNet [23]. Figure 9 shows the results on three images of ImageNet. In Figure 9, Khan’s results are derived from [23], and other results are obtained by our programming. Since our original images are from [23], we discarded the pixels in the lower right corner when calculating the PSNR. As shown in Figure 9, our method outperforms the other three methods. Our results are closest to the ground truth without color distortion. To test the image in DIV2K’s valid set, we resized the original image to 512. Figure 10 shows the results on 0898.png of DIV2K’s valid set. As we can see, our overall recovery rate is good, but there is still a blur in the detail area due to the ill condition of FlatCam. The average PSNR and SSIM values for 100 valid images in the DIV2K valid set are summarized in Table 7.

5. Conclusions

In this paper, we propose a DCNN denoiser based unrolled network for image restoration. We unfold the tedious iterative process in the model-based method into an end-to-end network consisting of several iterations, each of which has an analytic solution update step and a small multiscale deep denoiser network. Every DCNN serves as a denoiser rather than as the whole inverse process, which makes the network function easier to realize. In this way, our method can take advantage of optimization and the DCNNs. Specifically, we solve the convex problem in the form of an analytic solution that is faster than gradient descent under this framework. In addition, we use different multiscale prior networks in different iterations to better accommodate the image features. Compared with a fixed DCNN, it greatly reduces the number of computations when the total parameters are the same and the number of iterations is the same. Under an unrolled optimization framework, our method incorporates the physical model into the overall network, which provides a guarantee of high-quality image restoration. Visual effects and quantitative evaluation of the method for three IR tasks, including denoising, deblurring, and lensless imaging, indicate that our method achieves excellent performance in high quality image reconstruction.

Author Contributions

Conceptualization, methodology and writing, X.T.; formal analysis and software, H.Z.; validation and supervision, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by Advanced Science Key Research Project, Chinese Academy of Science (QYZDJ-SSW-JSC038); Key Research Project of International Cooperation, Chinese Academy of Science (181722KYSB20180015); National Natural Science Foundation of China (11903036, 61805243).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Boyat, A.K.; Joshi, B.K. A review paper: Noise models in digital image processing. arXiv 2015, arXiv:1505.03489. [Google Scholar] [CrossRef]
  2. Yang, C.; Feng, H.; Xu, Z.; Chen, Y.; Li, Q. Image Deblurring Utilizing Inertial Sensors and a Short-Long-Short Exposure Strategy. IEEE Trans. Image Process. 2020, 29, 4614–4626. [Google Scholar] [CrossRef]
  3. Zhang, L.; Zuo, W. Image Restoration: From Sparse and Low-rank Priors to Deep Priors. IEEE Signal. Process. Mag. 2017, 34, 172–179. [Google Scholar] [CrossRef]
  4. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-d transform-domaIn collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  5. Zoran, D.; Weiss, Y. From learning models of natural image patches to whole image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 468–479. [Google Scholar]
  6. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  7. Elad, M.; Aharon, M. Image denoising via Sparse and Redundant Representations over Learned Dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  8. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1712–1722. [Google Scholar]
  9. Anwar, S.; Barnes, N. Real image denoising with feature attention. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3155–3164. [Google Scholar]
  10. Chang, M.; Li, Q.; Feng, H.; Xu, Z. Spatial-Adaptive Network for Single Image Denoising. arXiv 2020, arXiv:2001.10291. [Google Scholar]
  11. Danielyan, A.; Katkovnik, V.; Egiazarian, K. BM3D Frames and Variational Image Deblurring. IEEE Trans. Image Process. 2012, 21, 1715–1728. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Ji, H.; Wang, K. Robust Image Deblurring With an Inaccurate Blur Kernel. IEEE Trans. Image Processi. 2012, 21, 1624–1634. [Google Scholar] [CrossRef] [PubMed]
  13. Schmidt, U.; Rother, C.; Nowozin, S.; Jancsary, J.; Roth, S. Discriminative Non-blind Deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 604–611. [Google Scholar]
  14. Pan, J.; Sun, D.; Pfister, H.; Yang, M. Blind image deblurring using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1628–1636. [Google Scholar]
  15. Kupyn, O.; Budzan, V.; Mykhailych, M.; Mishkin, D.; Matas, J. DeblurGAN: Blind Motion Deblurring Using Condational Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8183–8192. [Google Scholar]
  16. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image Super-resolution via Sparse Representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  17. Egiazarian, K.; Katkovnik, V. Single image super-resolution via BM3D sparse coding. In Proceedings of the European Signal Processing Conference, Nice, France, 31 August–4 September 2015; pp. 2849–2853. [Google Scholar]
  18. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  19. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Netwirk. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  20. Tai, Y.; Yang, J.; Liu, X. Image Super-resolution via Deep Recursive Residual Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  21. Asif, M.S.; Ayremlou, A.; Sankaranarayanan, A.; Veeraraghavan, A.; Baraniuk, R.G. FlatCam: Thin, Lensless Cameras Using Coded Aperture and Computation. IEEE Trans. Comput. Imaging 2017, 3, 384–397. [Google Scholar] [CrossRef]
  22. Canh, T.N.; Nagahara, H. Deep Compressive Sensing for Visual Privacy Protection in FlatCam Imaging. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019; pp. 3978–3986. [Google Scholar]
  23. Khan, S.S.; Adarsh, V.R.; Boominathan, V.; Tan, J.; Veeraraghavan, A.; Mitra, K. Towards photorealistic reconstruction of highly multiplexed lensless images. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 7860–7869. [Google Scholar]
  24. Monakhova, K.; Yurtsever, J.; Kuo, G.; Antipa, N.; Yanny, K.; Waller, L. Learned reconstructions for practical mask-based lensless imaging. Opt. Express 2019, 27, 28075–29090. [Google Scholar] [CrossRef] [PubMed]
  25. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Pho-to-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  26. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Krishnan, D.; Fergus, R. Fast image deconvolution using hyper-Laplacian priors. In Advances In Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2009; pp. 1033–1041. [Google Scholar]
  28. Bioucas-Dias, J.M.; Ario, M.; Figueiredo, A.T. A new TwIST: Two-step iterative shrinkage/thresholding algorithms for im-age restoration. IEEE Trans. Image Process. 2007, 16, 2992–3004. [Google Scholar] [CrossRef] [Green Version]
  29. Burger, H.C.; Schuler, C.J.; Harmeling, S. Image denoising: Can plaIn neural networks compete with BM3D? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2392–2399. [Google Scholar]
  30. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  31. Xu, L.; Ren, J.S.; Liu, C.; Jia, J. Deep convolutional neural network for image deconvolution. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2014; pp. 1790–1798. [Google Scholar]
  32. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  33. Osher, S.; Burger, M.; Goldfarb, D.; Xu, J.; Yin, W. An iterative regularization method for total variation-based image resto-ration. Multiscale Modelding Simul. 2005, 4, 460–489. [Google Scholar] [CrossRef]
  34. Mairal, J.; Elad, M.; Sapiro, G. Sparse representation for color image restoration. IEEE Trans. Image Process. 2008, 17, 53–69. [Google Scholar] [CrossRef] [Green Version]
  35. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD, An algorithm for designing overcomplete dictionaries for sparse representa-tion. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  36. Dong, W.; Li, X.; Zhang, L.; Shi, G. Sparsity-based image denoising via dictionary learning and structural clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 457–464. [Google Scholar]
  37. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  38. Xu, J.; Zhang, L.; Zuo, W.; Zhang, D.; Feng, X. Patch group based nonlocal self-similarity prior learning for image denoising. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 244–252. [Google Scholar]
  39. Dong, W.; Shi, G.; Li, X. Nonlocal image restoration with bilateral variance estimation: A low-rank approach. IEEE Trans. Image Process. 2013, 22, 700–711. [Google Scholar] [CrossRef]
  40. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  41. Barbu, A. Traing an active random field for real-time image denoising. IEEE Trans. Image Process. 2009, 18, 2451–2462. [Google Scholar] [CrossRef]
  42. Roth, S.; Black, M.J. Field of experts. Int. J. Comput. Vis. 2009, 82, 205–229. [Google Scholar] [CrossRef]
  43. Donoho, D.L. De-noising by soft-thresholding. IEEE Trans. Inf. Theory 1995, 41, 613–627. [Google Scholar] [CrossRef] [Green Version]
  44. Buades, A.; Coll, B.; Morel, J.M. Image denoising methods. A new nonlocal principle. SIAM Rev. 2010, 52, 113–147. [Google Scholar] [CrossRef]
  45. Cai, J.; Candes, E.J.; Shen, Z. A singular value thresholding algorithm for matrix completion. SIAM J. Optim. 2010, 20, 1956–1982. [Google Scholar] [CrossRef]
  46. Sun, J.; Tappen, M.F. Separable Markov random field model and its application in low level vision. IEEE Trans. Image Process. 2013, 22, 402–407. [Google Scholar] [CrossRef] [PubMed]
  47. Schmidt, U.; Roth, S. Shrinkage fields for effective image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2774–2781. [Google Scholar]
  48. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Zhang, Y.; Tian, Y.; Kong, Y.; Fu, B.Z.Y. Residual Dense Network for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2480–2495. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Venkatakrishnan, S.V.; Bouman, C.A.; Wohlberg, B. Plug-and-play priors for model based reconstruction. In Proceedings of the IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; pp. 945–948. [Google Scholar]
  51. Dong, W.; Wang, P.; Yin, W.; Shi, G.; Wu, F.; Lu, X. Denoising prior driven deep neural network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2305–2318. [Google Scholar] [CrossRef] [Green Version]
  52. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4539–4547. [Google Scholar]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  54. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  55. Goodfellow, I.J.; Abadie, J.P.; Mirza, M.; Xu, B.; Farley, D.W.; Ozair, S.; Courville, A. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar]
  56. Johnson, J.; Alahi, A.; Li, F. Perceptual loss for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
  57. Loffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  58. Zhang, J.; He, T.; Sra, S.; Jadbabaie, A. Why gradient clipping accelerates training: Theoretical justification for adaptivity. arXiv 2019, arXiv:1905.11881. [Google Scholar]
  59. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  60. Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1256–1272. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learing deep CNN denoiser prior for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3929–3938. [Google Scholar]
  62. Bertocchi, C.; Chouzenoux, E.; Corbineau, M.C.; Pesquet, J.C.; Prato, M. Deep unfolding of a proximal interior point method for image restoration. Inverse Probl. 2020, 36, 34005. [Google Scholar] [CrossRef] [Green Version]
  63. Teodoro, A.M.; Bioucas-Dias, J.M.; Figueiredo, M.A.T. Image restoration and reconstruction using variable splitting and class-adapted image priors. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 3518–3522. [Google Scholar]
  64. Kamilov, U.S.; Mansour, H.; Wohlberg, B. A Plug-and-play priors approach for solving nonlinear imaging inverse problems. IEEE Signal. Process. Lett. 2007, 24, 1872–1876. [Google Scholar] [CrossRef]
  65. Tirer, T.; Giryes, R. Image restoration by iterative denoising and backward projections. IEEE Trans. Image Process. 2019, 28, 1220–1234. [Google Scholar] [CrossRef]
  66. Brifman, A.; Romano, Y.; Elad, M. Turning a denoiser into a super-resolver using plug and play priors. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 1404–1408. [Google Scholar]
  67. Sun, Y.; Xu, S.; Li, Y.; Tian, L.; Wohlberg, B.; Kamilov, U.S. Regularized fourier ptychography using an online plug-and-play algorithm. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 7665–7669. [Google Scholar]
  68. Sreehari, S.; Venkatakrishnan, S.V.; Wohlberg, B.; Buzzard, G.T.; Drummy, L.F.; Simmons, J.P. Plug-and-play priors for bright field electron tomography and sparse interpolation. IEEE Trans. Comput. Imaging 2016, 2, 408–423. [Google Scholar] [CrossRef] [Green Version]
  69. Bigdeli, S.; Honzatko, D.; Susstrunk, S.; Dunbar, L.A. Image restoration using plug-and-play cnn map denoisers. arXiv 2019, arXiv:1912.09299. [Google Scholar]
  70. Chan, S.H.; Wang, X.; Elgendy, O.A. Plug-and-play ADMM for image restoration: Fixed-point convergence and applications. IEEE Trans. Comput. Imaging 2017, 3, 84–98. [Google Scholar] [CrossRef] [Green Version]
  71. Ryu, E.K.; Liu, J.; Wang, S.; Chen, X.; Wang, Z.; Yin, W. Plug-and-play methods provably converge with properly trained denoisers. arXiv 2019, arXiv:1905.05406. [Google Scholar]
  72. Zhang, J.; Ghanem, B. ISTA-Net: Interpretable Optimization-inspired deep network for image compressive sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1828–1837. [Google Scholar]
  73. Jeon, D.S.; Baek, S.H.; Yi, S.; Fu, Q.; Dun, X.; Heidrich, W.; Kim, M.H. Compact snapshot hyperspectral imaging with dif-fracted rotation. ACM Trans. Graph. 2019, 38, 1–13. [Google Scholar] [CrossRef] [Green Version]
  74. Zhou, H.; Feng, H.; Xu, W.; Xu, Z.; Li, Q.; Chen, Y. Deep denoiser prior based deep analytic network for lensless image res-toration. Opt. Express 2021, 29, 27237–27253. [Google Scholar] [CrossRef]
  75. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; Volume 28. [Google Scholar]
  76. Zhou, H.; Feng, H.; Hu, Z.; Xu, Z.; Li, Q.; Chen, Y. Lensless cameras using a mask based on almost perfect sequence through deep learning. Opt. Express 2020, 28, 30248–30262. [Google Scholar] [CrossRef]
  77. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Ten-sorFlow: Large-scale machine learning on Heterogeneous distributed systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  78. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  79. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  80. Levin, A.; Weiss, Y.; Durand, F.; Freeman, W.T. Understanding and evaluating blind deconvolution algorithms. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1964–1971. [Google Scholar]
  81. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
  82. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; BernsteIn, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  83. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2019, 2, 183–202. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Overview of our proposed method for image restoration. (a) The architecture of the proposed deep priors-based deep analytic network. (b) The structure of the deep prior network used in our method.
Figure 1. Overview of our proposed method for image restoration. (a) The architecture of the proposed deep priors-based deep analytic network. (b) The structure of the deep prior network used in our method.
Photonics 08 00376 g001
Figure 2. PSNR (dB) during the training epochs of three applications. (a) Three Gaussian denoising experiments. (b) Five image deblurring experiments. (c) Two FlatCam imaging experiments.
Figure 2. PSNR (dB) during the training epochs of three applications. (a) Three Gaussian denoising experiments. (b) Five image deblurring experiments. (c) Two FlatCam imaging experiments.
Photonics 08 00376 g002
Figure 3. The 10 commonly used images for our deblurring testing.
Figure 3. The 10 commonly used images for our deblurring testing.
Photonics 08 00376 g003
Figure 4. Image denoising results on Lena with noise level σ n = 15 . (a) The original image. (b) The result of BM3D. (c) The result of WNNM. (d) The result of FFDNet-cl. (e) Our result.
Figure 4. Image denoising results on Lena with noise level σ n = 15 . (a) The original image. (b) The result of BM3D. (c) The result of WNNM. (d) The result of FFDNet-cl. (e) Our result.
Photonics 08 00376 g004
Figure 5. Image denoising results on kodim15 in Kodak24 with noise level σ n = 25 . (a) The original image. (b) The result of WNNM. (c) The result of FFDNet-cl. (d) Our result.
Figure 5. Image denoising results on kodim15 in Kodak24 with noise level σ n = 25 . (a) The original image. (b) The result of WNNM. (c) The result of FFDNet-cl. (d) Our result.
Photonics 08 00376 g005
Figure 6. Image deblurring results on Boat in Figure 3 with 17 × 17 motion blur kernel and σ n = 2.55 . (a) The original image. (b) The result of EPLL. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Figure 6. Image deblurring results on Boat in Figure 3 with 17 × 17 motion blur kernel and σ n = 2.55 . (a) The original image. (b) The result of EPLL. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Photonics 08 00376 g006
Figure 7. Image deblurring results on Man in Figure 3 with 19 × 19 motion blur kernel and σ n = 2.55 . (a) The original image. (b) The result of IDD-BM3D. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Figure 7. Image deblurring results on Man in Figure 3 with 19 × 19 motion blur kernel and σ n = 2.55 . (a) The original image. (b) The result of IDD-BM3D. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Photonics 08 00376 g007
Figure 8. Image deblurring results on Plant in Figure 3 with 25 × 25 Gaussian kernel and σ n = 2 . (a) The original image. (b) The result of NCSR. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Figure 8. Image deblurring results on Plant in Figure 3 with 25 × 25 Gaussian kernel and σ n = 2 . (a) The original image. (b) The result of NCSR. (c) The result of IRCNN. (d) The result of DPDNN. (e) Our result.
Photonics 08 00376 g008
Figure 9. FlatCam imaging results on three images of ImageNet. (a) The results of image ‘Store’. (b) The results of image ‘Insect’. (c) The results of image ‘Boy’.
Figure 9. FlatCam imaging results on three images of ImageNet. (a) The results of image ‘Store’. (b) The results of image ‘Insect’. (c) The results of image ‘Boy’.
Photonics 08 00376 g009
Figure 10. FlatCam imaging results on 0898.png of DIV2K dataset. (a) The original image. (b) The result of Tikhonov. (c) The result of FISTA. (d) Our result.
Figure 10. FlatCam imaging results on 0898.png of DIV2K dataset. (a) The original image. (b) The result of Tikhonov. (c) The result of FISTA. (d) Our result.
Photonics 08 00376 g010
Table 1. The average PSNR of 10 images in Figure 3 by three methods.
Table 1. The average PSNR of 10 images in Figure 3 by three methods.
Kernel17 × 17 Kernel in [80]19 × 19 Kernel in [80]25 × 25 Gaussian
Noise σ2.557.652.557.652
DPDNN [51]31.9728.6532.5329.0131.01
DPDNN-AS32.3028.8932.8929.2031.24
Ours32.4229.0132.9129.2531.38
Table 2. The average PSNR and SSIM values of our method with different iteration numbers on the Kodak24 dataset.
Table 2. The average PSNR and SSIM values of our method with different iteration numbers on the Kodak24 dataset.
Number of IterationsPSNRSSIM
17 × 17 motion blur kernel of [80], σn = 2.55
Ours-131.020.857
Ours-231.480.865
Ours-431.800.872
Ours-6 (Ours)31.940.874
Ours-832.000.875
19 × 19 motion blur kernel of [80], σn = 2.55
Ours-131.320.865
Ours-231.810.873
Ours-432.180.880
Ours-6 (Ours)32.330.882
Ours-832.410.884
Table 3. Parameters and FLOPs of DPDNN and our method with different iteration number in the deblurring tasks.
Table 3. Parameters and FLOPs of DPDNN and our method with different iteration number in the deblurring tasks.
Image Size256 × 256 × 1
MethodDPDNNOurs-1Ours-2Ours-4OursOurs-8
Parameters1249K393K787K1573K2359K3146K
FLOPs794G29G59G118G177G236G
Run time (Sec)0.07120.02110.02560.04530.05750.0615
Table 4. The average PSNR values of test methods for denoising on the Bsd68 and Kodak24 datasets.
Table 4. The average PSNR values of test methods for denoising on the Bsd68 and Kodak24 datasets.
Method BM3D [4]WNNM [40]FFDNet-cl [48]TNRD [60]IRCNN [61]Ours
σ n PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
BSD681531.020.87331.230.87631.650.89031.42-31.63-31.700.891
2528.340.79728.520.80329.210.82928.92-29.15-29.250.831
5024.860.66924.810.66426.280.72525.97-26.19-26.280.726
Method BM3D [4]WNNM [40]FFDNet-cl [48]EPLL [5]DnCNN-S [30]Ours
σ n PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
Kodak241532.230.87732.460.88032.810.89232.100.88132.720.89032.850.893
2529.680.81429.890.81830.470.83829.540.81530.130.83230.510.840
5026.220.70726.230.70527.610.74825.940.69626.400.71727.530.750
Table 5. The PSNR values of the deblurred 10 images in Figure 3 by the test methods.
Table 5. The PSNR values of the deblurred 10 images in Figure 3 by the test methods.
ImageBoatC. ManFlowerHouseLena256ManMonar.ParrotsPeppersPlantAve.
Method
Gaussian blur with standard deviation 1.6, σ n   =   2
IDD-BM3D [11]29.9726.6528.4032.4929.5830.4328.3729.6229.4332.2529.72
EPLL [5]30.5526.6628.8132.8330.0330.6329.3729.8030.0232.8830.16
NCSR [26]31.1927.6229.2833.3330.3030.9329.8630.5230.2433.5630.68
IRCNN [61]31.2027.9429.6333.5330.4430.9930.5830.2430.7633.8630.92
DPDNN [51]31.1028.0829.6633.2730.7131.1330.7630.8130.6633.8931.01
Ours31.4428.5330.0133.7330.9531.2831.0631.1330.9634.6931.38
17   ×   17 motion blur kernel of [81], σ n   =   2.55
IDD-BM3D [11]30.2429.3628.7032.7130.3030.1127.3931.7028.9332.3430.18
EPLL [5]31.8529.9830.0333.9031.7031.2030.0232.2931.0333.2131.52
IRCNN [61]31.9530.8430.5133.4931.9031.3129.2033.1529.8034.0931.62
DPDNN [51]32.0230.4530.3933.9032.3531.6531.1532.8631.1333.8231.97
Ours32.5030.9030.7734.6632.7231.8831.6733.2931.3734.4532.42
17   ×   17 motion blur kernel of [81], σ n   =   7.65
IDD-BM3D [11]27.2225.7825.6130.2027.5927.2025.2527.8526.8629.2027.28
EPLL [5]26.9624.8725.0728.9327.3327.2423.7326.1427.0428.6526.60
IRCNN [61]28.5627.6926.9231.4028.8128.4127.2529.5527.7530.5228.69
DPDNN [51]28.6027.2826.8231.0828.8528.5127.4729.4528.1830.3028.65
Ours29.0327.6327.2031.7529.1728.7027.8429.7028.3830.7029.01
19   ×   19 motion blur kernel of [81], σ n   =   2.55
IDD-BM3D [11]30.2929.4229.3831.8230.4930.5228.9331.2128.9732.7230.38
EPLL [5]32.1330.5730.4733.1932.3131.5830.9132.6231.4133.7431.89
IRCNN [61]31.5930.5730.9332.0031.7431.3030.5232.4829.8834.1931.52
DPDNN [51]32.5931.0631.3633.6332.9432.0532.0233.3131.6634.6732.53
Ours33.1331.4231.8134.2733.2132.2532.4333.6131.8535.1532.91
19   ×   19 motion blur kernel of [81], σ n   =   7.65
IDD-BM3D [11]27.5426.3225.6230.1727.8927.3825.5828.0027.2929.4227.52
EPLL [5]27.0125.6124.8529.0427.5127.5424.4826.6327.4328.9926.91
IRCNN [61]28.8027.7727.4330.9229.2528.5228.0429.6828.4731.1229.00
DPDNN [51]28.8427.6127.2330.7829.2728.7228.2029.7628.6431.0229.01
Ours29.1727.7727.5331.2429.3928.8628.4329.9128.7731.4429.25
Table 6. The Average PSNR Values of The Test Methods For Deblurring on The Kodak24 Dataset.
Table 6. The Average PSNR Values of The Test Methods For Deblurring on The Kodak24 Dataset.
MethodGaussian Blur with Standard Deviation 1.6, σ n   =   2 17   ×   17 Motion Blur Kernel of [80], σ n   =   2.55 17   ×   17 Motion Blur Kernel of [80], σ n   =   7.65 19   ×   19 Motion Blur Kernel of [80], σ n   =   2.55 19   ×   19 Motion Blur Kernel of [80], σ n   =   7.65
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
IDD-BM3D [11]29.540.82830.360.82126.740.70630.300.82726.880.713
EPLL [5]29.480.82331.130.84226.820.70631.490.85227.090.715
NCSR [26]29.960.833--------
IRCNN [61]29.990.83131.330.84928.280.76530.880.84128.520.773
DPDNN [51]30.150.84231.570.86028.380.76532.000.87128.720.780
DPDNN-AS30.290.84731.840.87028.580.77532.280.87928.890.786
Ours30.430.85031.940.87428.680.77932.330.88228.990.790
Table 7. The average PSNR and SSIM values of 100 images in DIV2K’s valid set.
Table 7. The average PSNR and SSIM values of 100 images in DIV2K’s valid set.
MethodPSNRSSIM
Tikhonov [21]12.930.39
FISTA [83]14.910.42
Ours24.630.78
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tao, X.; Zhou, H.; Chen, Y. Image Restoration Based on End-to-End Unrolled Network. Photonics 2021, 8, 376. https://doi.org/10.3390/photonics8090376

AMA Style

Tao X, Zhou H, Chen Y. Image Restoration Based on End-to-End Unrolled Network. Photonics. 2021; 8(9):376. https://doi.org/10.3390/photonics8090376

Chicago/Turabian Style

Tao, Xiaoping, Hao Zhou, and Yueting Chen. 2021. "Image Restoration Based on End-to-End Unrolled Network" Photonics 8, no. 9: 376. https://doi.org/10.3390/photonics8090376

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop