Next Article in Journal
Efficient Characterization of Macroscopic Composite Cement Mortars with Various Contents of Phase Change Material
Next Article in Special Issue
A Spam Filtering Method Based on Multi-Modal Fusion
Previous Article in Journal
High-Gain Extended State Observer Based Adaptive Sliding Mode Path Following Control for An Underactuated Vessel Sailing in Restricted Waters
Previous Article in Special Issue
An Algorithm for Scene Text Detection Using Multibox and Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Deep CNN Denoiser Priors for Depth Image Inpainting

1
School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan 430081, Hubei, China
2
School of Physics and Electronic Engineering, Xinxiang College, Xinxiang 453000, Henan, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(6), 1103; https://doi.org/10.3390/app9061103
Submission received: 17 January 2019 / Revised: 11 March 2019 / Accepted: 12 March 2019 / Published: 15 March 2019
(This article belongs to the Special Issue Advances in Deep Learning)

Abstract

:
Due to the rapid development of RGB-D sensors, increasing attention is being paid to depth image applications. Depth images play an important role in computer vision research. In this paper, we address the problem of inpainting for single depth images without corresponding color images as a guide. Within the framework of model-based optimization methods for depth image inpainting, the split Bregman iteration algorithm was used to transform depth image inpainting into the corresponding denoising subproblem. Then, we trained a set of efficient convolutional neural network (CNN) denoisers to solve this subproblem. Experimental results demonstrate the effectiveness of the proposed algorithm in comparison with three traditional methods in terms of visual quality and objective metrics.

1. Introduction

Depth images are a key research topic in the field of 2D-to-3D technology, which is widely used in machine vision and 3D reconstruction. However, due to the limitations of hardware and the characteristics of 3D scenes, black holes can easily form in depth images, and they are susceptible to noise. Due to the practical benefits of using depth images, some scholars have begun to investigate the problem of depth image inpainting.
Depth images can be viewed as natural images without texture. Therefore, many natural image inpainting methods can be directly applied for depth image inpainting. In general, inpainting methods can be roughly divided into three categories, as discussed below.
Inpainting methods of the first type rely on spatial information. Several representative algorithms are described as follows. Telea et al. [1] proposed a fast depth image inpainting method based on the mechanism of human vision. This method has a high inpainting speed, but boundaries can easily become blurred. Shen et al. [2] proposed the weighted joint bilateral filter method. Similarly, although the speed of this method is high, the boundaries in the depth images are not well handled. Lee et al. [3] proposed an adaptive edge-oriented smoothing process to solve the depth image inpainting problem. This method achieves a good balance between time consumption and image quality. Methods of this type rely on general methods of image repair.
Inpainting methods of the second type use aligned color images as a guide. Several representative algorithms are described as follows. Liu et al. [4] used a deep-hole inpainting method with color guidance weights combined with bilateral differences. This method shows good robustness. Buyssens et al. [5] proposed a method based on block matching with the edges of aligned color images to guide the direction of inpainting. Deng et al. [6] trained a decision tree to identify edges in depth images and corresponding color images for inpainting. Methods of this type require constrained repair based on aligned color images, and the process is complex.
The third class of inpainting methods consists of model-based optimization methods. Several representative algorithms are described as follows. Lu et al. [7] proposed a low-rank matrix method for depth image enhancement. This method has a poor effect when large-area information is lost. Xue et al. [8] proposed the integration of low gradient regularization with low-rank regularization for depth image inpainting. This method offers improved inpainting quality. Methods of this type are flexible but time consuming. At present, they are attracting considerable attention and have become the mainstream approach to the problem at hand. Therefore, we plugged learned denoiser priors into such a model to solve the problem of depth image inpainting.
Within the framework of model-based optimization methods, we investigated inpainting for single depth images. Depth image inpainting is an ill-posed inverse problem. We can obtain a potential depth image by solving the following formula [9]:
x Λ = arg min x 1 2 | | H x y | | 2 2 + λ ψ ( x )
where x and y are the potential depth image and the degraded depth image, respectively; | | H x y | | 2 2 is the data-fidelity term; ψ ( x ) is the regularization term; and λ is a weight parameter. H is a binary template in which missing information is represented by values of 0 and all other information is represented by values of 1. We hope to obtain a potential depth image x from a degraded depth image y .
With the help of variable splitting techniques, many scholars have decoupled the data-fidelity term and the regularization term in Equation (1). Then, they have used various strategies to solve the resulting subproblems. The most commonly used variable splitting techniques are the iterative shrinkage/thresholding (IST) algorithm [10] and the split Bregman iteration (SBI) algorithm [11].
First, we adopt the SBI algorithm for the separation of the terms in Equation (1), and then, we train a set of effective CNN denoisers and integrate the results into the framework of model-based optimization methods to complete the task of depth image inpainting. The structure of the CNN includes rectified exponential linear units (RELUs) [12], batch normalization (BN) [13] and residual learning [14]. The proposed method can be used to more effectively complete the task of depth image inpainting. We chose CNN denoisers for the following reasons:
  • The CNN inference process is very efficient due to the parallel computation capability of Graphics Processing Units (GPUs).
  • A CNN with a deep architecture exhibits a powerful prior modeling capacity.
  • The deep architecture of the CNN facilitates discrimination learning.
Our contributions are summarized as follows: (1). The SBI algorithm is used to separate the data-fidelity term and the regularization term. Therefore, the depth image inpainting problem is transformed into the corresponding denoising subproblem. (2). We trained a set of effective CNN denoisers that are used as important components of a model-based optimization method to solve the denoising subproblem.
The remainder of our paper is organized as follows: Section 2 describes the related work on variable splitting techniques and the application of denoiser priors in image reconstruction. Section 3 introduces the method of training deep CNN denoisers for depth image inpainting. In Section 3, we also present important details of the CNN structure. In Section 4, we report experiments performed on our dataset and compare the results of various algorithms. In Section 5, we discuss future research directions.

2. Related Works

2.1. Variable Splitting Techniques

Variable splitting techniques are widely used in image reconstruction. The variable splitting techniques are usually adopted to decouple the data-fidelity term and the regularization term in Equation (1). This approach offers good convergence and effectiveness.
In Reference [9], Zhang et al. compared the performance of the IST and SBI algorithms. They then used the SBI algorithm to decouple the data-fidelity term and the sparse representation regularization term. Xue et al. used the alternating direction method of multipliers (ADMM) and the SBI algorithm to decouple the data-fidelity term and the regularization term in [8]. In Reference [15], the ADMM was adopted to decouple the data-fidelity term and the regularization term in class-adaptive image prior extraction for image reconstruction. Reports of similar approaches can also be found in References [16,17,18] and others.
Depth image inpainting is a branch of image reconstruction. In this paper, we adopt the SBI algorithm due to its simplicity. We then use the CNN denoiser to solve the resulting subproblems.

2.2. Convolutional Neural Network (CNN) Denoiser Priors

At present, many scholars are committed to using denoiser priors for image reconstruction. This method usually involves the separate processing of the data-fidelity subproblem and the denoising subproblem.
In Reference [14], within the framework of the ADMM, He et al. adopted a class-adaptive denoiser based on a Gaussian mixture model (GMM) to address image deblurring and image compression. In Reference [19], a block matching and 3D filtering (BM3D) denoiser model was proposed to solve the problem of single image super-resolution. In Reference [20], within the framework of spatially adaptive nonlocal patchwise estimation, a BM3D denoiser was used to address the problem of image deblurring. In Reference [21], a set of trained CNN denoisers was developed for image denoising, image super-resolution and image deblurring. Some similar approaches can also be found in References [22,23].
All above methods have shown that the existing denoising models solve many image restoration problems. Considering the performance and effectiveness of the various available algorithms, we chose the CNN-based denoiser approach.

3. Learning Deep CNN Denoiser Priors for Depth Image Inpainting

3.1. Split Bregman Iteration (SBI) Method

By introducing the auxiliary variable u , Equation (1) can be reformulated as a constrained optimization problem:
x ^ = arg min x 1 2 | | H x y | | 2 2 + λ ψ ( u ) s . t . u . = x
Then, by invoking the SBI algorithm, we separate the data-fidelity term and the regularization term. Equation (2) can be solved by solving the following two subproblems:
x k + 1 = arg min x 1 2 | | H x y | | 2 2 + μ 2 | | x u k b k | | 2 2
u k + 1 = arg min u λ ψ ( u ) + μ 2 | | x k + 1 u b k | | 2 2
b k + 1 = b k ( x k + 1 u k + 1 )
Given u , Equation (3a) is converted into a strict minimum problem of a strictly convex quadratic function. We can thus obtain a closed-form solution:
x k + 1 = ( H T H + μ I ) 1 [ H T y + μ ( u k + b k ) ]
where I is the identity matrix.
Given x , Equation (3b), in which the regularization term appears, can be converted into the following formula:
u k + 1 = arg min u ψ ( u ) + 1 2 ( λ / u ) 2 | | r k + 1 u | | 2 2
where r k + 1 = x k + 1 b k .
In Equation (5), we view r as a sort of noisy observation of u . Therefore, Equation (5) corresponds to the image u k + 1 produced by a Gaussian denoiser with noise level λ / μ . Therefore, Equation (5) can be rewritten as follows:
u k + 1 = d e n o i s e r ( r k + 1 , λ / μ )
In summary, the depth image inpainting problem is converted into the corresponding denoising subproblem.

3.2. CNN Denoiser

A CNN denoiser has the advantages of high efficiency and a strong modeling capability. Therefore, we adopted a CNN denoiser [24] for solving the u subproblem in depth image inpainting.
The structure of the CNN denoiser is illustrated in Figure 1. The basic units are Conv, RELU, and BN types.
The L layers of the CNN denoiser include three types of blocks with different functions.
The first layer is of the first block type: “Conv +RELU”. This layer generates 64 feature maps, and the RELU function enhances the nonlinear relationship between the layers.
The second through (L−1)th layers are of the second block type: “Conv + BN + RELU”. The BN unit is added between the Conv and RELU components in this block type. BN can speed up training and improve the denoising performance.
The Lth layer is of the third block type: “Conv”. This layer is used to reconstruct the output.
The important details of the CNN will be given below.

3.2.1. Network Depth and Receptive Field Size

According to the rules presented in References [25,26], the size of the convolution filter should be set to 3 × 3 . Therefore, the receptive field size, s , and the network depth, d , satisfy s = 2 d + 1 . In a CNN, a larger receptive field can better utilize the known information for denoising.
To balance efficiency and effectiveness, it is important to choose a reasonable network depth.
Depth images can be regarded as textureless natural images with low noise levels [21,26]. Therefore, the network depth can be set to 7, and the receptive field size is 15 × 15 .

3.2.2. Integration of Residual Learning and Batch Normalization (BN)

We cite the classic design of the Denoising convolutional neural network (DnCNN), which combines residual learning and Batch Normalization (BN) [27,28]. This integrated approach can achieve the optimal denoising performance.
The residual learning method is used for training to obtain the complete image or residual image from potential information. When the original mapping approaches an identity mapping, this method is more suitable for obtaining the residual image. Therefore, we adopt a single residual unit to predict the residual image.
The BN operation is similar to the regularization operation. BN alleviates internal covariate transfer by means of a normalization operation before the nonlinear operation in each layer. This step can speed up the convergence of the network and reduce the dependence on the initial parameters. BN units have been widely used in deep learning networks [23].
In the proposed network structure, the BN operation is introduced to speed up learning and prevent overfitting.
In summary, this integrated approach not only enables stable training but also boosts the denoising performance.

3.2.3. Zero Padding

In depth image inpainting, the sizes of the input and output need to be consistent. However, this may lead to boundary artifacts. There are two ways to solve this problem: symmetrical padding and zero padding. We chose zero padding to maintain the image size while avoiding boundary artifacts.

3.2.4. Learning of CNN Denoiser Priors

The depth image inpainting problem is reduced to the denoising subproblem in our approach. As the proposed method requires iteration, it is necessary to train denoisers for different image noise levels. In depth images, the noise level is generally low; therefore, we trained a set of denoisers for noise levels in the range (0,30) in increments of two.
Once the denoising subproblem had been solved as described above, we adopted an effective solution to ensure the optimal solution of the overall algorithm.

4. Experiments and Results

It is well known that CNNs benefit from the availability of large training data. To learn the CNN denoiser, we followed Reference [23] in using 2000 images from ImageNet [29] for training. The noisy images were obtained by adding Gaussian noise for noise levels in the range (0,30) in increments of two. The hardware simulation platform was a Lenovo R720 computer (Lenovo company, Haidian district, Beijin, China) with an Nvidia GPU. The software platform used for simulation was MATLAB R2017a with MatConvNet. It takes about one day to train the CNN denoiser. We empirically found that using a larger training database can slightly improve the effect and greatly increase time.
Under the same settings, the inpainting effect should be evaluated in terms of both subjective visual quality and objective metrics. As depth image inpainting is not yet a mature area of research, there are no public datasets or algorithms available for comparison; therefore, we introduced three sets of depth images for comparative experiments: the Mask depth image set, the Teddy depth image set and the Art depth image set. These image sets include both natural areas and artificial areas that need to be repaired. For comparison with our method, we used the total variation (TV) [30], group-based sparse representation (GSR) [9] and low rank with L0 gradient regularization (LRL0) [8] algorithms.
The objective metrics can be assessed on the basis of two objective metrics: the peak signal-to-noise ratio (PSNR) and the feature similarity (FSIM) index [31].
In Experiment (1), we obtained degraded depth images directly from a Kinect sensor, as shown in Figure 2.
In Experiment (2), we obtained degraded depth images by adding artificial text, as shown in Figure 3. The original images were the images repaired with the GSR algorithm in Experiment (1).
Subjectively, all four algorithms meet the visual requirements, as shown in Figure 4 and Figure 5. However, the inpainting effect of the TV algorithm is blurred, and the boundaries are smoothed. The GSR algorithm achieves a better repair effect at the boundaries but can easily lead to blur in areas where the gray values are similar, also resulting in unclear boundaries. The boundary repair effect is improved in both the LRL0 algorithm and the proposed algorithm, and the inpainting effect is better for fine image details.
Objectively, it can be seen from Table 1 and Table 2 that the proposed algorithm is superior to the other three algorithms. For different types of images, the objective metrics of our proposed algorithm are the best.
In summary, the proposed algorithm has advantages in terms of its inpainting effect. Hence, the proposed algorithm is an effective inpainting method for single depth images.

5. Conclusions

The main research issue addressed in this paper is inpainting for single depth images. With the help of the SBI algorithm, we transformed the problem of single depth image inpainting into the corresponding denoising subproblem. Then, we cited and trained a set of effective CNN denoisers for solving the denoising subproblem. Experimental results prove that the proposed algorithm has value for practical applications and that its inpainting effect is superior to that of other algorithms.
However, the proposed approach still has certain shortcomings. On the one hand, it will be interesting to investigate how to balance the efficiency and effect of the CNN denoiser. On the other hand, an effective compression method is needed for the CNN architecture. These issues will be addressed in our future research.

Author Contributions

L.Z.: methodology, software, validation, analysis, writing; W.J.: writing-review and editing.

Funding

This work was supported by the National Natural Science Foundation of China under Grant Nos. U1704132 and 11747089 and by The Ninth Group of Key Disciplines in Henan Province under Grant No. 2018119.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Telea, A. An image inpainting technique based on the fast marching method. J. Graph. Tools 2004, 9, 23–24. [Google Scholar] [CrossRef]
  2. Shen, Y.; Li, J.; Lu, C. Depth map enhancement method based on joint bilateral filter. In Proceedings of the 2014 7th International Congress on Image and Signal Processing, Dalian, China, 14–16 October 2014; pp. 153–158. [Google Scholar]
  3. Lee, P.J. Nongeometric Distortion Smoothing Approach for Depth Map Preprocessing. IEEE Trans. Multimed. 2011, 13, 246–254. [Google Scholar] [CrossRef]
  4. Liu, W.; Chen, X.; Yang, J.; Wu, Q. Robust Color Guided Depth Map Restoration. IEEE Trans. Image Process. 2017, 26, 315–327. [Google Scholar] [CrossRef] [PubMed]
  5. Buyssens, P.; LeMeur, O.; Daisy, M. Depth-guided disocclusion inpainting of synthesized RGB-D images. IEEE Trans. Image Process. 2017, 26, 525–538. [Google Scholar] [CrossRef] [PubMed]
  6. Deng, H.; Wu, J.; Zhu, L.; Yan, Z.; Yu, L. Texture edge-guided depth recovery for structured light-based depth sensor. Multimed. Tools Appl. 2017, 76, 4211–4226. [Google Scholar] [CrossRef]
  7. Lu, S.; Ren, X.; Liu, F. Depth enhancement via low-rank matrix completion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3390–3397. [Google Scholar]
  8. Xue, H.; Zhang, S.; Cai, D. Depth image inpainting: Improve Low Rank Matrix completion with Low Gradient Regularization. IEEE Trans. Image Process. 2017, 26, 4311–4320. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, J.; Zhao, D.; Gao, W. Group-based Sparse Representation for Image Restoration. IEEE Trans. Image Process. 2014, 23, 3336–3351. [Google Scholar] [CrossRef] [PubMed]
  10. Bioucas-Dias, J.; Figueiredo, M. A new TwIST: Two-step iterative shrinkage/thresholding algorithm for image restoration. IEEE Trans. Image Process. 2007, 16, 1620–1630. [Google Scholar] [CrossRef]
  11. Goldstein, T.; Osher, S. The split Bregman algorithm for L1 regularized problem. SIAM J. Imaging Sci. 2009, 2, 323–343. [Google Scholar] [CrossRef]
  12. Clevert, D.; Unterthiner, T.; Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). In Proceedings of the International Conference on Learning Representations, San Juan, PR, USA, 2–4 May 2016; pp. 1–14. [Google Scholar]
  13. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  14. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  15. Afonso, M.; Teodoro, M.; Bioucas-Dias, M. Image restoration and reconstruction using variable splitting and class-adapted image priors. In Proceedings of the 2016 IEEE International Conference on Image Processingv, Phoenix, AZ, USA, 25–28 September 2016; pp. 25–28. [Google Scholar]
  16. Danieyan, A.; Katkovnik, V.; Egiazarian, K. BM3D frames and variational image deblurring. IEEE Trans. Image Process. 2014, 21, 1715–1728. [Google Scholar] [CrossRef] [PubMed]
  17. Chambolle, A.; Pock, T. A first-order primal-dual algorithm for convex problem with application to imaging. J. Math. Imaging Vision 2011, 40, 120–145. [Google Scholar] [CrossRef]
  18. Heide, F.; Steinberger, M.; Tsai, Y. A flexible camera image processing frame-work. ACM Trans. Graph. 2014, 33, 231–248. [Google Scholar] [CrossRef]
  19. Egiazarian, K.; Katkovnik, V. Single image super-resolution via BM3D sparse coding. In Proceedings of the European Signal Processing Conference, Nice, France, 31 August–4 September 2015; pp. 2849–2853. [Google Scholar]
  20. Danieyan, A.; Katkovnik, V.; Egiazarian, K. Image deblurring by argumented lagrangian with BM3D frame prior. In Proceedings of the Workshop on information Theoretic Method in Science and Engineering, Tampere, Finland, 16–18 August 2010; pp. 16–18. [Google Scholar]
  21. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN denoiser Prior for Image Restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1–10. [Google Scholar]
  22. Rond, A.; Giryes, R.; Elad, M. Possion inverse problems by the plug-and-play scheme. J. Vis. Commun. Image Represent. 2016, 2, 96–108. [Google Scholar] [CrossRef]
  23. Zhenyu, G.; Yujuan, S.; Muwei, J.; Xiaofeng, Z. Deep Residual Network with Sparse Feedback for Image Restoration. Appl. Sci. 2018, 8, 2417. [Google Scholar]
  24. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2014, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  25. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 2–5. [Google Scholar]
  26. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolution network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
  27. Timofte, R.; De Smet, V.; Van Gool, L. A+: Adjust anchored neighborhood regression for fast super-resolution. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 111–126. [Google Scholar]
  28. Kiku, D.; Monno, Y.; Tanaka, M.; Okutomi, M. Residual interpolation for color image demosaicking. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 2304–2308. [Google Scholar]
  29. Deng, J.; Dong, W.; Socher, R. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  30. Chan, T.; Shen, J. Mathermatical models for Local Non-texture inpaintings. SIAM J. Appl. Math. 2002, 62, 1019–1043. [Google Scholar]
  31. Zhang, L.; Zhang, L.; Mouxuan, Q.; Zhang, D. FSIM: A Feature Similarity Index for Image Quality Assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. The structure of the proposed convolutional neural network (CNN) denoiser. RELUs: rectified exponential linear units, BN: batch normalization.
Figure 1. The structure of the proposed convolutional neural network (CNN) denoiser. RELUs: rectified exponential linear units, BN: batch normalization.
Applsci 09 01103 g001
Figure 2. Experimental test images (1): (a) represents the Mask depth image; (b) represents the Teddy depth image; (c) represents the Art depth image.
Figure 2. Experimental test images (1): (a) represents the Mask depth image; (b) represents the Teddy depth image; (c) represents the Art depth image.
Applsci 09 01103 g002
Figure 3. Experimental test images (2): (a) represents the Mask depth image; (b) represents the Teddy depth image; (c) represents the Art depth image.
Figure 3. Experimental test images (2): (a) represents the Mask depth image; (b) represents the Teddy depth image; (c) represents the Art depth image.
Applsci 09 01103 g003
Figure 4. Visual contrasting of the inpainting results (1): (a)(d) represent the inpainting effects of the four algorithms (total variation (TV), group-based sparse representation (GSR), low rank with L0 gradient regularization (LRL0) and the proposed algorithm) for the mask depth image; (e)–(h) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the toy depth image; and (i)–(l) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the Art depth image.
Figure 4. Visual contrasting of the inpainting results (1): (a)(d) represent the inpainting effects of the four algorithms (total variation (TV), group-based sparse representation (GSR), low rank with L0 gradient regularization (LRL0) and the proposed algorithm) for the mask depth image; (e)–(h) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the toy depth image; and (i)–(l) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the Art depth image.
Applsci 09 01103 g004
Figure 5. Visual contrasting of the inpainting results (1). (a)–(d) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the mask depth image; (e)–(h) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the toy depth image; and (i)–(l) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the Art depth image.
Figure 5. Visual contrasting of the inpainting results (1). (a)–(d) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the mask depth image; (e)–(h) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the toy depth image; and (i)–(l) represent the inpainting effects of the four algorithms (TV, GSR, LRL0 and the proposed algorithm) for the Art depth image.
Applsci 09 01103 g005
Table 1. Peak signal-to-noise ratio (PSNR) and feature similarity (FSIM) result in Experiment (1).
Table 1. Peak signal-to-noise ratio (PSNR) and feature similarity (FSIM) result in Experiment (1).
ImageAlgorithm (PSNR/dB)/FSIM
TVGSRLRL0Proposed
Mask Depth Image22.1304/0.925723.2678/0.928824.2544/0.999324.3889/0.9993
Teddy Depth Image21.3046/0.931422.5604/0.936923.9979/0.999324.0529/0.9993
Art Depth Image27.2198/0.968328.4787/0.976530.0326/0.999930.4045/0.9999
Table 2. PSNR and FSIM result in Experiment (2).
Table 2. PSNR and FSIM result in Experiment (2).
ImageAlgorithm (PSNR/dB)/FSIM
TVGSRLRL0Proposed
Mask Depth Image25.2308/0.929730.7068/0.986433.4979/133.7880/1
Teddy Depth Image27.9409/0.932032.9311/0.988433.8560/134.3918/1
Art Depth Image27.5757/0.975031.1104/0.983931.4780/131.9801/1

Share and Cite

MDPI and ACS Style

Li, Z.; Wu, J. Learning Deep CNN Denoiser Priors for Depth Image Inpainting. Appl. Sci. 2019, 9, 1103. https://doi.org/10.3390/app9061103

AMA Style

Li Z, Wu J. Learning Deep CNN Denoiser Priors for Depth Image Inpainting. Applied Sciences. 2019; 9(6):1103. https://doi.org/10.3390/app9061103

Chicago/Turabian Style

Li, Zun, and Jin Wu. 2019. "Learning Deep CNN Denoiser Priors for Depth Image Inpainting" Applied Sciences 9, no. 6: 1103. https://doi.org/10.3390/app9061103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop