Next Article in Journal
Reducing False Arrhythmia Alarms Using Different Methods of Probability and Class Assignment in Random Forest Learning Methods
Next Article in Special Issue
Single-Image Depth Inference Using Generative Adversarial Networks
Previous Article in Journal
A New Maritime Moving Target Detection and Tracking Method for Airborne Forward-looking Scanning Radar
Previous Article in Special Issue
Combining Weighted Contour Templates with HOGs for Human Detection Using Biased Boosting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Depth Map Upsampling via Multi-Modal Generative Adversarial Network

1
Department of Computer Science and Information Engineering, National Taiwan University of Science and Technology, Taipei 10607, Taiwan
2
Center for Automation Research, College of Computer Studies, De La Salle University, Manila 1004, Philippines
3
Center for Cyber-Physical System Innovation, National Taiwan University of Science and Technology, Taipei 10607, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(7), 1587; https://doi.org/10.3390/s19071587
Submission received: 16 February 2019 / Revised: 28 March 2019 / Accepted: 29 March 2019 / Published: 2 April 2019
(This article belongs to the Special Issue Selected Papers from INNOV 2018)

Abstract

:
Autonomous robots for smart homes and smart cities mostly require depth perception in order to interact with their environments. However, depth maps are usually captured in a lower resolution as compared to RGB color images due to the inherent limitations of the sensors. Naively increasing its resolution often leads to loss of sharpness and incorrect estimates, especially in the regions with depth discontinuities or depth boundaries. In this paper, we propose a novel Generative Adversarial Network (GAN)-based framework for depth map super-resolution that is able to preserve the smooth areas, as well as the sharp edges at the boundaries of the depth map. Our proposed model is trained on two different modalities, namely color images and depth maps. However, at test time, our model only requires the depth map in order to produce a higher resolution version. We evaluated our model both quantitatively and qualitatively, and our experiments show that our method performs better than existing state-of-the-art models.

1. Introduction

The standard digital camera tries to capture three-dimensional scenes and projects them onto a two-dimensional image plane. This process inevitably loses much information, particularly the depths or the distances between objects and the camera: small objects close to the camera will appear much larger, and big objects that are far away will appear much smaller than they actually are. This is why depth information is crucial in many tasks that rely on visual perception such as robot grasping, obstacle avoidance, and navigation, which are necessary tasks for smart homes and smart cities.
With the availability of affordable depth cameras such as Microsoft Kinect sensors, many systems are now being supplemented with depth range information to solve various computer vision tasks. However, the depth maps produced by these sensors have significantly lower resolution than the color images due to the intrinsic physical constraints of the sensors [1,2]. We desire accurate and high-resolution depth maps in order to perform many robotic tasks effectively, especially those involving higher risks, such as autonomous robots and self-driving cars, where small errors could lead to large costs.
A simple way of increasing the resolution of the depth map is to use the high-resolution color image. First, the depth map and its corresponding color image need to be aligned and registered since there would be a small offset coming from the side-by-side placements of the range and image sensors that needs to be corrected. Homographic warping and multi-view camera calibrations are sufficient to perform this task. Next, we can up-sample the depth maps and use the edges of the corresponding color image in order to refine the depth maps at the discontinuities where the values are likely to be uncertain. In many cases, this naive approach would work well. However, for objects with rich color textures, the edges of the textures will be treated as an object boundary, which would lead to unwanted texture-copying artifacts [3,4]. Another issue is when the object boundaries in the color image are not clear, which happens when the the object and its background have similar colors or low contrast.
In this paper, we propose a Generative Adversarial Network (GAN)-based framework that accepts multi-modal inputs, more specifically RGB color images and low-resolution depth maps, to generate higher resolution depth maps. GANs have been shown to model high frequency components well [5,6]. Since these high frequency components would correspond to the edges or object boundaries, it would be highly beneficial to incorporate them into the model. Moreover, as the discriminator in the GAN framework tries to separate real images from fake images, it would discourage the model from producing unrealistic values at the depth discontinuities as they would be considered as fake. While our model is trained on color images and depth maps, it only requires a single depth map input to perform super-resolution at test time. This is a highly desirable property since there are tasks, such as gesture recognition, action recognition, and event recognition, where the depth maps alone may be sufficient. This allows for faster models with a lesser memory footprint since we do not need to process the RGB image. There may also be applications with privacy restrictions, which would make it difficult to work with RGB images, as they are highly identifiable, unlike depth maps.

2. Related Work

Common approaches for depth map super-resolution can be grouped into two categories. One uses a single depth map as the input, and the other uses RGB images together with the depth maps as input. Single depth map up-sampling is desirable for applications where privacy is necessary or for applications where it is sufficient to use only depth maps. Aodha et al. [7] collected a set of training images containing large numbers of low-resolution patches and high-resolution patches. They then performed patch matching in order to synthesize the super-resolved depth map. Hornacek et al. [8] further extended this patch-based matching strategy by exploiting patch-wise self-similarity structures across depth resolutions. Li et al. [9], on the other hand, proposed to consider semantic object composition as an auxiliary regularization for assembling the high-resolution depth outputs. For these types of approaches, the collected training data have a high influence on the model’s performance.
Methods that use only single depth map inputs usually perform poorer than methods that use both RGB images and the depth maps as inputs. This is to be expected because of the added high-resolution information that they get from the RGB images. Algorithms with RGB-D inputs can be further divided into filtering-based methods and learning-based ones. Filtering-based methods are based on performing a weighted average of the neighboring pixels. The most common is the Bilateral Filter (BF) [10], which is an edge-preserving filter for image upsampling. Kopf et al. [11] proposed a Joint Bilateral Upsampling (JBU) framework that operates on both the high-resolution color image and the low-resolution depth map. They leveraged the color image to preserve the edges of the depth map. Chan et al. [12] and Li et al. [13] proposed an extended JBU method by providing a noise-aware characteristic, which can determine the effects of color information on the upsampled depth map. Yang et al. [14] proposed a Joint Bilateral Filtering (JBF) method, which is similar to JBU. Kim et al. [15] developed a JBF framework by integrating the numerical analysis of the local distribution in color images and depth maps. Liu et al. [16] also proposed an improved JBF method using geodesic distance for better-preserved depth edges. The above methods present some promising results, but color or lighting variations may produce incorrect discontinuity regions in the process of integrating color information into the upsampled depth map. Jung [17] proposed a filtering method that applies matching patterns between local color and depth patches to the filtering kernels. However, some ambiguous pattern assignments caused prediction errors in their results. Choi et al. [18] proposed another filtering method that suppresses texture-transfer and depth-bleeding artifacts. They first applied region classification to split the color image into different regions and processed each region differently. They used the Weighted Mode Filter (WMF) [19] as their main filter kernel. A problem with this approach is that depth edges at some regions are ambiguous due to color similarities. Lo et al. proposed a series of methods to improve this filtering method [4,20,21] and reported promising results.
Learning-based methods, on the other hand, rely on extracting information from a training dataset that can be generalized to different examples. Diebel and Thrun et al. [22] proposed a multi-labeling optimization problem based on Markov Random Fields (MRF), which defines a consistency term, which encourages consistency between depth values across resolutions, and a smoothness term, which encourages neighboring pixels with similar colors to have similar depth values. Revised MRF methods focusing on depth discontinuities were proposed in [3,23,24,25] to improve depth map super-resolution. Park et al. [26] improved the smoothness term by incorporating semi-local neighborhood information extracted from Non-Local Means (NLM) regularization and an edge weighting scheme, which enhances color details. Other learning-based methods like [27] phrased the depth map up-sampling task as a convex optimization problem with higher order regularization guided by anisotropic diffusion tensors extracted from high-resolution intensity images.
In this paper, we present a novel multi-modal GAN framework to perform the depth map up-sampling task. Our method combines the best of both worlds where we train on both the RGB color image and the depth maps, but only require a single depth map at testing time. We show in our experiments that this approach outperforms several state-of-the-art baselines.

3. Proposed Framework

Figure 1 shows an overview of our GAN-based depth map super-resolution framework. Our generator takes inputs from two modalities: low-resolution depth maps and low-resolution scene images. Since the scene images and depth maps are structurally very similar to each other, as shown in Figure 2, we can leverage the information from the scene images to preserve the discontinuity regions of the depth maps. In order to ensure that the generator outputs valid depth maps, we train it adversarially using a discriminator that tries to distinguish whether the output of the generator is similar to real depth maps sampled from the dataset or fake depth maps synthesized by the generator.
We first formally define the problem of depth map super-resolutions in Section 3.1. Next, we discuss the details of our model starting with generative adversarial networks in Section 3.2 followed by our loss functions (Section 3.3) and network architecture (Section 3.4). Lastly, we discuss our multi-modal mini-batch scheme in Section 3.5.

3.1. Problem Formulation

A standard camera and a range sensor can capture a high-resolution color image x i H R and a low-resolution depth map x d L R . Suppose that we have a dataset with the corresponding ground truth high-resolution depth maps x d H R ; our goal is to learn a function G : x d L R x d H R that can generate a high-resolution version of the low-resolution depth map. This is more commonly referred to as super-resolution. The function G is modeled as a convolutional neural network, which we refer to as the generator.

3.2. Generative Adversarial Network

The framework of Generative Adversarial Networks (GANs) introduces a discriminator D, which is a separate classifier, to guide the learning process of the generator G. This transforms the learning problem into a two-player minimax game, where the optimal solution is a Nash equilibrium.
The role of the discriminator is to learn how to tell apart real images from fake images. The generator G, on the other hand, is our super-resolution model, which generates high-resolution depth maps from its low-resolution input and tries to make it as realistic looking as possible in order to trick the discriminator D into classifying the generated depth map as real. This is represented as a min-max optimization in the form shown in Equation (1), where p ( x ) represents the distributions of the data and x H R and x L R are the high-resolution and low-resolution inputs, respectively. The first term accounts for the objective where we want the discriminator to classify the high-resolution inputs x H R coming from the dataset as real, while the second term accounts for the objective where we want to classify the super-resolved outputs of the generator ( G ( x L R ) ) from low-resolution inputs x L R as fake. In the formulation of these networks, the generator G has access to the gradients of the discriminator D and therefore has some form of instruction as to how to improve itself. This enables the generator to learn how to produce realistic-looking depth maps that are indistinguishable from the ground truth depth maps.
min G max D L GAN = E x H R p ( x ) [ log D ( x H R ) ] + E x L R p ( x ) [ log 1 D ( G ( x L R ) ) ]
In the beginning of the training process, the fake images generated by G are extremely poor and are rejected by D with high confidence. Therefore, when performing the optimization, it is better for G to optimize for log ( D ( G ( x L R ) ) ) instead of log ( 1 D ( G ( x L R ) ) ) . Both objectives result in the same fixed point, but log ( D ( G ( x L R ) ) ) provides stronger gradients in the early stages of learning.

3.3. Loss Function

Our loss function is a combination of two components. The first component is a content loss or a reconstruction error, which measures how different the generated depth maps are from the ground truth depth map. It is implemented as the mean absolute error or L 1 distance between the generated and the ground truth depth map, as shown in Equation (2). As shown in previous works [5,6], this term alone will produce blurry images as the solution would turn towards the mean. However, this shows that it can capture the low frequency components well.
L Content = E x H R , x L R p ( x ) x H R G ( x L R ) 1
The second component is an adversarial loss that we have defined in the previous Section 3.2. This term encourages the outputs of G to reside on the manifold of the ground truth depth maps. It is also able to model the high frequency components as shown by [5,6], which makes it suitable for our problem. The final objective function is shown in Equation (3), where λ is a hyper-parameter that controls the relative importance of the two components.
min G max D L = L Content + λ L GAN

3.4. Network Architecture

Figure 3 details the network architecture of our generator. Inspired by Ledig et al. [28] and Lim et al. [29], we use a single convolutional layer followed by a series of sixteen residual blocks. A skip-connection is then performed, where the output of the residual blocks is combined with the output of the initial convolutional layer using an element-wise sum. The idea behind this design is that the low-resolution depth map already contains many of the pixel values for the output, and adding the skip-connection allows the network to have easier passing of information from the lower layers, which are closer to the original input, to the upper layers, which are closer to the output. We then use a two sub-pixel convolutional layer as proposed by [30], which up-samples the depth maps to four times their input resolution. We use the ReLU activation function after every convolutional layer without any batch normalization layers. We remove the batch normalization layers since it has been shown in [29] that it is detrimental to super-resolution tasks since it reduces the range flexibility by normalizing the features.
As shown in Figure 4, the discriminator D is composed of a series of six convolutional layers where every other layer has a stride of two that downsamples the image representations by half. This is then followed by two fully-connected layers and a sigmoid activation function that outputs the probability of being real. All the layers use LeakyReLU activations, and batch normalization layers are inserted after every convolutional layer except for the first layer.

3.5. Multi-Modal Mini-Batch

We would like to incorporate the scene image during the learning process; however, adding an explicit branch with convolutional layers to encode and extract features from RGB scene image would require us to use the RGB image during test time. This is undesirable since our goal is to enable our super-resolution model to handle depth maps independent from the scene image at test time.
Inspired by multi-task learning and domain adaptation frameworks, we mix scene images with the depth images and perform super-resolution on both modalities during training. More specifically, we down-sample the scene image to have the same resolution as the depth map and convert it to gray scale. Now that they both have the same dimensions, we can combine them in a mini-batch and perform the training. This way, the network will be able to share what it learned from one domain (super-resolving scene images) and apply it to the other (super-resolving depth maps).

4. Experiments

4.1. Dataset

A standard dataset used widely in depth-related tasks is the NYU Depth dataset Version 2 [31]. The images and its corresponding depth maps were captured from 464 different locations coming from a variety of indoor scenes such as bathrooms, bedrooms, offices, and kitchens. We followed the same train-test experiment setup as previous studies [3,4,21]. We downsampled the ground truth depth maps from 640 × 480 to 160 × 120 (by a factor of four) in order to obtain the low-resolution depth maps to be used as the input by our model.
We also experimented on the Middlebury stereo dataset [32]. The ground truth depth maps were estimated using state-of-the-art stereo depth estimation methods that are able to handle high-resolution images. Similarly, we downsampled the depth maps by a factor of four to get our training inputs.

4.2. Implementation Details

We trained our model on a single NVIDIA GTX 1080 GPU with a mini-batch size of 16. During training, we randomly cropped 96 × 96 patches from the low-resolution depth maps, and our model up-sampled it by a factor of four producing a patch size of 384 × 384 . Note that since our model is composed of fully-convolutional layers, it can handle arbitrary sizes at test time. We usd Adam optimizer [33] with β 1 = 0.9 and β 2 = 0.999 to train our model. Our network was trained with an initial learning rate of 1 e 4 . We set λ = 1 e 3 in our loss function. We implemented our models using the TensorLayer and TensorFlow framework.

4.3. Performance Evaluations

Following previous works [3,4,21,32], we quantitatively evaluated our model using the percentage of Bad matching Pixels (BP%). Bad matching Pixels (BP) is defined as the percentage of pixels that have a difference larger than a pre-defined threshold from the ground truth values. This can be expressed mathematically as shown in Equation (4), where Ω refers to the set of all pixels in the depth map. To compute the percentage of bad pixels (BP%), the output depth maps have to be scaled to the same range. We used the same threshold ( δ d = 1 ) from previous works [3,4,21,32] to maintain a fair comparison. In addition to the whole image statistics (referenced as “all”), we also computed local statistics at non-occluded regions (“nonocc.”) and at depth discontinuity regions (“disc.”), as defined in [32,34].
BP % = 1 | Ω | Ω 𝟙 ( | x H R G ( x L R ) | > δ d ) × 100 %
We evaluated our model’s super-resolved depth maps at an upsampling factor of four. We compared against several state-of-the-art baselines: (1) Diebel et al. [22]; (2) JBU [11]; (3) Chan et al. [12]; (4) Lu et al. [23]; (5) Kim et al. [24]; (6) Jung [17]; (7) Park et al. [26]; (8) Ferstl et al. [27]; (9) Choi et al. [18]; (10) Lo et al. [3]; (11) Lo et al. [4]; and (12) Lo et al. [21]. We used the outputs from the respective authors’ publicly available code for the comparisons. However, there is no information on the parameter settings used in [12,22]; hence, we performed a simple grid search and reported the best performing results. Table 1 lists the performance of the various methods in terms of BP%. It can be observed that our method consistently performed better in terms of percentage of bad matching pixels.
We also evaluated our model using the Mean Squared Error (MSE) with respect to the ground truth depth maps. Our method was able to achieve state-of-the-art results, as shown in Table 2. MSE weights large differences more than small differences. We can observe that these large differences usually occurred at the depth discontinuity regions (“disc.”) since the local MSE values were significantly higher than the global MSE (“all”). By visually inspecting the super-resolved depth maps, we can further evaluate the performances of the various models. Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 show comparisons of super-resolved depth maps on several examples of patches at regions with depth discontinuity. We can see that most methods failed to capture the geometry at the edges properly. This is more noticeable when we zoom in to the patches where the straight lines were no longer straight for other methods, while our method successfully preserved these structures.

5. Conclusions

We proposed a multi-modal model based on the generative adversarial network for depth map super-resolution. Our method was able to preserve detailed discontinuity regions such as sharp edges, as well as respect the geometric structure of the objects. Moreover, our model only required a single depth map as an input during test time, even though it was trained with both RGB and depth maps. The results of our quantitative and qualitative experiment show that our method outperforms state-of-the-art depth map super-resolution methods, which confirms the effectiveness of our method.

Author Contributions

Conceptualization, D.S.T.; methodology, D.S.T. and J.-M.L.; software, D.S.T. and J.-M.L.; validation, D.S.T., J.-M.L., and K.-L.H.; formal analysis, D.S.T., J.-M.L., and K.-L.H.; investigation, J.-M.L.; resources, K.-L.H.; data curation, J.-M.L.; writing, original draft preparation, D.S.T. and K.-L.H.; writing, review and editing, D.S.T., Y.-C.L., J.I., and K.-L.H.; visualization, D.S.T.; supervision, Y.-C.L., J.-M.L., J.I., and K.-L.H.; project administration, K.-L.H.; funding acquisition, K.-L.H.

Funding

This work was financially supported by the Center for Cyber-physical System Innovation and Center of Intelligent Robots from The Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education (MOE) in Taiwan and the Ministry of Science and Technology of Taiwan under Grants MOST107-2218-E-011-014, MOST106-2221-E-011-154-MY2, and MOST106-2218-E-011-009-MY2.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RGBRed, Green, and Blue
GANGenerative Adversarial Network
LRLow-Resolution
HRHigh-Resolution
SRSuper-Resolution
BFBilateral Filter
JBUJoint Bilateral Upsampling
JBFJoint Bilateral Filtering
WMFWeighted Mode Filter
MRFMarkov Random Field
NLMNon-Local Means
BPBad Pixel
MSEMean Squared Error

References

  1. Schuon, S.; Theobalt, C.; Davis, J.; Thrun, S. High-quality scanning using time-of-flight depth superresolution. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW’08), Anchorage, AK, USA, 23–28 June 2008; pp. 1–7. [Google Scholar]
  2. Guomundsson, S.A.; Aanaes, H.; Larsen, R. Environmental effects on measurement uncertainties of Time-of-Flight cameras. In Proceedings of the International Symposium on Signals, Circuits and Systems (ISSCS), Iasi, Romania, 12–13 July 2007; Volume 1, pp. 1–4. [Google Scholar]
  3. Lo, K.H.; Hua, K.L.; Wang, Y.C.F. Depth map super-resolution via Markov random fields without texture-copying artifacts. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 1414–1418. [Google Scholar]
  4. Lo, K.H.; Wang, Y.C.F.; Hua, K.L. Joint trilateral filtering for depth map super-resolution. In Proceedings of the Visual Communications and Image Processing (VCIP), Kuching, Malaysia, 17–20 November 2013; pp. 1–6. [Google Scholar]
  5. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
  6. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar]
  7. Aodha, O.M.; Campbell, N.D.; Nair, A.; Brostow, G.J. Patch based synthesis for single depth image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; pp. 71–84. [Google Scholar]
  8. Hornacek, M.; Rhemann, C.; Gelautz, M.; Rother, C. Depth super-resolution by rigid body self-similarity in 3d. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 1123–1130. [Google Scholar]
  9. Li, J.; Lu, Z.; Zeng, G.; Gan, R.; Zha, H. Similarity-aware patchwork assembly for depth image super-resolution. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 24–27 June 2014; pp. 3374–3381. [Google Scholar]
  10. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Bombay, India, 4–7 January 1998; pp. 839–846. [Google Scholar]
  11. Kopf, J.; Cohen, M.; Lischinski, D.; Uyttendaele, M. Joint bilateral upsampling. ACM Trans. Graph. 2007, 26, 96–101. [Google Scholar] [CrossRef]
  12. Chan, D.; Buisman, H.; Theobalt, C.; Thrun, S. A noise-aware filter for real-time depth upsampling. In Proceedings of the ECCV Workshop on Multi-camera and Multi-modal Sensor Fusion Algorithms and Applications, Marseille, France, 18 October 2008. [Google Scholar]
  13. Li, Y.; Zhang, L.; Zhang, Y.; Xuan, H.; Dai, Q. Depth map super-resolution via iterative joint-trilateral-upsampling. In Proceedings of the IEEE Visual Communications and Image Processing (VCIP), Valletta, Malta, 7–10 December 2014; pp. 386–389. [Google Scholar]
  14. Yang, Q.; Yang, R.; Davis, J.; Nister, D. Spatial-depth super-resolution for range images. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Minneapolis, MN, USA, 18–23 June 2007; pp. 1–8. [Google Scholar]
  15. Kim, J.; Lee, J.; Han, S.; Kim, D.; Min, J.; Kim, C. Trilateral filter construction for depth map upsampling. In Proceedings of the IEEE IVMSP Workshop, Seoul, Korea, 10–12 June 2013; pp. 1–4. [Google Scholar]
  16. Liu, M.Y.; Tuzel, O.; Taguchi, Y. Joint geodesic upsampling of depth images. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 169–176. [Google Scholar]
  17. Jung, S.W. Enhancement of Image and Depth Map Using Adaptive Joint Trilateral Filter. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 258–269. [Google Scholar] [CrossRef]
  18. Choi, O.; Jung, S.W. A Consensus-Driven Approach for Structure and Texture Aware Depth Map Upsampling. IEEE Trans. Image Process. 2014, 23, 3321–3335. [Google Scholar] [CrossRef] [PubMed]
  19. Min, D.; Lu, J.; Do, M.N. Depth Video Enhancement Based on Weighted Mode Filtering. IEEE Trans. Image Process. 2012, 21, 1176–1190. [Google Scholar] [PubMed] [Green Version]
  20. Hua, K.L.; Lo, K.H.; Wang, Y.C.F. Extended Guided Filtering for Depth Map Upsampling. IEEE MultiMedia Mag. 2016, 23, 72–83. [Google Scholar] [CrossRef]
  21. Lo, K.H.; Wang, Y.C.F.; Hua, K.L. Edge-preserving depth map upsampling by joint trilateral filter. IEEE Trans. Cybern. 2018, 48, 371–384. [Google Scholar] [CrossRef] [PubMed]
  22. Diebel, J.; Thrun, S. An application of Markov random fields to range sensing. In Proceedings of the MIT Press Conference on Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 5–8 December 2005; pp. 291–298.
  23. Lu, J.; Min, D.; Pahwa, R.S.; Do, M.N. A revisit to MRF-based depth map super-resolution and enhancement. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 985–988. [Google Scholar]
  24. Kim, D.; Yoon, K. High quality depth map up-sampling robust to edge noise of range sensors. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 553–556. [Google Scholar]
  25. Lu, J.; Forsyth, D. Sparse Depth super-resolution. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 2015; pp. 2245–2253. [Google Scholar]
  26. Park, J.; Kim, H.; Tai, Y.W.; Brown, M.; Kweon, I. High quality depth map upsampling for 3d-tof cameras. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 1623–1630. [Google Scholar]
  27. Ferstl, D.; Reinbacher, C.; Ranftl, R.; Ruether, M.; Bischof, H. Image guided depth upsampling using anisotropic total generalized variation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013; pp. 993–1000. [Google Scholar]
  28. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; Shi, W. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  29. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  31. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from RGBD images. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012; pp. 746–760. [Google Scholar]
  32. Scharstein, D.; Szeliski, R. A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. Int. J. Comput. Vis. 2001, 47, 7–42. [Google Scholar] [CrossRef]
  33. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  34. Middlebury Stereo. Available online: http://vision.middlebury.edu/stereo/ (accessed on 4 January 2019).
Figure 1. Overview of our proposed framework for depth map super-resolution.
Figure 1. Overview of our proposed framework for depth map super-resolution.
Sensors 19 01587 g001
Figure 2. Observation of the similarities of discontinuity regions between a depth image and the corresponding scene image.
Figure 2. Observation of the similarities of discontinuity regions between a depth image and the corresponding scene image.
Sensors 19 01587 g002
Figure 3. Generator network of the proposed generative adversarial network framework for depth map super-resolution with f: the number of feature maps, k: the kernel size, and s: the stride indicated for each convolutional layer.
Figure 3. Generator network of the proposed generative adversarial network framework for depth map super-resolution with f: the number of feature maps, k: the kernel size, and s: the stride indicated for each convolutional layer.
Sensors 19 01587 g003
Figure 4. Discriminator network of the proposed generative adversarial network framework for depth map super-resolution with f: the number of feature maps, k: the kernel size, and s: the stride indicated for each convolutional layer.
Figure 4. Discriminator network of the proposed generative adversarial network framework for depth map super-resolution with f: the number of feature maps, k: the kernel size, and s: the stride indicated for each convolutional layer.
Sensors 19 01587 g004
Figure 5. Example depth map SR results of Venus with two Regions Of (nterest (ROI). (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 1.16), (d) JBU [11] (0.72), (e) Lu et al. [23] (0.31), (f) Kim et al. [24] (0.30), (g) Jung [17] (0.59), (h) Lo et al. [3] (0.16), (i) Park et al. [26] (0.66), (j) Ferstl et al. [27] (0.41), (k) Choi et al. [18] (0.49), (l) Lo et al. [21] (0.16), (m) Our method (0.04), and (n) the associated ground truth.
Figure 5. Example depth map SR results of Venus with two Regions Of (nterest (ROI). (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 1.16), (d) JBU [11] (0.72), (e) Lu et al. [23] (0.31), (f) Kim et al. [24] (0.30), (g) Jung [17] (0.59), (h) Lo et al. [3] (0.16), (i) Park et al. [26] (0.66), (j) Ferstl et al. [27] (0.41), (k) Choi et al. [18] (0.49), (l) Lo et al. [21] (0.16), (m) Our method (0.04), and (n) the associated ground truth.
Sensors 19 01587 g005
Figure 6. Example depth map SR results of Dolls with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 6.64), (d) JBU [11] (5.55), (e) Lu et al. [23] (6.69), (f) Kim et al. [24] (4.37), (g) Jung [17] (5.04), (h) Lo et al. [3] (3.45), (i) Park et al. [26] (5.07), (j) Ferstl et al. [27] (2.80), (k) Choi et al. [18] (3.17), (l) Lo et al. [21] (1.78), (m) Our method (1.74), and (n) the associated ground truth.
Figure 6. Example depth map SR results of Dolls with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 6.64), (d) JBU [11] (5.55), (e) Lu et al. [23] (6.69), (f) Kim et al. [24] (4.37), (g) Jung [17] (5.04), (h) Lo et al. [3] (3.45), (i) Park et al. [26] (5.07), (j) Ferstl et al. [27] (2.80), (k) Choi et al. [18] (3.17), (l) Lo et al. [21] (1.78), (m) Our method (1.74), and (n) the associated ground truth.
Sensors 19 01587 g006
Figure 7. Example depth map SR results of Moebius with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 9.58), (d) JBU [11] (6.48), (e) Lu et al. [23] (6.36), (f) Kim et al. [24] (5.46), (g) Jung [17] (6.03), (h) Lo et al. [3] (4.33), (i) Park et al. [26] (6.49), (j) Ferstl et al. [27] (3.53), (k) Choi et al. [18] (4.97), (l) Lo et al. [21] (3.04), (m) Our method (1.53), and (n) the associated ground truth.
Figure 7. Example depth map SR results of Moebius with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 9.58), (d) JBU [11] (6.48), (e) Lu et al. [23] (6.36), (f) Kim et al. [24] (5.46), (g) Jung [17] (6.03), (h) Lo et al. [3] (4.33), (i) Park et al. [26] (6.49), (j) Ferstl et al. [27] (3.53), (k) Choi et al. [18] (4.97), (l) Lo et al. [21] (3.04), (m) Our method (1.53), and (n) the associated ground truth.
Sensors 19 01587 g007
Figure 8. Example depth map SR results of Reindeer with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.95), (d) JBU [11] (3.36), (e) Lu et al. [23] (3.48), (f) Kim et al. [24] (2.14), (g) Jung [17] (2.91), (h) Lo et al. [3] (1.97), (i) Park et al. [26] (3.86), (j) Ferstl et al. [27] (1.93), (k) Choi et al. [18] (2.42), (l) Lo et al. [21] (1.42), (m) Our method (0.74), and (n) the associated ground truth.
Figure 8. Example depth map SR results of Reindeer with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.95), (d) JBU [11] (3.36), (e) Lu et al. [23] (3.48), (f) Kim et al. [24] (2.14), (g) Jung [17] (2.91), (h) Lo et al. [3] (1.97), (i) Park et al. [26] (3.86), (j) Ferstl et al. [27] (1.93), (k) Choi et al. [18] (2.42), (l) Lo et al. [21] (1.42), (m) Our method (0.74), and (n) the associated ground truth.
Sensors 19 01587 g008
Figure 9. Example depth map SR results of Teddy with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 8.17), (d) JBU [11] (6.61), (e) Lu et al. [23] (5.60), (f) Kim et al. [24] (6.20), (g) Jung [17] (5.65), (h) Lo et al. [3] (3.69), (i) Park et al. [26] (5.24), (j) Ferstl et al. [27] (4.32), (k) Choi et al. [18] (3.88), (l) Lo et al. [21] (3.20), (m) Our method (2.76), and (n) the associated ground truth.
Figure 9. Example depth map SR results of Teddy with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 8.17), (d) JBU [11] (6.61), (e) Lu et al. [23] (5.60), (f) Kim et al. [24] (6.20), (g) Jung [17] (5.65), (h) Lo et al. [3] (3.69), (i) Park et al. [26] (5.24), (j) Ferstl et al. [27] (4.32), (k) Choi et al. [18] (3.88), (l) Lo et al. [21] (3.20), (m) Our method (2.76), and (n) the associated ground truth.
Sensors 19 01587 g009
Figure 10. Example depth map SR results of Cones with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 8.10), (d) JBU [11] (7.54), (e) Lu et al. [23] (4.51), (f) Kim et al. [24] (5.33), (g) Jung [17] (5.93), (h) Lo et al. [3] (3.82), (i) Park et al. [26] (5.73), (j) Ferstl et al. [27] (3.44), (k) Choi et al. [18] (5.52), (l) Lo et al. [21] (3.19), (m) Our method (2.58), and (n) the associated ground truth.
Figure 10. Example depth map SR results of Cones with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 8.10), (d) JBU [11] (7.54), (e) Lu et al. [23] (4.51), (f) Kim et al. [24] (5.33), (g) Jung [17] (5.93), (h) Lo et al. [3] (3.82), (i) Park et al. [26] (5.73), (j) Ferstl et al. [27] (3.44), (k) Choi et al. [18] (5.52), (l) Lo et al. [21] (3.19), (m) Our method (2.58), and (n) the associated ground truth.
Sensors 19 01587 g010
Figure 11. Example depth map SR results of Midd with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 2.96), (d) JBU [11] (1.95), (e) Lu et al. [23] (3.00), (f) Kim et al. [24] (2.92), (g) Jung [17] (2.10), (h) Lo et al. [3] (1.80), (i) Park et al. [26] (1.68), (j) Ferstl et al. [27] (1.40), (k) Choi et al. [18] (1.75), (l) Lo et al. [21] (0.72), (m) Our method (0.52), and (n) the associated ground truth.
Figure 11. Example depth map SR results of Midd with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 2.96), (d) JBU [11] (1.95), (e) Lu et al. [23] (3.00), (f) Kim et al. [24] (2.92), (g) Jung [17] (2.10), (h) Lo et al. [3] (1.80), (i) Park et al. [26] (1.68), (j) Ferstl et al. [27] (1.40), (k) Choi et al. [18] (1.75), (l) Lo et al. [21] (0.72), (m) Our method (0.52), and (n) the associated ground truth.
Sensors 19 01587 g011
Figure 12. Example depth map SR results of Kitchen1 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.45), (d) JBU [11] (2.30), (e) Lu et al. [23] (2.32), (f) Kim et al. [24] (2.24), (g) Jung [17] (2.00), (h) Lo et al. [3] (1.45), (i) Park et al. [26] (2.21), (j) Ferstl et al. [27] (1.76), (k) Choi et al. [18] (1.86), (l) Lo et al. [21] (1.16), (m) Our method (0.48), and (n) the associated ground truth.
Figure 12. Example depth map SR results of Kitchen1 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.45), (d) JBU [11] (2.30), (e) Lu et al. [23] (2.32), (f) Kim et al. [24] (2.24), (g) Jung [17] (2.00), (h) Lo et al. [3] (1.45), (i) Park et al. [26] (2.21), (j) Ferstl et al. [27] (1.76), (k) Choi et al. [18] (1.86), (l) Lo et al. [21] (1.16), (m) Our method (0.48), and (n) the associated ground truth.
Sensors 19 01587 g012
Figure 13. Example depth map SR results of Kitchen2 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 6.04), (d) JBU [11] (3.21), (e) Lu et al. [23] (3.01), (f) Kim et al. [24] (3.23), (g) Jung [17] (2.71), (h) Lo et al. [3] (2.96), (i) Park et al. [26] (3.64), (j) Ferstl et al. [27] (2.78), (k) Choi et al. [18] (2.86), (l) Lo et al. [21] (1.73), (m) Our method (0.74), and (n) the associated ground truth.
Figure 13. Example depth map SR results of Kitchen2 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 6.04), (d) JBU [11] (3.21), (e) Lu et al. [23] (3.01), (f) Kim et al. [24] (3.23), (g) Jung [17] (2.71), (h) Lo et al. [3] (2.96), (i) Park et al. [26] (3.64), (j) Ferstl et al. [27] (2.78), (k) Choi et al. [18] (2.86), (l) Lo et al. [21] (1.73), (m) Our method (0.74), and (n) the associated ground truth.
Sensors 19 01587 g013
Figure 14. Example depth map SR results of Store1 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.09), (d) JBU [11] (3.60), (e) Lu et al. [23] (2.56), (f) Kim et al. [24] (2.46), (g) Jung [17] (2.86), (h) Lo et al. [3] (2.39), (i) Park et al. [26] (3.59), (j) Ferstl et al. [27] (2.24), (k) Choi et al. [18] (2.78), (l) Lo et al. [21] (2.15), (m) Our method (0.58), and (n) the associated ground truth.
Figure 14. Example depth map SR results of Store1 with two ROI. (a) Ground truth depth map, (b) color regions, and SR results of (c) Diebel et al. [22] (BP% = 3.09), (d) JBU [11] (3.60), (e) Lu et al. [23] (2.56), (f) Kim et al. [24] (2.46), (g) Jung [17] (2.86), (h) Lo et al. [3] (2.39), (i) Park et al. [26] (3.59), (j) Ferstl et al. [27] (2.24), (k) Choi et al. [18] (2.78), (l) Lo et al. [21] (2.15), (m) Our method (0.58), and (n) the associated ground truth.
Sensors 19 01587 g014
Table 1. Comparisons of Bad Pixel percentages (BP%) with upsampling factors of 4. For each image, the ranking of our approach is shown in parenthesis, and the best performance is shown in bold. nonocc., non-occluded; disc., discontinuity.
Table 1. Comparisons of Bad Pixel percentages (BP%) with upsampling factors of 4. For each image, the ranking of our approach is shown in parenthesis, and the best performance is shown in bold. nonocc., non-occluded; disc., discontinuity.
BP% of 4× SRVenusTeddyConesDollsMidd2MoebiusReindeerKitchen1Kitchen2Store1
nonocc.alldisc.nonocc.alldisc.nonocc.alldisc.all
Diebel et al. [22]0.851.163.937.468.1718.026.988.1019.826.642.969.583.953.456.043.09
JBU [11]0.400.725.625.756.6120.086.437.5419.185.551.956.483.362.303.213.60
Chan et al. [12]0.410.695.745.646.4819.756.247.3118.784.961.875.533.301.962.512.84
Lu et al. [23]0.240.313.275.145.6014.473.734.5110.076.693.006.363.482.323.012.56
Kim et al. [24]0.170.302.315.336.2016.864.865.3314.624.372.925.462.142.243.232.46
Jung [17]0.270.593.764.775.6515.994.835.9314.585.042.106.032.912.002.712.86
Lo et al. [3]0.120.161.673.353.699.593.153.829.433.451.804.331.971.452.962.39
Lo et al. [4]0.090.131.243.934.3212.813.543.9210.622.120.753.031.811.321.822.17
Park et al. [26]0.380.665.154.295.2413.884.075.7312.275.071.686.493.862.213.643.59
Ferstl et al. [27]0.270.413.773.754.3210.032.473.448.172.801.403.531.931.762.782.24
Choi et al. [18]0.280.493.893.233.8811.414.835.5214.653.171.754.972.421.862.862.78
Lo et al. [21]0.080.161.102.733.208.522.673.198.091.780.723.041.421.161.732.15
Our method0.04 (1)0.07 (1)0.57 (1)2.13 (1)2.76 (1)6.54 (1)2.16 (1)2.58 (1)6.54 (1)1.74 (1)0.52 (1)1.53 (1)0.74 (1)0.48 (1)0.74 (1)0.58 (1)
Table 2. Comparisons of MSE with upsampling factors of 4. For each image, the ranking of our approach is shown in the parenthesis, and the best performance is shown in bold.
Table 2. Comparisons of MSE with upsampling factors of 4. For each image, the ranking of our approach is shown in the parenthesis, and the best performance is shown in bold.
MSE of 4× SRVenusTeddyConesDollsMidd2MoebiusReindeerKitchen1Kitchen2Store1
nonocc.alldisc.nonocc.alldisc.nonocc.alldisc.all
Diebel et al. [22]9.5512.6949.9840.6351.95135.4260.8866.38180.3714.7042.1927.7240.9523.3744.3928.46
JBU [11]2.905.0436.2761.0372.47224.3680.1284.11241.5017.1825.3123.9951.367.8920.3810.39
Chan et al. [12]2.774.7936.4160.7872.17223.8779.7983.71241.0616.3524.3222.5850.627.3518.619.37
Lu et al. [23]6.3425.4339.5416.9720.9844.5656.1660.36161.6614.0235.7321.8854.2513.7636.5115.59
Kim et al. [24]2.833.5919.2440.2251.04141.7561.3063.01182.7713.2731.0814.8429.9411.5138.4515.06
Jung [17]2.424.7230.0849.7460.84179.5269.1575.27208.7814.7627.6221.9251.187.5320.068.03
Lo et al. [3]2.162.5418.6726.0732.4080.0267.8170.98198.3911.5224.4917.4639.588.6731.8411.69
Lo et al. [4]1.401.7417.4418.1122.5763.8259.1659.54178.977.3415.0310.8680.0810.2417.8019.01
Park et al. [26]2.894.9635.0849.4657.58111.6549.1283.94141.6011.3220.8316.0364.979.8817.8814.21
Ferstl et al. [27]2.013.2129.4939.8850.98109.6042.6258.1597.529.0219.6710.3444.875.9518.1812.94
Choi et al. [18]2.383.4430.6517.3422.5663.2439.1741.94118.488.7218.0614.0343.755.558.426.64
Lo et al. [21]1.332.1316.5315.1919.1651.5641.2442.84124.635.8414.0011.2524.016.5717.728.95
Our method0.30 (1)0.43 (1)2.82 (1)17.03 (3)22.47 (3)62.75 (3)21.30 (1)24.98 (1)64.19 (1)7.97 (3)9.66 (1)8.93 (1)11.50 (1)1.27 (1)1.60 (1)1.40 (1)

Share and Cite

MDPI and ACS Style

Tan, D.S.; Lin, J.-M.; Lai, Y.-C.; Ilao, J.; Hua, K.-L. Depth Map Upsampling via Multi-Modal Generative Adversarial Network. Sensors 2019, 19, 1587. https://doi.org/10.3390/s19071587

AMA Style

Tan DS, Lin J-M, Lai Y-C, Ilao J, Hua K-L. Depth Map Upsampling via Multi-Modal Generative Adversarial Network. Sensors. 2019; 19(7):1587. https://doi.org/10.3390/s19071587

Chicago/Turabian Style

Tan, Daniel Stanley, Jun-Ming Lin, Yu-Chi Lai, Joel Ilao, and Kai-Lung Hua. 2019. "Depth Map Upsampling via Multi-Modal Generative Adversarial Network" Sensors 19, no. 7: 1587. https://doi.org/10.3390/s19071587

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop