Next Article in Journal
Direct Electron Transfer of Dehydrogenases for Development of 3rd Generation Biosensors and Enzymatic Fuel Cells
Next Article in Special Issue
Geometric Integration of Hybrid Correspondences for RGB-D Unidirectional Tracking
Previous Article in Journal
Multi-Layer Artificial Neural Networks Based MPPT-Pitch Angle Control of a Tidal Stream Generator
Previous Article in Special Issue
In Situ 3D Monitoring of Geometric Signatures in the Powder-Bed-Fusion Additive Manufacturing Process via Vision Sensing Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Depth Reconstruction from Single Images Using a Convolutional Neural Network and a Condition Random Field Model

1
Faculty of Geomatics, East China University of Technology, Nanchang 330013, China
2
Key Laboratory of Virtual Geographic Environment, Nanjing Normal University, Nanjing 210023, China
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(5), 1318; https://doi.org/10.3390/s18051318
Submission received: 30 March 2018 / Revised: 19 April 2018 / Accepted: 20 April 2018 / Published: 24 April 2018
(This article belongs to the Special Issue Depth Sensors and 3D Vision)

Abstract

:
This paper presents an effective approach for depth reconstruction from a single image through the incorporation of semantic information and local details from the image. A unified framework for depth acquisition is constructed by joining a deep Convolutional Neural Network (CNN) and a continuous pairwise Conditional Random Field (CRF) model. Semantic information and relative depth trends of local regions inside the image are integrated into the framework. A deep CNN network is firstly used to automatically learn a hierarchical feature representation of the image. To get more local details in the image, the relative depth trends of local regions are incorporated into the network. Combined with semantic information of the image, a continuous pairwise CRF is then established and is used as the loss function of the unified model. Experiments on real scenes demonstrate that the proposed approach is effective and that the approach obtains satisfactory results.

1. Introduction

Measuring the depth of a scene is a significant topic of research in photogrammetry and computer vision, which plays an essential role in various applications in 3D reconstruction, video surveillance, and robotics, etc. Much prior work has performed depth acquisition from multiple images taken in accordance with certain requirements [1,2,3], but in fact, many photos may well be not taken by photogrammetric purposes, but rather taken by the public or amateur photographers. Scene structure cannot be correctly recovered through the traditional photogrammetry, due to lack of corresponding features, or too big or small a baseline between these images. Moreover, there usually exists only a single image of a scene, such as historic photos and images from the Internet. Therefore, depth reconstruction from a single image is a basic task with important research value in photogrammetry and computer vision.
The task is an ill-posed and inherently ambiguous problem, as one given image may correspond to an infinite number of possible real world scenes [4]. Therefore, depth acquisition from a single image it is a still challenging issue. Some previous works solve this problem using some depth cues such as geometric characteristics [5,6,7,8], shading [9], texture [10] and contour [11]. However, these works only infere the relative depth of the scene from an image but can’t get the absolute depth. In recent years, many researchers have applied machine learning to the problem and obtained some good results [12,13,14,15,16,17]. A common characteristic of these methods is that they rely on hand-crafted features. Saxena et al. [12] extracted three local features from images: haze, texture variations and texture gradient. Shape- and location-based features were added in [13] for better feature representation. However, these low-level features are still not enough to predict the exact depth values of pixels in an image. Based on Saxena et al. [13], Liu et al. [14] used semantic labels to guide depth reconstruction from a single image. Another challenging problem of these methods is how to utilize extracted image features to measure the depth of each pixel in the image. Many of these methods use a Markov Random Field (MRF) to build the relationships between image features and depth. Unfortunately, it is sensitive to multicolored objects in the image, and involves many assumptions to make the decision.
Recently, the Convolutional Neural Network (CNN) method has become a mainstream of image processing research. Compared with those traditional methods applied to depth reconstruction, CNN can learn a high-level of representation automatically without any manual interventions. Eigen et al. [18] used a multi-scale deep network to estimate depth maps from a single image. To perform pixel-level depth inference, Hu et al. [19] trained a CNN with raw RGB image patches cropped by a large window centered at each pixel. Li et al. [20] presented a framework for depth estimation from a single image, which consists of depth regression on superpixels via a deep CNN model and refining from superpixels to pixels via a hierarchical Conditional Random Field (CRF). Similarly, Wang et al. [21] performed depth prediction via regression on CNN model, combined with a post-processing refining step with a hierarchical CRF, but they joined depth and semantic inference, considering that the two problems are mutually beneficial. Unlike the above methods, Liu et al. [22,23], Xu et al. [24] formulated depth prediction as a continuous CRF learning problem, and used a CNN model to learn the feature representation of the image. The approach combined the strength of the CNN and CRF in a unified framework. However, they ignored the importance of semantic information to depth reconstruction and did not resolve depth ambiguities of a scene.
In this paper, a unified CNN framework is presented for depth reconstruction from a single image, joining a CNN and a continuous pairwise CRF model. A deep CNN network is firstly designed to automatically learn a hierarchical feature representation of the image. To get local details of the image, relative depth trends of local regions inside the image are integrated into the CNN network. Then, a continuous pairwise CRF is established as the loss function of the unified model through semantic information of a scene and the results of the CNN network in the first step. Depth reconstruction is formulated as a CRF learning problem and can be solved by maximum a posteriori (MAP) inference.

2. Methods

The approach performs pixel-level depth reconstruction from a single image in a unified CNN model framework, shown in Figure 1. The unified model joins a CNN and a continuous pairwise CRF, in which the continuous pairwise CRF is used as the loss function of the CNN. The model architecture consists of three parts: a unitary part, a pairwise part and a CRF loss layer. (1) In the unitary part, a convolutional network is used to obtain convolutional feature maps from the input image. To get feature maps of the superpixels, the convolutional feature maps are fed into a superpixel pooling layer along with the superpixels inside the image. These feature maps are then followed by three fully-connected layers. (2) In the pairwise part, sematic information and similarities of neighboring superpixels are considered and are fed into one fully-connected layer to produce the output. (3) In the loss layer, a continuous pairwise CRF is used as the loss function of the unified CNN framework, which is established via the outputs of the unary and pairwise part.

2.1. Depth Reconstruction Using CRF Model

Given an image x = { x 1 , x 2 , , x n } with corresponding to depth labels y = { y 1 , y 2 , , y n } , where n indexes superpixels via over-segmentation, the pairwise CRF are modeled as:
P ( y   |   x ; ϑ ) = 1 Z ( x , ϑ ) e x p { E ( y , x ; ϑ ) }
where ϑ are model parameters and Z ( x , ϑ ) = y e x p { E ( y , x ; ϑ ) }   d y the normalization term. The energy E ( y , x ; ϑ ) over superpixels N and edges S takes the following form:
E ( y , x ; ϑ ) = p N φ   ( y p , x ; ϑ ) + ( p , q ) S ϕ   ( y p , y q , x ; ϑ )
where φ   ( y p , x ; ϑ ) and ϕ   ( y p , y q , x ; ϑ ) represent the unary and pairwise potentials respectively.
Once the parameters ϑ are learned, depth map of an image can be predicted by MAP inference, written as:
y * = arg max y P ( y   |   x ; ϑ )

2.2. Unitary Part

The unitary part to obtain depth regression of each superpixel in the image uses a deep CNN model for learning feature representation of all the superpixels. The unitary potential φ   ( y p , x ; ϑ ) of the CNN model is defined as a Euclidean loss associated with the ground-truth depth value y p , p = 1 , 2 , , n and the predication z p :
φ   ( y p , x ; θ ) = ( y p z p ( θ ) ) 2
Usually, the depth of a superpixel is calculated with a single value. However, it is too coarse since depth values of different pixels inside the superpixel may be different. Fortunately, there are many local regions with similar structure from a sematic class, which means that their relative depth trends are nearly same, shown in Figure 2. Therefore, the relative depth trends from the same semantic class can be expressed with a limited normalized depth map called a depth template. A normalized depth map of a superpixel is calculated by the depth value of superpixel centers and scale factors. Given the normalized depth map t p , the depth value at the superpixel center c p and the scale factor s p , the depth map of the superpixel can be defined as: z p = s p t p + c p .
To obtain depth templates for each semantic label, the normalized depth maps of all the superpixels with the same sematic label are clustered. In this paper, relative depth trends of the superpixels, which are represented by the depth templates, are incorporated into the CNN network. To obtain the absolute depth values of each pixel inside a superpixel, the outputs of the CNN network for the unary part are designed as the depth value at the superpixel center and its normalized scale factor. The structure of the CNN model is similar to that described by Liu et al. [23], but their outputs are different because this paper joins the relative depth trends of the superpixels.

2.3. Pairwise Part

The pairwise part considers the depth relationships between neighboring superpixels, combined with their similarity and semantic information. The pairwise potential of the CRF model is constructed as:
ϕ   ( y p , y q , x ; β , w ) = 1 2 R p q ( y p y q ) 2 + 1 2 w   ( l p , l q )   ( y p y q ) 2
where, β , w are parameters. The first term 1 2 R p q ( y p y q ) 2 represents the consistency information of the neighboring superpixels p , q with their similarity matrix S p q . S p q is established with color in LUV space, color histogram and texture of Local Binary Pattern. R p q is produced by one fully-connected layer with S p q , defined as:
R p q = β T [ S p q ( 1 ) , S p q ( 2 ) , , S p q ( K ) ] = k = 1 K β k S p q ( k ) ( K = 3 )
The second term 1 2 w ( l p , l q )   ( y p y q ) 2 in Equation (6) represents the depth smoothness of the neighboring superpixels p , q with their semantic labels. Here l p , l q are respectively the sematic labels of p , q and w ( l p , l q ) represents the semantic weight between them. The higher the weight value is, the smoother the depth between the neighboring superpixels is. A weight matrix w is formed with all the sematic weights. w is a C × C matrix, where C is the number of the sematic labels in the scene. In the weight matrix, w ( l p , l q ) ( l p = 1 , , C ; l q = 1 , , C ) represents the semantic weight of the semantic labels l p , l q , and w ( l p , l q ) = w ( l q , l p ) .

2.4. CRF Loss Layer

The loss function of the depth reconstruction model uses the negative log-loss of the pairwise CRF, shown in Equation (1). According to Equations (4) and (5), the potential of the CRF can be expressed as:
E ( y , x ; ϑ ) = p N ( y p z p ( θ ) ) 2   + 1 2 ( p , q ) S R p q   ( y p y q ) 2 + 1 2 ( p , q ) S w ( l p , l q )   ( y p y q ) 2
Then Equation (1) can be written as:
L o s s = log P ( y   | x   ; ϑ ) = log ( 1 Z ( x , ϑ ) e x p { E ( y , x ; ϑ ) } )
Here ϑ = { θ , β , w } are parameters that can be learned by minimizing Equation (8).

3. Results

The proposed method is evaluated on the Make3D dataset [12]. The Make3D dataset contains 534 images of outdoor scenes composed of eight semantic classes including sky, tree, road, water, grass, building, mountain and foreground objects. The method is quantitatively evaluated by several common measures used in prior work [20,23]:
(1)
mean relative error (Rel): 1 T i T | d i d i * | d i
(2)
root mean squared error (Rmse): 1 T i T ( d i d i * ) 2
(3)
mean log10 error (Log10): 1 T i T | log 10 ( d i ) log 10 ( d i * ) |
where d i * is predicted depth at pixel i , di is the corresponding ground-truth depth, and T is the number of pixels in the image.
As pointed out in [17], the range of pixels in Make3D is limited to a depth range of 0~81 m, due to the limited range and the resolution of the sensor. As done in [17], two criteria are used to measure the errors: (1) C1 errors are calculated with pixels of the ground-truth depth less than 70 m; (2) C2 errors are computed with all pixels in the image.
To evaluate the quantitative results of the proposed method, several state-of-the-art methods are used for comparison. Additionally, considering the influence of the constraint information including sematic information, relative depth trends and CRF on the results, experiments with the dataset are performed, which share the same model with the proposed approach except integrating the constraint information.

3.1. The Experiments with Different Constraint Information

In the experiments, depth maps are predicted via the CNN model with different constraint information. The results are shown in Table 1, where Unconstrained represents the model without integrating the semantic information and relative depth trends of local regions. Sematic_constrained represents the model with integrating only the semantic information. Local_constrained represents the model with integrating only the relative depth trends of local regions. Eucli_loss represents the model in which the loss function of the model replaces the CRF loss with a Euclidean loss and depth reconstruction becomes a regression problem as done in much existing work. A qualitative comparison of depth reconstruction with these methods is presented in Figure 3.
From the results illustrated in Table 1, the following considerations can be outlined.
(1)
The method through Sematic_constrained can get more satisfactory results compared with Unconstrained, which demonstrates the semantic information is an effective cue for depth reconstruction.
(2)
Likewise, the relative depth trends of local regions are helpful to depth reconstruction because the results via Local_constrained outperform Unconstrained.
(3)
The errors of depth reconstruction through Eucli_loss are lower than Unconstrained. This is mainly because their loss functions are different. Eucli_loss uses a Euclidean loss as the loss function of the model. Unlike Eucli_loss, Unconstrained uses a pairwise CRF to establish the loss function, which can consider depth consistency and smoothness between the neighboring superpixels.
(4)
As result of the semantic information, the relative depth trends and the pairwise CRF incorporated into the model, the proposed approach can get more satisfactory results than other methods.

3.2. The Experiments with Different Methods

To show the effectiveness of the proposed approach, several state-of-the-art methods are tested for comparison:
Saxena et al. [13]: The method learns the relation between image features and depth values using MRF. The image features including haze, texture variations and gradient, and shape- and location-based features are manually extracted and represented.
Liu et al. [14]: Based on Saxena et al. [13], Liu et al. [14] added semantic labels to guide depth reconstruction from a single image, but the method still depends on hand-crafted features.
Depth transfer [25]: The method is a non-parametric learning, which avoids explicitly defining a parametric model and requires fewer assumptions as in other methods [13,14]. Likewise, it still depends on hand-crafted features.
DC CRF [17]: In the method, depth prediction is formulated as a discrete-continuous optimization problem, which is solved via particle belief propagation in a graphical model.
DCNF [23]: The method performs depth reconstruction by jointing CNN and CRF. Unlike the proposed approach, the method does not consider semantic information and local detail information from images.
The results of these methods are shown in Table 2. A qualitative comparison of depth reconstruction is presented in Figure 4.
From the results illustrated in Table 2, the following considerations can be noted:
(1)
DCNF [23] and the proposed method significantly outperform the other four methods. This is mainly because the other four methods predict depth maps from a single image via hand-crafted features. Instead, DCNF [23] and the proposed method use the CNN model which can automatically learn a high-level of feature representation without any manual intervention.
(2)
The proposed approach can get more satisfactory results than DCNF [23], because the proposed approach integrated into the semantic information and relative depth trends of local regions.
Besides, depth maps are reconstructed for some images not in the Make3D dataset, but from the Internet, which further demonstrate the effectiveness of the proposed approach in Figure 5.

4. Discussion

Through the experiments, it is observed that the proposed method is successful at depth reconstruction from a single image with satisfactory accuracy. The proposed approach for depth reconstruction uses a unified CNN framework, joining the advantages of the CNN and the continuous pairwise CRF model. On the one hand, it can the automatically learn hierarchical feature representation of the image via CNN model rather than hand-crafted mode. On the other hand, depth reconstruction is formulated as a CRF learning problem rather than a regression problem due to the loss function that uses a continuous pairwise CRF instead of a Euclidean loss. In the continuous pairwise CRF, the depth consistency and smoothness of neighboring superpixels are considered. Additionally, the unified framework incorporates into the sematic information and relative depth trends of local regions, which can be helpful to resolve depth ambiguities and provide more local details in the image. Therefore, depth reconstruction through the proposed approach is effectiveness and has some improvements.

5. Conclusions

In this paper, the development and implementation of a new approach for depth reconstruction from a single image is presented. A unified framework joining a CNN and pairwise CRF model is used to obtain depth information. A particular feature of the approach is that semantic information and relative depth trends of local regions are integrated into the unified framework. A series of experiments on Make3D dataset are presented in this paper. The experiments with different constraint information demonstrate that the semantic information, the relative depth trends of local regions and CRF model are helpful to depth reconstruction from a single image. The experimental results show that the proposed method is effective and suitable for depth reconstruction.

Author Contributions

Dan Liu and Xuejun Liu conceived and designed the experiments; Dan Liu performed the experiments; Dan Liu and Xuejun Liu analyzed the data; Yiguang Wu contributed reagents/materials/analysis tools; Dan Liu wrote the paper.

Acknowledgments

The work described in this paper was supported by the National Natural Science Foundation of China (Project No.: 41701437), the Scientific Research Fund of East (Project No.: DHBK2016102) and Natural Science Foundation of Jiangxi Province ((Project No.: 20161BAB213092).

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Appendix A

The CNN architecture in the unitary part and some implementation details for the proposed approach are shown in this section.

Appendix A.1. CNN Architecture in the Unary Part

The structure of the CNN model in the unary part is shown in Figure A1. The input image is first fed into seven convolutional layers (Conv1 … Conv7) to produce the convolutional feature maps of the image. Then these feature maps are fed to a superpixel pooling layer to transfer into the convolutional feature maps of the superpixels, which are followed by three fully-connected layers. The outputs of the model are the depth at the center of the superpixels and the normalized scale factors.
Figure A1. The CNN architecture in the unary part.
Figure A1. The CNN architecture in the unary part.
Sensors 18 01318 g0a1

Appendix A.2. Implementation Details of the Experiments

In the experiments, the superpixels in the images are firstly over-segmented by Simple Linear Iterative Clustering (SLIC) [26], which clusters pixels in the combined five-dimensional color and image plane space to efficiently generate compact, nearly uniform superpixels. For the Make3D dataset in this paper, the minimum size of the extracted superpixels is set as 10. To generate depth templates, the normalized depth maps of the superpixels in the same sematic label are clustered by Affinity Propagation (AP) [27]. During training and testing, the ground-truth depth values are transferred into log-space. The proposed unified network is trained via stochastic gradient decent with momentum of 0.9, and weight decay of 0.0005. The learning rate is initialized as 0.001, and divided by 2 after 10 cycles. Weights for the convolution layers Conv1 ... Conv5 are initialized by the pre-trained CNN-S model [28].The weights of the other layers are randomly initialized with standard deviation 0.01. In the actual scene, the photograph distance from the sky, that is much larger than the other objects, can be approximated as infinity. Thus, the depth of sky regions in the image can be directly assigned as a maximum value.

References

  1. Bolles, R.C.; Baker, H.H.; Marimont, D.H. Epipolar-Plane Image Analysis: An Approach to Determining Structure from Motion; Springer: Berlin, Germany, 1987; Volume 1, pp. 7–55. [Google Scholar]
  2. Pollefeys, M.; Koch, R.; Vergauwen, M.; Van Gool, L. Automated reconstruction of 3D scenes from sequences of images. ISPRS J. Photogramm. Remote Sens. 2000, 55, 251–267. [Google Scholar] [CrossRef]
  3. Zhang, G.F.; Jia, J.Y.; Wong, T.-T.; Bao, H. Consistent depth maps recovery from a video sequence. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 974–988. [Google Scholar] [CrossRef] [PubMed]
  4. Eigen, D.; Puhrsch, C.; Fergus, R. Depth Map Prediction from a Single Image using a Multi-Scale Deep Network. In Proceedings of the Neural Information Process Systems, Montréal, QC, Canada, 8–13 December 2014. [Google Scholar]
  5. Wilczkowiak, M.; Boyer, E.; Sturm, P. Camera Calibration and 3D Reconstruction from Single Images Using Parallelepipeds. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 142–148. [Google Scholar]
  6. Wang, R.S.; Ferrie, F.P. Self-calibration and metric reconstruction from single image. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Beijing, China, 3–11 July 2008; pp. 639–644. [Google Scholar]
  7. Liu, D.; Liu, X.J.; Wang, M.Z. Camera self-calibration with lens distortion from a single image. Photogramm. Eng. Remote Sens. 2016, 82, 325–334. [Google Scholar] [CrossRef]
  8. Antensteiner, D.; Štolc, S.; Pock, T. A Review of Depth and Normal Fusion Algorithms. Sensors 2018, 18, 431. [Google Scholar] [CrossRef] [PubMed]
  9. Wang, G.H.; Liu, S.Z.; Han, J.Q.; Zhang, X. A Novel Shape from Shading Algorithm for Non-Lambertian Surfaces. In Proceedings of the 3th Measuring Technology and Mechatronics Automation, Shanghai, China, 6–7 January 2011; pp. 222–225. [Google Scholar]
  10. Lobay, A.; Forsyth, A.D. Shape from Texture without Boundaries. Int. J. Comput. Vis. 2006, 67, 71–91. [Google Scholar] [CrossRef]
  11. Toppe, E.; Oswald, M.R.; Cremers, D.; Rother, C. Silhouette-Based Variational Methods for Single View Reconstruction. Video Process. Comput. Video 2011, 7082, 104–123. [Google Scholar]
  12. Saxena, A.; Chung, S.H.; Ng, A.Y. Learning Depth from Single Monocular Images. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 5–8 December 2005; MIT Press: Cambridge, MA, USA, 2015. [Google Scholar]
  13. Saxena, A.; Chung, S.H.; Ng, A.Y. Learning 3D Scene Structure from a Single Still Image. In Proceedings of the International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007. [Google Scholar]
  14. Liu, B.; Koller, D.; Gould, S. Single image depth estimation from predicted semantic labels. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1253–1260. [Google Scholar]
  15. Cao, Y.; Xia, Y.; Wang, Z. A Close-Form Iterative Algorithm for Depth Inferring from a Single Image. In Proceedings of the 2010 European Conference on Computer Vision, Crete, Greece, 5–11 September 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  16. Lin, Y.; Cheng, W.; Miao, H.; Ku, T.-H.; Hsieh, Y.-H. Single image depth estimation from image descriptors. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, 25–30 March 2012; pp. 809–812. [Google Scholar]
  17. Liu, M.; Salzmann, M.; He, X. Discrete-Continuous Depth Estimation from a Single Image. In Proceedings of the 2014 IEEE International Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 716–723. [Google Scholar]
  18. Eigen, D.; Fergus, R. Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture. In Proceedings of the 2015 IEEE International Conference on Europe Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2650–2658. [Google Scholar]
  19. Tian, H.; Zhuang, B.J.; Hua, Y.; Cai, A. Depth Inference with Convolutional Neural Network. In Proceedings of the Visual Communications and Image Processing Conference, Valletta, Malta, 7–20 December 2014; pp. 169–172. [Google Scholar]
  20. Li, B.; Shen, C.H; Dai, Y.C.; van den Hengel, A.; He, M. Depth and surface normal estimation from monocular images using regression on deep features and hierarchical CRFs. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1119–1127. [Google Scholar]
  21. Wang, P.; Shen, X.; Lin, Z.; Cohen, S.; Price, B.; Yuille, A. Towards unified depth and semantic prediction from a single image. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2016; pp. 2800–2809. [Google Scholar]
  22. Liu, F.Y.; Shen, C.H.; Lin, G.S. Deep Convolutional Neural Fields for Depth Estimation from a Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5162–5170. [Google Scholar]
  23. Liu, F.Y.; Shen, C.H.; Lin, G.S.; Reid, I. Learning Depth from Single Monocular Images using Deep Convolutional Neural Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 2024–2039. [Google Scholar] [CrossRef] [PubMed]
  24. Xu, D.; Ricci, E.; Ouyang, W.; Wang, X.; Sebe, N. Multi-scale Continuous CRFs as Sequential Deep Networks for Monocular Depth Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 161–169. [Google Scholar]
  25. Karsch, K.; Liu, C.; Kang, S.B. Depth Transfer: Depth Extraction from Video Using Non-parametric Sampling. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2144–2158. [Google Scholar] [CrossRef] [PubMed]
  26. Achanta, R.; Shaji, A.; Smith, K.; Lucci, A.; Fua, P.; Susstrunk, S. SLIC Superpixels; EPFL Technical Report 149300; EPFL: Lausanne, Switzerland, 2010. [Google Scholar]
  27. Frey, B.J.; Delbert, D. Clustering by passing messages between data points. Science 2007, 315, 972–976. [Google Scholar] [CrossRef] [PubMed]
  28. Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the Devil in the Details: Delving Deep into Convolutional Nets. Comput. Sci. 2014. submitted. [Google Scholar]
Figure 1. The overall framework of the unified CNN model.
Figure 1. The overall framework of the unified CNN model.
Sensors 18 01318 g001
Figure 2. Some local regions of similar relative depth trends from a same sematic label. (a,b) Some different local regions (superpixels) from a same sematic label (in the red box); (c) relative depth trends of the local regions in (a,b) are similar.
Figure 2. Some local regions of similar relative depth trends from a same sematic label. (a,b) Some different local regions (superpixels) from a same sematic label (in the red box); (c) relative depth trends of the local regions in (a,b) are similar.
Sensors 18 01318 g002
Figure 3. Qualitative comparison of depth reconstruction via the proposed approach and Unconstrained. Color indicates depth (red is far, blue is close). (a) Test images (b) Unconstrained (c) Proposed approach (d) Ground-truth.
Figure 3. Qualitative comparison of depth reconstruction via the proposed approach and Unconstrained. Color indicates depth (red is far, blue is close). (a) Test images (b) Unconstrained (c) Proposed approach (d) Ground-truth.
Sensors 18 01318 g003
Figure 4. Qualitative comparison of depth reconstruction via the proposed approach and depth transfer [25]. (a) Test images (b) depth transfer [25] (c) Proposed approach (d) Ground-truth.
Figure 4. Qualitative comparison of depth reconstruction via the proposed approach and depth transfer [25]. (a) Test images (b) depth transfer [25] (c) Proposed approach (d) Ground-truth.
Sensors 18 01318 g004
Figure 5. Depth reconstruction for images from the Internet.
Figure 5. Depth reconstruction for images from the Internet.
Sensors 18 01318 g005
Table 1. Errors of depth reconstruction with different constraints.
Table 1. Errors of depth reconstruction with different constraints.
MethodsC1 ErrorC2 Error
RelLog10 (m)Rmse (m)RelLog10 (m)Rmse (m)
Eucli_loss0.3660.1378.630.3630.14814.41
Unconstrained0.3120.1139.100.3050.12013.24
Semantic_constrained0.2910.1098.740.2870.11412.10
Local_constrained0.2950.1058.530.2910.10911.95
Proposed approach0.2600.0927.160.2450.10310.07
Table 2. Quantitative comparisons with other methods.
Table 2. Quantitative comparisons with other methods.
MethodsC1 ErrorC2 Error
RelLog10 (m)Rmse (m)RelLog10 (m)Rmse (m)
Saxena et al. [13]---0.3700.187-
Liu et al. [14]---0.3790.148-
Depth transfer [25]0.3550.1279.200.3610.14815.10
DC CRF [17]0.3350.1379.490.3380.13412.60
DCNF [23]0.3120.1139.100.3050.12013.24
Proposed approach0.2600.0927.160.2450.10310.07

Share and Cite

MDPI and ACS Style

Liu, D.; Liu, X.; Wu, Y. Depth Reconstruction from Single Images Using a Convolutional Neural Network and a Condition Random Field Model. Sensors 2018, 18, 1318. https://doi.org/10.3390/s18051318

AMA Style

Liu D, Liu X, Wu Y. Depth Reconstruction from Single Images Using a Convolutional Neural Network and a Condition Random Field Model. Sensors. 2018; 18(5):1318. https://doi.org/10.3390/s18051318

Chicago/Turabian Style

Liu, Dan, Xuejun Liu, and Yiguang Wu. 2018. "Depth Reconstruction from Single Images Using a Convolutional Neural Network and a Condition Random Field Model" Sensors 18, no. 5: 1318. https://doi.org/10.3390/s18051318

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop