Next Article in Journal
Reliability of Repeated Nordic Hamstring Strength in Rugby Players Using a Load Cell Device
Previous Article in Journal
EEG Signal Processing for Biomedical Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Refinement Scheme for Depth Estimation Networks

by
Amin Alizadeh Naeini
,
Mohammad Moein Sheikholeslami
and
Gunho Sohn
*
Department of Earth and Space Science and Engineering, Lassonde School of Engineering, York University, 4700 Keele Street, Toronto, ON M3J1P3, Canada
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9755; https://doi.org/10.3390/s22249755
Submission received: 17 October 2022 / Revised: 18 November 2022 / Accepted: 9 December 2022 / Published: 13 December 2022
(This article belongs to the Section Sensors and Robotics)

Abstract

:
Deep learning has proved to be a breakthrough in depth generation. However, the generalization ability of deep networks is still limited, and they cannot maintain a satisfactory performance on some inputs. By addressing a similar problem in the segmentation field, a feature backpropagating refinement scheme (f-BRS) has been proposed to refine predictions in the inference time. f-BRS adapts an intermediate activation function to each input by using user clicks as sparse labels. Given the similarity between user clicks and sparse depth maps, this paper aims to extend the application of f-BRS to depth prediction. Our experiments show that f-BRS, fused with a depth estimation baseline, is trapped in local optima, and fails to improve the network predictions. To resolve that, we propose a double-stage adaptive refinement scheme (DARS). In the first stage, a Delaunay-based correction module significantly improves the depth generated by a baseline network. In the second stage, a particle swarm optimizer (PSO) delineates the estimation through fine-tuning f-BRS parameters—that is, scales and biases. DARS is evaluated on an outdoor benchmark, KITTI, and an indoor benchmark, NYUv2, while for both, the network is pre-trained on KITTI. The proposed scheme was effective on both datasets.

1. Introduction

Dense depth maps play a crucial role in a variety of applications, such as simultaneous localization and mapping (SLAM) [1], visual odometry [2], and object detection [3]. With the advent of deep learning (DL) and its ever-growing success in most fields, DL methods have also been utilized for generating dense depth (DD) maps and have demonstrated a prominent improvement in this field.
Depending on the primary input, DL-based depth generation methods can be categorized into depth completion and estimation methods. Depth completion methods try to fill the gaps present in input sparse depth (SD) maps [4,5], whereas depth estimation ones attempt to estimate depth for each pixel of an input image [6,7,8,9,10]. Although the results provided by depth completion methods [4,11] are usually more accurate than those from depth estimation ones, they need to be supplied by a remarkably large number of DD maps as targets in the training stage. This is while collecting such data in a real-world application is an expensive and time-consuming task [8,12].
In parallel, the input to depth estimation methods includes no depth maps; however, supervised ones take either SD [13] or DD maps [14,15,16] as the target during training. Between these two supervised depth estimation approaches, using SD maps usually leads to less accurate results but is more viable than using DD ones. Because SD-based methods only need SD maps which can be provided using a LiDAR sensor and without any need for post-processing or labeling effort. Considering the above issues, supervised depth estimation methods which use sparse depth maps are preferred, especially for most real-world cases in which access to large-enough accurate DD maps is difficult or even impossible.
Similar to all DL methods, DL-based depth estimation methods, no matter supervised or unsupervised, suffer from the generalization problem. In other terms, DL models trained on an arbitrary dataset are not able to preserve their satisfying performance neither on unseen data samples nor even on hard seen ones. This problem is more severe in applications such as SLAM and autonomous vehicles, where the test environment is under a constant change. Hence, the high variety in input samples leads to accuracy degradation [17].
In the segmentation field, a similar problem, i.e., limited generalization ability, has been alleviated by the backpropagating refinement scheme (BRS) [18]. This method has been proposed to optimize an input to a segmentation network based on user clicks. The method performs this process via backpropagating through the whole network; thus, it suffers from the computational burden. To speed up the process, feature BRS (f-BRS) [19] has been proposed to only backpropagate through several last layers and optimizes the activation responses of an intermediate layer using some introduced parameters. Optimizing those parameters during inference can be viewed as an adaptive approach because an intermediate activation function and indeed the output are adapted to input data. In other words, f-BRS converts the baseline network to a functionally adaptive method, in which the shape of activation function changes in the inference time [20,21].
As user clicks and SD maps are both sparse labels, it seems that the application of the above-mentioned scheme, f-BRS [19], can be generalized to depth estimation. Accordingly, we present an inference-time functionally adaptive refinement scheme for depth estimation networks (see Figure 1). For this purpose, f-BRS [19] is used, which can be injected into any DL baseline and adapts the model to hard and unseen environments or different datasets. Nevertheless, f-BRS suffers from two fundamental issues which prevent it from being applicable in depth estimation. The first problem is associated with its nature which cannot manage highly inaccurate products (here, the depth predicted by the network), since f-BRS has been originally proposed for interactive segmentation [19], where is often no need for a considerable modification. To resolve this, a sliced Delaunay correction (SDC) module is designed to carry out a correction using SD maps and provide an appropriate initial value for the optimizer. The second problem is that f-BRS is trapped in local optima due to its local optimizer. To address this, f-BRS is equipped with particle swarm optimization (PSO) as a global optimizer [22]. For simplicity, this novel generalization of f-BRS, which is applicable to depth estimation, is named as the double-stage adaptive refinement scheme (DARS), where the first stage is carried out through SDC and the second stage is conducted by optimizing the f-BRS parameters, i.e., adapting the activation maps.
Overall, our contributions can be summarized as:
  • A novel double-stage adaptive refinement scheme for monocular depth estimation networks. The proposed scheme needs neither offline data gathering nor offline training, because it uses available pre-trained weights.
  • Introduction of functional adaptation schemes in the field of depth generation, for the first time. Using the proposed adaptive scheme, pre-trained networks can be straightforwardly used for unseen datasets through adjusting the shape of activation functions of an intermediate layer.
  • A model-agnostic scheme which can be plugged into any baseline. In this paper, we selected Monodepth2 [23] as one of the most widely used baselines for depth estimation.

2. Related Work

Here, we initially provide an overview for unsupervised and supervised depth estimation methods. In the last part, a brief review of functionally adaptive networks is displayed.

2.1. Unsupervised Depth Estimation Methods

These methods use color consistency loss between stereo images [24], temporal ones [25], or a combination of both [23] to train a monocular depth estimation model. Many attempts have been made to rectify the self-supervision by new loss terms such as left–right consistency [26], temporal depth consistency [27], or cross-task consistency [28,29,30]. Of these improvements, Monodepth2 has attracted substantial attention because of the different sets of techniques it has used for modification [23]. To the best of our knowledge, methods in this category have been presented for either outdoor environments, such as the above ones, or indoor environments, as in [31]. Not being applicable for both indoor and outdoor datasets can be regarded as a drawback of these methods. Another problem of these methods is that they suffer from low accuracy.

2.2. Supervised Depth Estimation Methods

The inputs to these methods are only images and they use either DD maps or SD maps as targets. This group can be categorized into DD-based and SD-based methods. Of these two, DD-based ones need DD maps during their training. DD-based methods, such as Adabins [14] and BTS [15], learn based on the error between predicted depth maps and DD maps. The main disadvantage of these methods is that they need DD maps for training.
Unlike DD-based methods, SD-based ones use SD maps only. Training data are not an issue for these methods because current robots and mapping systems can capture both images and SD maps simultaneously. The distance between predictions and SD maps are used as loss functions [32,33,34]. These methods are also known as semi-supervised [35].

2.3. Functionally Adaptive Neural Networks

Neural networks are called adaptive when they can adapt themselves to unseen environments, i.e., new inputs [36,37]. There are different techniques for designing adaptive networks, among which weight modification and functional adaptation can be mentioned. The former optimizes the network weights for new inputs while the latter modifies the slope and shape of the activation functions usually through a relatively few number of additional parameters [36]. Functional adaptation can be categorized under activation response optimization methods [38,39,40,41], in which the aim is to update activation responses while the network weights are fixed. The reason behind keeping the network weights fixed is to preserve the semantics learned by the network during the training process. On the other hand, one or several activation responses are modified to optimize the performance on inevitable unseen objects and scenes so that the network maintains its proficient performance in constantly changing environments [19].
The adaptation process can happen in either the training stage [20] or the inference stage for some tasks, such as interactive segmentation or SLAM, where some ground truth (even though sparse) is available on the fly [37]. In addition, the networks can adapt to a sequence of images or a single image. In a single-image adaptation, the core merit optimizes the prediction for a specific image or even an object, and the adaptation is discarded for the next image [37]. Thus, single-image adaptation can be beneficial, especially when scenes are prone to varying significantly.
Inspired from the biological neurons, some investigations have been conducted on the adaptive activation functions such as PReLU, which shows that adaptation behaviour in such activation functions can improve the accuracy and generalization of neural networks [20]. In [19], some parameters are introduced to adapt the activation functions to user clicks during the inference of the interactive segmentation task. An adaptive instance normalization layer is proposed in [21], which enables the style transfer networks to adapt to arbitrary new styles, adding a negligible computational cost.

3. Theoretical Background

In this section, some theoretical background needed for understanding the proposed scheme is provided. Firstly, Delaunay-based interpolation is explained, which is used in the correction stage to densify sparse correction maps. Subsequently, the particle swarm optimization (PSO) algorithm is displayed as the optimizer utilized in the optimization stage of the proposed scheme.

3.1. Delaunay-Based Interpolation

The first step of the interpolation is to conduct triangulation. Considering that there are many different triangulations for a given point set, we should obtain an optimal triangulation method, avoiding poorly shaped triangles. The Delaunay triangulation method has proved to be the most robust and widely used triangulation approach. This method connects all the neighbouring points in a Voronoi diagram to obtain a triangulation [42].
To find the value of any new point by interpolation, its corresponding triangle in which it lies should be identified. Suppose P ( x , y ) is a new point that belongs to a triangle with vertices of P 1 ( x 1 , y 1 ) , P 2 ( x 2 , y 2 ) and P 3 ( x 3 , y 3 ) with the values of z 1 , z 2 and z 3 , respectively, to linearly interpolate the value z of P, we should fit a plane (Equation (1)) to the vertices P 1 , P 2 and P 3 .
z = a x + b y + c
By inserting the known points ( x 1 , y 1 , z 1 ) , ( x 2 , y 2 , z 2 ) and ( x 3 , y 3 , z 3 ) in Equation (1) and solving a linear system of equations, the unknown coefficients ( a , b , c ) of the plane are estimated. Finally, applying Equation (1) and having ( a , b , c ) , the value z for any arbitrary point P ( x , y ) is interpolated within the triangle (Figure 2).

3.2. PSO

PSO is a population-based stochastic optimization technique inspired by the social behavior of birds within a flock or fish schooling [22]. PSO has two main components which need to be specifically defined for each application. One component is the introduction of particles, and the other is an objective function for particle evaluation.
Each particle has the potential of solving the problem; this means they must contain all the arguments needed for the problem in question.
The velocity and position of each particle are calculated using Equation (2) and Equation (3), respectively, [22]. Optimum values of unknown parameters are iteratively updated using the position equation, which is itself dependent on the velocity.
V i ( t + 1 ) = w V i ( t ) + c 1 r 1 ( t ) [ p b e s t i ( t ) X i ( t ) ] + c 2 r 2 ( t ) [ g b e s t i ( t ) X i ( t ) ]
X i ( t + 1 ) = X i ( t ) + V i ( t + 1 )
In Equation (2), V i ( t ) is the velocity of a particle i at time t, and p b e s t i ( t ) and g b e s t i ( t ) are personal and global best positions found by the particle i and all the particles by the iteration t, respectively. The w parameter is an inertia weight scaling the previous time step velocity. Parameters c 1 and c 2 are two acceleration coefficients that scale the influence of p b e s t i ( t ) and g b e s t i ( t ) , respectively. In addition, parameters r 1 and r 2 are random variables between 0 and 1 obtained from a uniform distribution. The next position of each particle ( X i ( t + 1 ) ) can be calculated using Equation (3).

4. Proposed Method

Supervised depth estimation methods suffer from the generalization problem. In other words, they usually need to be retrained for achieving a proficient performance on an unseen dataset. To alleviate this, a double-stage adaptive refinement scheme (DARS) is proposed to equip pre-trained depth estimation networks with inference-time optimization for improving the performance on both seen and unseen datasets. The proposed scheme (Figure 3) consists of several components including a deep baseline model, a correction module which applies the first stage of refinement, and an activation optimization as the second stage. The baseline model could be any supervised or unsupervised pre-trained depth estimation network. The predicted depth by the baseline is given to the correction module to provide the optimization module with a sufficiently accurate depth map. In the second stage, scale and bias parameters are applied on a set of intermediate feature maps in the baseline, and they are optimized by a PSO to improve the accuracy of the final depth. The tasks and details of each module, and the overall proposed scheme, are displayed below. In the following subsections, s and d superscripts, respectively, indicate that depth maps are sparse or dense.

4.1. Baseline

Given an input monocular RGB image I R w × h × 3 , we rely on a depth estimation network F : I D 0 d to provide us with an initial depth map D 0 d R w × h . The proposed scheme can utilize any monocular depth estimation network. In this study, Monodepth2 [23] has been selected as the baseline, as one of most widely used depth estimation networks. The baseline is pre-trained and the weights are kept fixed.

4.2. Correction

The depth map D 0 d predicted by the baseline lacks sufficient accuracy, especially for an unseen input. Thus, D 0 d is not a proper initial value for the optimization stage. As a solution, in the first stage of the proposed refinement scheme, a sliced Delaunay correction (SDC) C : R w × h R w × h is used to correct D 0 d , using the available sparse depth map D s . In SDC, first a correction value δ d s Δ D s for any available depth pixel d s D s is calculated:
δ d s = d 0 s d s
where d 0 s D 0 d are the pixels in D 0 d corresponding to the ones in D s . Then, the sparse correction map Δ D s is divided into three overlapped slices (see Figure 4). Neighbouring pixels are intuitively assumed to share a similar error pattern, and slices can represent a simplistic segmentation based on the error pattern.
In each slice, a Delaunay-based interpolation (see Section 3.1) J : R 2 R is utilized to estimate a dense correction map Δ D d = J ( Δ D s ) , given the sparse one Δ D s . For the pixels in overlapped areas (see Figure 4), the average of the values coming from two adjacent slices is considered as the final depth correction value. As a result of this stage, a corrected depth D ^ d = D 0 d + Δ D d is generated, yet with marginal errors.
Regarding the number of slices, three was selected as the optimal number of slices on both datasets based on our experiments. Lower numbers could not result in homogeneous areas, and hence, a remarkable correction performance. On the other hand, a larger number was not considered because the improvement in accuracy was negligible with respect to the computational overload.

4.3. Activation Optimization

Given the initial value from the first stage (correction), the core part of network adaptation is conducted in the second stage. The technique chosen for the network adaptation is to modify an intermediate set of activation outputs [36]. This is usually carried out by freezing the weights and optimizing some auxiliary parameters. This way, not only are the valuable learned semantics preserved, but also the network can adapt itself to inputs. Inspired from works such as f-BRS [19] in an interactive segmentation field, we apply channel-wise scale and bias parameters on intermediate features of the baseline network. The scales are initialized to ones and biases to zeros; they are then optimized based on a cost function. To describe the algorithm of the optimization module better, the overall scheme, i.e., from the baseline to optimization module, is explained, followed by some details about the optimizer.

4.3.1. Overall Scheme

Given an input RGB image I R w × h × 3 , denote the intermediate feature set as G ( I ) R m × n × c where G : R w × h R m × n × c is the network body and m, n, and c are, respectively, width, height, and number of channels. The auxiliary parameters, scales S R c and biases B R c are applied on G ( I ) , and the depth D 0 d = H ( S G ( I ) B ) is predicted, where H : R m × n × c R w × h is the network head, and ⨂ and ⨁ represent channel-wise multiplication and addition. Afterwards, the correction module C : R w × h R w × h carries out the first refinement stage on D 0 d and returns D ^ d :
D ^ d = C ( D 0 d , D s )
The auxiliary parameters X R 2 c , i.e., channel-wise scales and biases, are learnable. Therefore, the following optimization problem can be formulated as:
L ( D ^ d ( I , X + Δ X ) , D s ) min Δ X .
where Δ X is the corrections applied to the parameters and L is the cost function given to the optimizer.

4.3.2. Optimizer

The above optimization problem can be given to any type of optimizers. The default optimizer of f-BRS is limited-memory Broyden–Fletcher–Goldfarb–Shanno (L-BFGS) [43,44]. This optimizer, due to its local gradient-based nature, is trapped in local optima. To overcome this problem, L-BFGS is replaced with PSO [22]. PSO iteratively updates scale and bias parameters in each particle based on the below distance loss:
L = 1 T i = 1 T log ( d ^ s ) log ( d s ) ,
where T is the total number of pixels with depth values in D s . Figure 5 shows the algorithm flow of the PSO and its parameters in DARS.

5. Experiments

In this section, we first briefly describe the datasets used in the experiments. Secondly, the metrics are introduced, and after that, an ablation study is discussed to show the effectiveness of each module. Finally, the results by the proposed scheme are compared with those of the state of the art.

5.1. Datasets

Two datasets are used in the experiments, KITTI [45] and NYUv2 [46]. KITTI is a well-known outdoor dataset, on which the baseline is trained, while NYUv2 is an indoor benchmark dataset and the adaptation performance of the scheme is highlighted through testing on it.

5.1.1. KITTI

The KITTI dataset [45] consists of stereo RGB images and corresponding SD and DD maps of 61 outdoor scenes acquired by 3D mobile laser scanners. The RGB images have a resolution of 1241 × 376 pixels, while the corresponding SD maps are of very low density with lots of missing data. The dataset is divided into 23,488 train and 697 test images, according to [47]. For testing, 652 images associated with DD maps are selected from the test split. Sample data have been shown in Figure 6 for the KITTI dataset.

5.1.2. NYUv2

The NYUv2 dataset [46] contains 120,000 RGB and depth pairs of 640 × 480 pixels in size, acquired as video sequences using a Microsoft Kinect from 464 indoor scenes. The official train/test split contains 249 and 215 scenes, respectively. Given that NYUv2 does not contain any SD maps, SD maps with 80% sparsity have been randomly synthesized from DD maps for the experiments of the proposed method. Sample data for NYUv2 dataset, including the synthetic SD maps, are illustrated in Figure 6.

5.2. Assessment Criteria

Assessment criteria proposed by [47] include error and accuracy metrics. The error metrics are root mean square error (RMSE), logarithmic RMSE ( RMSE log ), absolute relative error (Abs Rel), and square relative error (Sq Rel), whereas the accuracy rate metrics contain t h r = 1 . 25 t where t = 1 , 2 , 3 . These criteria are formulated as follows:
R M S E = 1 T i T | | d i d i g t | | 2 ,
R M S E l o g = 1 T i T | | l o g ( d i ) l o g ( d i g t ) | | 2 ,
A b s R e l = 1 T i T d i d i g t d i g t ,
S q R e l = 1 T i T | | d i d i g t | | 2 d i g t , a n d
a c c u r a c i e s = % of d i subject to m a x ( d i d i g t , d i g t d i ) = δ < t h r
where d i and d i g t are the predicted and target (ground truth) depth, respectively, at the pixel indexed by i, and T is the total number of pixels in all the evaluated images.

5.3. Network Architecture

As the proposed scheme is by design model agnostic, the network architecture is not the focus of this study. Thus, we used the standard monocular version of the Monodepth2 [23] model with the input size of 640 × 192 × 3 .

5.4. Implementation Details

We have used monocular Monodepth2 pre-trained on KITTI as our baseline. The input images were resampled to 640 × 192 and then were fed to the network. The weights were fixed and the network was run in inference mode. In SDC, the number of slices were three and the overlap between slices was set to 50%. Moreover, PSO paramters, i.e., c 1 , c 2 , number of particles, and number of iterations were, respectively, set to 0.5, 0.3, 10, and 30 in all the experiments. Furthermore, all the implementations were conducted in PyTorch [48].

5.5. Ablation Studies

This ablation study aims to prove the effectiveness of different stages and modules in the proposed scheme. To do this, starting from the baseline, we have enabled the correction and optimization modules in several steps (see Table 1). First of all, the result of Monodepth2 [23] with median scaling is discussed for comparison, while the version without any kind of post-processing (Monodepth2*) is also reported as our baseline. It means that the baseline results are without median scaling by target DD maps. As a result, they suffer from scale ambiguity and low accuracy. In addition, DC is introduced to show the efficacy of slicing in our proposed SDC as the correction module. The difference between SDC and DC is that, in the latter, Delaunay interpolation and correction are carried out on the entire depth maps instead of separately on each slice. For the sake of brevity, these two methods have just been surveyed for KITTI.
From Table 1, the worst results on KITTI in terms of all the metrics was recorded by the baseline (Monodepth2*), which was expected because of scale ambiguity. Using DC as the correction module improved the results by 13% in terms of RMSE, while SDC showed a significantly higher improvement over the baseline by 91%. This not only proves the contribution of the correction module but also indicates the effectiveness of the slicing process in SDC. Furthermore, SDC, without any use of target DD maps, yielded over 57% improvement with respect to Monodepth2, which means that SDC not only addresses the scale ambiguity problem but also corrects the given depth map significantly. Moreover, this observation supports the assumption that adjacent pixels in depth maps share a similar error pattern. First because adjacent pixels usually belong to same objects. Second, the error in LiDAR sensor has a correlation with distance from sensor, and as a result, pixels which are in an approximately equal distance to the sensor are likely to have close error magnitudes. From another perspective, the proposed slicing proved to be a simplistic segmentation based on the error pattern and was able to remarkably contribute to the correction stage.
According to Table 1, the results obtained when using L-BFGS as the optimizer are equal to ones without optimization on both KITTI and NYUv2 datasets. This means that L-BFGS could not improve the results because, unlike PSO, it does not have the capability for global search. In better words, it seems that it was trapped in local optima, i.e., the depth provided by SDC. Therefore, due to the identical performances and for the sake of conciseness, just one row is dedicated to both SDC and L-BFGS in Figure 7 and Figure 8.
In the meanwhile, PSO improved the results significantly in terms of all metrics and on both KITTI and NYUv2 datasets. For instance, PSO showed nearly 50% enhancement in AbsRel and 14% in RMSE on KITTI and 6% and 86%, respectively, in terms of AbsRel and RMSE on NYUv2.
If we compare the improvement of PSO over L-BFGS on KITTI and that on NYUv2, it can be observed that the improvement was more remarkable on NYUv2. Thus, considering that the baseline was trained on KITTI, one can conclude that the optimization module with PSO as its optimizer plays a significant role in the adaptation process. This observation also demonstrated the capability and efficacy of the activation optimization used in the proposed scheme.
To conclude, both of the proposed correction and optimization stages in DARS, i.e., SDC and activation optimization using PSO, proved to be effective and led to considerable improvements. Moreover, DARS proved its capability in network adaptation, given its performance on NYUv2.
As is clear from error patterns in Figure 7, related to KITTI and Figure 8 pertaining to NYUv2, the introduction of PSO has led to considerable improvements. The improvements can be specifically observed in more distant pixels which are usually of a higher error magnitude.

5.6. Comparison with SOTA

Proficient generalization is necessary for DL-based depth estimation methods, especially in applications with constantly changing environments, such as SLAM and autonomous vehicles. To deal with this problem, an inference-time refinement scheme is proposed to help pre-trained networks adapt to new inputs. To show the generalization performance of the proposed scheme, it has been compared with a range of unsupervised and supervised methods. On the other hand, to evaluate its adaptation performance, DARS with pre-trained weight on KITTI is applied on an unseen benchmark dataset, namely NYUv2. As is clear from Table 2, DARS outperformed competing methods in terms of almost all assessment criteria except for δ 1 . 25 2 and δ 1 . 25 3 . From the perspective of these two criteria, the performance of our method was not as good as the second-place rival. However, DARS led to better performance in terms of δ 1.25 , which is the primary criterion for accuracy assessment. Although DARS utilizes a self-supervised baseline, Monodepth2, it outperformed its supervised rivals by a 39% margin in terms of RMSE on KITTI. This confirms the superiority of the proposed DARS even over supervised approaches and in dealing with harder scenes in a seen dataset. A visual comparison between DARS and the second best method in terms of RMSE on KITTI is presented in Figure 9.
Regarding the second dataset, NYUv2, DARS outperformed the competing methods in terms of all criteria according to Table 3. In terms of AbsRel and RMSE, DARS reached improvements of, respectively, 83% and 70% with respect to the best competing method. Furthermore, this table indicates how the proposed method successfully adapted to an unseen dataset. Note that unlike DARS, the other methods in Table 3 have been trained on NYUv2. Hence, one can deduce that DARS not only could adapt a network to an unseen dataset but also outperformed the methods trained on the exact same dataset. Furthermore, it suggests DARS as a possible alternative to supervised approaches which suffer from complicated generalization problems in practice. This adaptation capability is extremely advantageous in applications with constantly changing environments such as SLAM, where the scenes are of an unlimited variety and sparse LiDAR maps are available on the fly. A visual comparison between DARS and the second best method in terms of RMSE on NYUv2 is presented in Figure 10.

6. Conclusions

This paper deals with one of the main problems of available deep learning-based depth estimation networks, which is their limited generalization capability. This problem specifically restricts the practical usage of such models in applications with a constantly changing environment, such as SLAM. To alleviate this problem, a new double-stage adaptive refinement scheme for depth estimation networks, namely, DARS based on the combination of f-BRS and PSO, is proposed in this paper. DARS, here, is injected into Monodepth2 as the baseline and adapts the pre-trained network to each input during inference. Experimental results on KITTI and NYUv2 datasets demonstrated the efficacy of the proposed scheme not only for KITTI but also for NYUv2, while the baseline model was pre-trained only on KITTI. Although our approach is model agnostic by design, this paper did not explore the effects of using different baselines. In future work, we will, therefore, replace our unsupervised baseline with other networks, ranging from unsupervised to supervised in order to investigate the effectiveness of our proposed scheme on different baselines.

Author Contributions

A.A.N.: conceptualization, methodology, software, formal analysis, writing-original draft preparation, and visualization; M.M.S.: methodology, software, and writing—review and editing; G.S.: Conceptualization, Supervision, writing—review and editing, and project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) Collaborative Research and Development (CRD).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yousif, K.; Taguchi, Y.; Ramalingam, S. MonoRGBD-SLAM: Simultaneous localization and mapping using both monocular and RGBD cameras. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; IEEE: Piscataway Township, NJ, USA, 2017; pp. 4495–4502. [Google Scholar]
  2. Li, R.; Wang, S.; Long, Z.; Gu, D. Undeepvo: Monocular visual odometry through unsupervised deep learning. In Proceedings of the 2018 IEEE international conference on robotics and automation (ICRA), Brisbane, Australia, 21–25 May 2018; IEEE: Piscataway Township, NJ, USA, 2018; pp. 7286–7291. [Google Scholar]
  3. Dimas, G.; Gatoula, P.; Iakovidis, D.K. MonoSOD: Monocular Salient Object Detection based on Predicted Depth. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway Township, NJ, USA, 2021; pp. 4377–4383. [Google Scholar]
  4. Hu, M.; Wang, S.; Li, B.; Ning, S.; Fan, L.; Gong, X. PENet: Towards Precise and Efficient Image Guided Depth Completion. arXiv 2021, arXiv:2103.00783. [Google Scholar]
  5. Park, J.; Joo, K.; Hu, Z.; Liu, C.K.; So Kweon, I. Non-local spatial propagation network for depth completion. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 120–136. [Google Scholar]
  6. Gurram, A.; Tuna, A.F.; Shen, F.; Urfalioglu, O.; López, A.M. Monocular Depth Estimation through Virtual-world Supervision and Real-world SfM Self-Supervision. arXiv 2021, arXiv:2103.12209. [Google Scholar] [CrossRef]
  7. Hirose, N.; Koide, S.; Kawano, K.; Kondo, R. Plg-in: Pluggable geometric consistency loss with wasserstein distance in monocular depth estimation. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway Township, NJ, USA, 2021; pp. 12868–12874. [Google Scholar]
  8. Liu, J.; Li, Q.; Cao, R.; Tang, W.; Qiu, G. MiniNet: An extremely lightweight convolutional neural network for real-time unsupervised monocular depth estimation. ISPRS J. Photogramm. Remote Sens. 2020, 166, 255–267. [Google Scholar] [CrossRef]
  9. Hwang, S.J.; Park, S.J.; Kim, G.M.; Baek, J.H. Unsupervised Monocular Depth Estimation for Colonoscope System Using Feedback Network. Sensors 2021, 21, 2691. [Google Scholar] [CrossRef]
  10. Wang, Y.; Zhu, H. Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network. Sensors 2022, 22, 6344. [Google Scholar] [CrossRef]
  11. Cheng, X.; Wang, P.; Guan, C.; Yang, R. Cspn++: Learning context and resource aware convolutional spatial propagation networks for depth completion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10615–10622. [Google Scholar]
  12. Kuznietsov, Y.; Stuckler, J.; Leibe, B. Semi-supervised deep learning for monocular depth map prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6647–6655. [Google Scholar]
  13. Huang, Y.K.; Liu, Y.C.; Wu, T.H.; Su, H.T.; Chang, Y.C.; Tsou, T.L.; Wang, Y.A.; Hsu, W.H. S3: Learnable Sparse Signal Superdensity for Guided Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16706–16716. [Google Scholar]
  14. Bhat, S.F.; Alhashim, I.; Wonka, P. Adabins: Depth estimation using adaptive bins. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4009–4018. [Google Scholar]
  15. Lee, J.H.; Han, M.K.; Ko, D.W.; Suh, I.H. From big to small: Multi-scale local planar guidance for monocular depth estimation. arXiv 2019, arXiv:1907.10326. [Google Scholar]
  16. Lee, S.; Lee, J.; Kim, B.; Yi, E.; Kim, J. Patch-Wise Attention Network for Monocular Depth Estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 1873–1881. [Google Scholar]
  17. Liu, Y.; Yixuan, Y.; Liu, M. Ground-aware monocular 3d object detection for autonomous driving. IEEE Robot. Autom. Lett. 2021, 6, 919–926. [Google Scholar] [CrossRef]
  18. Jang, W.D.; Kim, C.S. Interactive Image Segmentation via Backpropagating Refinement Scheme. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  19. Sofiiuk, K.; Petrov, I.; Barinova, O.; Konushin, A. f-brs: Rethinking backpropagating refinement for interactive segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 8623–8632. [Google Scholar]
  20. Lau, M.M.; Lim, K.H. Review of adaptive activation function in deep neural network. In Proceedings of the 2018 IEEE-EMBS Conference on Biomedical Engineering and Sciences (IECBES), Sarawak, Malaysia, 3–6 December 2018; IEEE: Piscataway Township, NJ, USA, 2018; pp. 686–690. [Google Scholar]
  21. Huang, X.; Belongie, S. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1501–1510. [Google Scholar]
  22. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; IEEE: Piscataway Township, NJ, USA, 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  23. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2017; pp. 3828–3838. [Google Scholar]
  24. Zhou, T.; Brown, M.; Snavely, N.; Lowe, D.G. Unsupervised learning of depth and ego-motion from video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1851–1858. [Google Scholar]
  25. Ye, X.; Ji, X.; Sun, B.; Chen, S.; Wang, Z.; Li, H. DRM-SLAM: Towards dense reconstruction of monocular SLAM with scene depth fusion. Neurocomputing 2020, 396, 76–91. [Google Scholar] [CrossRef]
  26. Godard, C.; Mac Aodha, O.; Brostow, G.J. Unsupervised monocular depth estimation with left-right consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 270–279. [Google Scholar]
  27. Bian, J.; Li, Z.; Wang, N.; Zhan, H.; Shen, C.; Cheng, M.M.; Reid, I. Unsupervised scale-consistent depth and ego-motion learning from monocular video. Adv. Neural Inf. Process. Syst. 2019, 32, 35–45. [Google Scholar]
  28. Yin, Z.; Shi, J. Geonet: Unsupervised learning of dense depth, optical flow and camera pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–20 June 2018; pp. 1983–1992. [Google Scholar]
  29. Cai, H.; Matai, J.; Borse, S.; Zhang, Y.; Ansari, A.; Porikli, F. X-Distill: Improving Self-Supervised Monocular Depth via Cross-Task Distillation. arXiv 2021, arXiv:2110.12516. [Google Scholar]
  30. Feng, T.; Gu, D. Sganvo: Unsupervised deep visual odometry and depth estimation with stacked generative adversarial networks. IEEE Robot. Autom. Lett. 2019, 4, 4431–4437. [Google Scholar] [CrossRef] [Green Version]
  31. Ji, P.; Li, R.; Bhanu, B.; Xu, Y. Monoindoor: Towards good practice of self-supervised monocular depth estimation for indoor environments. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 12787–12796. [Google Scholar]
  32. Fei, X.; Wong, A.; Soatto, S. Geo-supervised visual depth prediction. IEEE Robot. Autom. Lett. 2019, 4, 1661–1668. [Google Scholar] [CrossRef] [Green Version]
  33. dos Santos Rosa, N.; Guizilini, V.; Grassi, V. Sparse-to-continuous: Enhancing monocular depth estimation using occupancy maps. In Proceedings of the 2019 19th International Conference on Advanced Robotics (ICAR), Belo Horizonte, Brazil, 2–6 December 2019; IEEE: Piscataway Township, NJ, USA, 2019; pp. 793–800. [Google Scholar]
  34. Ma, F.; Cavalheiro, G.V.; Karaman, S. Self-supervised sparse-to-dense: Self-supervised depth completion from lidar and monocular camera. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway Township, NJ, USA, 2019; pp. 3288–3295. [Google Scholar]
  35. Ming, Y.; Meng, X.; Fan, C.; Yu, H. Deep learning for monocular depth estimation: A review. Neurocomputing 2021, 438, 14–33. [Google Scholar] [CrossRef]
  36. Palnitkar, R.M.; Cannady, J. A review of adaptive neural networks. In Proceedings of the IEEE SoutheastCon, Greensboro, NC, USA, 26–29 March 2004; IEEE: Piscataway Township, NJ, USA, 2004; pp. 38–47. [Google Scholar]
  37. Kontogianni, T.; Gygli, M.; Uijlings, J.; Ferrari, V. Continuous adaptation for interactive object segmentation by learning from corrections. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 579–596. [Google Scholar]
  38. Gatys, L.; Ecker, A.S.; Bethge, M. Texture synthesis using convolutional neural networks. In Advances in Neural Information Processing Systems 28; Curran Associates, Inc.: Red Hook, NY, USA, 2015. [Google Scholar]
  39. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
  40. Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv 2013, arXiv:1312.6034. [Google Scholar]
  41. Zhang, J.; Bargal, S.A.; Lin, Z.; Brandt, J.; Shen, X.; Sclaroff, S. Top-down neural attention by excitation backprop. Int. J. Comput. Vis. 2018, 126, 1084–1102. [Google Scholar] [CrossRef]
  42. Amidror, I. Scattered data interpolation methods for electronic imaging systems: A survey. J. Electron. Imaging 2002, 11, 157–176. [Google Scholar] [CrossRef] [Green Version]
  43. Zhu, C.; Byrd, R.H.; Lu, P.; Nocedal, J. Algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound-constrained optimization. ACM Trans. Math. Softw. (TOMS) 1997, 23, 550–560. [Google Scholar] [CrossRef]
  44. Byrd, R.H.; Lu, P.; Nocedal, J.; Zhu, C. A limited memory algorithm for bound constrained optimization. SIAM J. Sci. Comput. 1995, 16, 1190–1208. [Google Scholar] [CrossRef]
  45. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  46. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  47. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. arXiv 2014, arXiv:1406.2283. [Google Scholar]
  48. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
  49. Luo, C.; Yang, Z.; Wang, P.; Wang, Y.; Xu, W.; Nevatia, R.; Yuille, A. Every pixel counts++: Joint learning of geometry and motion with 3d holistic understanding. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2624–2641. [Google Scholar] [CrossRef] [PubMed]
  50. Zhang, J.; Li, W.; Gou, H.; Fang, L.; Yang, R. LEAD: LiDAR Extender for Autonomous Driving. arXiv 2021, arXiv:2102.07989. [Google Scholar]
  51. Lee, M.; Hwang, S.; Park, C.; Lee, S. EdgeConv with Attention Module for Monocular Depth Estimation. arXiv 2021, arXiv:2106.08615. [Google Scholar]
Figure 1. Overall performance of the proposed scheme: (a) Train on dataset 1, (b) inference on the same or another dataset. Keeping the pre-trained weights fixed, channel-wise scales and biases are applied to an intermediate activation map and are optimized to improve the predictions for the input dataset. The input dataset can be identical to the training dataset or a new one.
Figure 1. Overall performance of the proposed scheme: (a) Train on dataset 1, (b) inference on the same or another dataset. Keeping the pre-trained weights fixed, channel-wise scales and biases are applied to an intermediate activation map and are optimized to improve the predictions for the input dataset. The input dataset can be identical to the training dataset or a new one.
Sensors 22 09755 g001
Figure 2. Delaunay-based interpolation on a set of points. First, a Delaunay triangulation is carried out on the points. Then, a plane is fitted to each triangle, and finally, the value for points on each of them is obtained based on the fitted plane.
Figure 2. Delaunay-based interpolation on a set of points. First, a Delaunay triangulation is carried out on the points. Then, a plane is fitted to each triangle, and finally, the value for points on each of them is obtained based on the fitted plane.
Sensors 22 09755 g002
Figure 3. The double-stage adaptive refinement scheme (DARS). DARS refines depth maps estimated by baselines through optimizing (green arrows) an intermediate set of feature maps during inference (red arrows). Hence, given a pretrained baseline with fixed weights and an RGB input image, an initial depth D 0 d is estimated. In the first stage (correction), a Sliced Delaunay Correction (SDC) module corrects D 0 d with the guidance of a SD map D s . Afterwards, an optimizer module (PSO) tries to update the intermediate feature maps to minimize the distance between the dense corrected depth D ^ d and D s . As the scheme can be used for any deep architecture, the baseline (here, monodepth2) is illustrated minimally.
Figure 3. The double-stage adaptive refinement scheme (DARS). DARS refines depth maps estimated by baselines through optimizing (green arrows) an intermediate set of feature maps during inference (red arrows). Hence, given a pretrained baseline with fixed weights and an RGB input image, an initial depth D 0 d is estimated. In the first stage (correction), a Sliced Delaunay Correction (SDC) module corrects D 0 d with the guidance of a SD map D s . Afterwards, an optimizer module (PSO) tries to update the intermediate feature maps to minimize the distance between the dense corrected depth D ^ d and D s . As the scheme can be used for any deep architecture, the baseline (here, monodepth2) is illustrated minimally.
Sensors 22 09755 g003
Figure 4. (a) Delaunay correction (DC), and (b) sliced Delaunay correction (SDC). In DC, the Delaunay-based interpolation is conducted on the whole sparse correction map. While in SDC, each SD map is first divided into three slices with overlap, then the correction value ( Δ D s ) is interpolated using Delaunay-based interpolation in each of them, independently. In the overlapped areas, the average of the values from the two slices is taken.
Figure 4. (a) Delaunay correction (DC), and (b) sliced Delaunay correction (SDC). In DC, the Delaunay-based interpolation is conducted on the whole sparse correction map. While in SDC, each SD map is first divided into three slices with overlap, then the correction value ( Δ D s ) is interpolated using Delaunay-based interpolation in each of them, independently. In the overlapped areas, the average of the values from the two slices is taken.
Sensors 22 09755 g004
Figure 5. Flow chart of PSO in DARS, where position matrix X contains scales and biases. The theoretical background of PSO is provided in Section 3.2.
Figure 5. Flow chart of PSO in DARS, where position matrix X contains scales and biases. The theoretical background of PSO is provided in Section 3.2.
Sensors 22 09755 g005
Figure 6. A sample of the used datasets.
Figure 6. A sample of the used datasets.
Sensors 22 09755 g006
Figure 7. Visual results related to the ablation study of KITTI dataset. Numbers on the right side of error patterns are in meters.
Figure 7. Visual results related to the ablation study of KITTI dataset. Numbers on the right side of error patterns are in meters.
Sensors 22 09755 g007
Figure 8. Visual results related to ablation study of NYUv2 dataset. Numbers on the right side of error patterns are in meters.
Figure 8. Visual results related to ablation study of NYUv2 dataset. Numbers on the right side of error patterns are in meters.
Sensors 22 09755 g008
Figure 9. Visual results related to the comparative study of the KITTI dataset. The results of Adabins [14], as the second best method, are brought. Also, numbers on the right side of error patterns are in meters.
Figure 9. Visual results related to the comparative study of the KITTI dataset. The results of Adabins [14], as the second best method, are brought. Also, numbers on the right side of error patterns are in meters.
Sensors 22 09755 g009
Figure 10. Visual results related to comparative study on NYUv2 dataset. The results of Adabins [14], as the second best method, are brought. Also, numbers on the right side of error patterns are in meters.
Figure 10. Visual results related to comparative study on NYUv2 dataset. The results of Adabins [14], as the second best method, are brought. Also, numbers on the right side of error patterns are in meters.
Sensors 22 09755 g010
Table 1. Ablation study on KITTI and NYUv2. Monodepth2* is the version of Monodepth2 without median scaling.
Table 1. Ablation study on KITTI and NYUv2. Monodepth2* is the version of Monodepth2 without median scaling.
DatasetModulesLower Is BetterHigher Is Better
BaselineCorrectionOptimizerAbsRelRMSE RMSE log δ 1.25 δ 1 . 25 2 δ 1 . 25 3
KITTIMonodepth2--0.0903.9420.1370.9140.9830.995
Monodepth2*--0.99619.3245.7150.0000.0000.000
Monodepth2*DC-0.86416.8883.1490.1830.3300.447
Monodepth2*SDC-0.0461.6760.0910.9760.9910.995
Monodepth2*SDCL-BFGS0.0461.6760.0910.9760.9910.995
Monodepth2*SDCPSO0.0241.4400.0710.9850.9930.996
NYUv2Monodepth2*SDC-0.0180.7660.7470.9720.9740.975
Monodepth2*SDCL-BFGS0.0180.7660.7470.9720.9740.975
Monodepth2*SDCPSO0.0170.1090.0440.9930.9960.999
Table 2. Comparative study on KITTI. The first part from above contains unsupervised methods while the second part is dedicated to supervised ones.
Table 2. Comparative study on KITTI. The first part from above contains unsupervised methods while the second part is dedicated to supervised ones.
MethodLower Is BetterHigher Is Better
AbsRelSqRelRMSE RMSE log δ 1.25 δ 1 . 25 2 δ 1 . 25 3
[49]0.1200.7894.7550.1770.8560.9610.987
[28]0.1320.9945.2400.1930.8330.9530.985
[26]0.1140.8984.9350.2060.8610.9490.976
[23]0.0900.5453.9420.1370.9140.9830.995
[50]0.0900.4243.4190.1330.9160.9840.996
[51]0.0600.2312.6420.0940.9580.9940.999
[14]0.0580.1902.3600.0880.9640.9950.999
DARS0.0240.1371.4420.0710.9850.9930.996
Table 3. Comparative study on NYUv2.The first part from above contains unsupervised methods while the second part is dedicated to supervised ones.
Table 3. Comparative study on NYUv2.The first part from above contains unsupervised methods while the second part is dedicated to supervised ones.
MethodLower Is BetterHigher Is Better
AbsRelRMSE RMSE log δ 1.25 δ 1 . 25 2 δ 1 . 25 3
[47]0.1580.641-0.7690.9500.988
[15]0.1100.3920.0470.8850.9780.994
[51]0.1070.3730.0460.8930.9850.997
[14]0.1030.3640.0440.9030.9840.997
DARS0.0170.1090.0440.9930.9960.999
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alizadeh Naeini, A.; Sheikholeslami, M.M.; Sohn, G. An Adaptive Refinement Scheme for Depth Estimation Networks. Sensors 2022, 22, 9755. https://doi.org/10.3390/s22249755

AMA Style

Alizadeh Naeini A, Sheikholeslami MM, Sohn G. An Adaptive Refinement Scheme for Depth Estimation Networks. Sensors. 2022; 22(24):9755. https://doi.org/10.3390/s22249755

Chicago/Turabian Style

Alizadeh Naeini, Amin, Mohammad Moein Sheikholeslami, and Gunho Sohn. 2022. "An Adaptive Refinement Scheme for Depth Estimation Networks" Sensors 22, no. 24: 9755. https://doi.org/10.3390/s22249755

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop