You are currently viewing a new version of our website. To view the old version click .
Symmetry
  • Article
  • Open Access

15 May 2022

Adaptive Superpixel-Based Disparity Estimation Algorithm Using Plane Information and Disparity Refining Mechanism in Stereo Matching

and
Graduate Institute of Communication Engineering, National Taiwan University, No. 1, Sec. 4, Roosevelt Road, Taipei 10617, Taiwan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Selected Papers from IIKII 2021 Conferences

Abstract

The motivation of this paper is to address the limitations of the conventional keypoint-based disparity estimation methods. Conventionally, disparity estimation is usually based on the local information of keypoints. However, keypoints may distribute sparsely in the smooth region, and keypoints with the same descriptors may appear in a symmetric pattern. Therefore, conventional keypoint-based disparity estimation methods may have limited performance in smooth and symmetric regions. The proposed algorithm is superpixel-based. Instead of performing keypoint matching, both keypoint and semiglobal information are applied to determine the disparity in the proposed algorithm. Since the local information of keypoints and the semi-global information of the superpixel are both applied, the accuracy of disparity estimation can be improved, especially for smooth and symmetric regions. Moreover, to address the non-uniform distribution problem of keypoints, a disparity refining mechanism based on the similarity and the distance of neighboring superpixels is applied to correct the disparity of the superpixel with no or few keypoints. The experiments show that the disparity map generated by the proposed algorithm has a lower matching error rate than that generated by other methods.

1. Introduction

In image processing, disparity estimation is to determine the difference of object locations in two images; it is an important technique in distance information retrieval, image stitching, and stereoscopic image processing. It is critical for entertainment, robotic mapping, navigation, object recognition, the advanced driver assistance system, 3D modeling, gesture recognition, etc. Therefore, how to accurately generate the disparity map of two stereo images (i.e., a pair of images are generated at the same time with a known camera distance) becomes a popular, interesting, and exciting research topic.
The disparity map is to represent the number of displacements for each of the pixels in an image pair. Usually, disparity estimation is performed by detecting the matching keypoints between two stereo images. Then, the distance from the object to the camera can be estimated from the disparity. The disparity value is nearly inversely proportional to the scene depth for each corresponding pixel location.
The main motivation of this work is to address the limitation of conventional disparity estimation methods, which are usually based on keypoint matching. Keypoints may distribute sparsely in a smooth region. Moreover, in an image, there are usually many symmetric patterns, and keypoints with very similar descriptors may appear repeatedly in a symmetric region. These problems may lead to conventional keypoint-based methods having limited performance in smooth and symmetric regions.
In this paper, an advanced disparity estimation algorithm that applies both point and semi-global information is proposed. Instead of performing only keypoint-by-keypoint matching, we also apply superpixel information. Since the local information of keypoints and the semi-global information of superpixels are both applied, a more accurate disparity map can be generated, especially for smooth, textural, and symmetric regions. Moreover, several disparity refinement mechanisms are also developed. We use the disparities of neighboring superpixels to correct the disparities of the superpixels that have no or few keypoints. The disparity is refined adaptively according to the surrounding information and the distance of the superpixel centroids. An accurate estimation result can be achieved by the proposed algorithm, because it uses SIFT keypoint-based features, plane-based superpixel information, and the hierarchical fine-tune mechanism based on the Euclidean distance between the superpixel gravities and the weighted coefficients from neighboring information.
In Section 2, a review on related work is presented. Then, the proposed algorithm and the adopted techniques are illustrated in detail in Section 3. In Section 4, the experimental results of the proposed algorithm on the popular Middlebury database [1] are shown. Finally, a conclusion is provided in Section 5.

3. Proposed Algorithm

In the proposed algorithm, the disparity map is generated by both local and semi-global features. Local information is integrated into the semiglobal one using the entropy rate superpixel (ERS) [31]. Furthermore, after generating the disparity map by the ERS, the result is further optimized by disparity fusion, extension, and refinement. After applying both the pixel-based and the superpixel-based disparities together with several adjusting mechanisms, the final disparity estimation result for a stereoscopic image pair is created [32]. The flowchart of the proposed algorithm is shown in Figure 1. The details of (i) “SIFT keypoint extraction”, (ii) “local-matching cost function”, and (iii) “pixel-based disparity value” are described in Section 3.1. The details of (iv) “ERS segmentation” and (v) “plane-based disparity map” are described in Section 3.2. The details of (vi) “disparity fusion map”, (vii) “disparity extended map”, and (viii) “disparity refinement” are described in Section 3.3.1, Section 3.3.2 and Section 3.3.3, respectively.
Figure 1. Process of the proposed disparity estimation algorithm.

3.1. Feature Extraction by Applying SIFT

The purpose of the SIFT (scale-invariant feature transform) [33] is to detect or describe local keypoints in an image. It includes multiscale difference of Gaussian (DoG) operations, keypoint localization, orientation determination, and descriptor generation. First, it detects the points of interest (keypoints). The image is convolved with multi-sized Gaussian functions, and the keypoints are taken as the local maxima or minima of the output of the DoG at multiple scales. Then, the keypoints that are noise-like or localized along an edge are excluded by a scoring mechanism. Then, each keypoint is assigned one or more orientations based on the direction of the local gradient. With the multi-scale and orientation assignment mechanisms, the SIFT matching process is robust to image location, scaling, and rotation. Finally, the descriptor with 128 elements is created by the histogram on 4 × 4-pixel neighboring patches with 8 quantized directions.
For a pair of stereo data, there are two images captured from the left-sided and right-sided viewpoints separately. After SIFT keypoint extraction, each keypoint on the left-sided image is applied to match the keypoint on the right-sided image. The matching cost function θ is defined as:
θ ( L m ,   R n ) = cos 1 ( L m · R n ) k = arg min m θ ( L m ,   R n ) M a t c h ( m , n ) = { 1 , i f m = k a n d θ ( L m ,   R n ) T 0 , o t h e r w i s e
where Lm and Rn are the normalized descriptor vectors of the mth and the nth SIFT keypoints from the input L and the reference image R, respectively. The operator (●) means the dot product. From (1), the local matching cost function is designed as the dot product of the descriptor vectors between the left and the right images, respectively. If the arccosine of the dot product of two descriptor vectors is lower than a threshold T, then the corresponding two keypoints are treated as a candidate of a matched keypoint pair. Otherwise, these keypoints are considered to be mismatched.
Note that, in (1), if (LmRn) is near to 1, then θ(Lm, Rn) has a smaller value. If:
   θ(Lk, Rn) < θ(Lk, Rm) for all mn and
θ(Lk, Rn) ≤ threshold T = 0.6,
then the kth SIFT keypoint of the input image is considered to match the nth SIFT keypoint of the reference image.
After the matched keypoint pair is determined, the disparity can be calculated by the absolute value of the horizontal coordinate difference in the two matched keypoints.
The SIFT provides an effective way to determine the disparity. However, since SIFT keypoints only reflect local information, we apply several techniques, including the ERS and the refinement mechanism, to integrate local information into semi-global information and perform disparity estimation more accurately.

3.2. Entropy Rate Superpixels for Semi-Global-Based Disparity

ERS segmentation [31] adopts an objective function consisting of two parts: the entropy rate of the random walk on a graph and a balancing term. The entropy rate is helpful to make each region homogeneous, and the balancing term is to constrain the number of regions. Segmentation is executed by the graph that maximizes the objective function under the matroid constraint [31]. In this work, the ERS method is applied to segment the input image into many superpixels for disparity assignment and disparity map creation. First, the number of superpixels is chosen as Bs:
I m a x = max ( { I [ n ] | n = 1 , 2 , , T } ) , I m i n = min ( { I [ n ] | n = 1 , 2 , , T } ) B s = I m a x I m i n b
where I[n] is the pixel intensity, n is the index, and T is the total number of pixels. Then, b was chosen as 3 in our experiments. The idea is to adaptively determine the number of superpixels for each image according to the contrast of luminance.
Then, the ERSs that have at least one SIFT keypoint are chosen for disparity assignment. If there are more than one SIFT keypoint in the superpixel, then the disparity of the whole superpixel is determined by the average of the disparity values (i.e., davg) of the SIFT keypoints within this superpixel. The formula of disparity assignment for the superpixel S is:
d a v g [ S ] = i = 1 N d i / N , d i S ,
where N is the total number of the matched SIFT keypoints within the superpixel S, and di is the disparity value of the ith matched SIFT keypoint. With this process, the disparity of the superpixel with at least one matched SIFT point can be determined well.
Note that, after applying the above method, the disparity of the superpixel without any matched SIFT keypoint has not been determined yet. Moreover, some superpixels may have the disparity value quite different from that of the adjacent superpixels. Therefore, some adjusting methods are required to obtain an even more accurate disparity estimation result, which is described in the next subsection.

3.3. Disparity Optimization

Disparity optimization is to adjust the disparity value obtained by the method in Section 3.2 using the techniques of (i) depth fusion, (ii) depth extension, and (iii) depth refinement.
The goal of disparity fusion is to modify the disparity value of a superpixel by that of the neighboring superpixels. Its purpose is to remove unexpected or unreasonable disparity and increase the accuracy of disparity estimation. Moreover, after applying the method in Section 3.2, the superpixel without any SIFT keypoint (we call it the non-SIFT superpixel) does not have its own disparity value yet. To address this problem, the technique of disparity extension is applied to calculate the disparity of the non-SIFT superpixel. The most similar neighboring superpixel is chosen to generate the disparity value until all of the non-SIFT superpixels have been assigned a disparity value. Finally, a disparity refinement mechanism is performed to achieve an even more accurate disparity estimation result. Each of the adjusting techniques is illustrated in detail as follows.
In disparity optimization, the disparity values of the adjacent superpixels are adopted for disparity fusion, extension, and refinement. First, suppose that the coordinates of the pixels in the Sth superpixel are (x[i], y[i]), i = 1, 2, …, L, where L is the number of pixels in the Sth superpixel. Then, the center of gravity of superpixel S (denoted by xg[S], yg[S]) is determined from:
x g [ S ] = i = 1 L x i / L w h e r e x i S , y g [ S ] = i = 1 L y i / L w h e r e y i S .
Then, the Euclidean distance R of the adjacent superpixels S1 and S2 is calculated by using their centers of gravity:
R ( S 1 , S 2 ) = ( x g [ S 1 ] x g [ S 2 ] ) 2 + ( y g [ S 1 ] y g [ S 2 ] ) 2 .

3.3.1. Disparity Fusion

Disparity fusion is to apply a criterion to update the disparity value of a superpixel. Suppose that S1 and S2 are two adjacent superpixels. Then, the disparity value d of superpixel S1 is replaced by that of superpixel S2 if the following inequality is satisfied:
| d [ S 1 ] d [ S 2 ] | d [ S 1 ] > t h r e s h o l d = k 1 × ( 1 + N [ S 1 ] N [ S 2 ] N [ S 1 ] + N [ S 2 ] × k 2 ) ,
where d[Sj] is the disparity value, and N[Sj] (j = 1, 2) is the number of SIFT keypoints of superpixel Sj, respectively. The parameters k1 and k2 were chosen as 0.3 and 0.1, respectively, in our experiments. Note that the threshold in the right-hand side of (7) is adjusted according to the number of keypoints in S1 and S2. If N[S1] is larger than N[S2], the threshold value is increased. By contrast, the threshold is decreased if N[S1] is smaller than N[S2]. From (7), one can see that the inequality is easier to be held if:
(i)
the difference of d[S1] and d[S2] is large;
(ii)
N[S1] is much smaller than N[S2].
In this case, the disparity of S1 is replaced by that of S2.

3.3.2. Disparity Extension

The goal of disparity extension is to populate the disparity value for all of the non-SIFT superpixels. Assume that:
{S1,n | n = 1, 2, , A} and {S2,m | m = 1, 2, , M}
are the sets of non-SIFT superpixels and the superpixels with matched SIFT keypoints (we call them SIFT-superpixels), respectively, and L and M are the total numbers of non-SIFT superpixels and SIFT superpixels, respectively.
Non-SIFT superpixels have not been assigned a disparity value because there is no matched SIFT keypoint as reference. We assign the disparity value of the non-SIFT superpixel by an iterative process. It is supposed that a more accurate disparity value can be determined by the non-SIFT superpixel with more neighboring SIFT superpixels. Therefore, we first process the non-SIFT superpixel with the largest value of Ω1,n where Ω1,n is the number of neighboring SIFT-superpixels of S1,n. In other words, if:
P = arg M a x n = 1 , 2 , , L { Ω 1 , n }
then the non-SIFT superpixel S1,P are processed before other non-SIFT superpixels. The above process is repeated until all non-SIFT superpixels have been processed.
To assign the disparity value on non-SIFT superpixel S1,P, the disparity values of the adjacent SIFT-superpixels should be applied. However, if there are more than one adjacent SIFT-superpixel, we choose the most proper one for disparity assignment. First, if the Euclidean distance between the centers of gravity of S1,P and S2,m is larger than a threshold, then the superpixel S2,m is not applied to determine the disparity of S1,P. That is, the adopted SIFT superpixel S2,m should satisfy that S2,m is adjacent to S1,P and that:
R(S1,P, S2,m) < k3,
where R is defined in (6). In our experiments, we set k3 = 100.
Second, if there are more matched SIFT keypoints in the superpixel S2,m and the Euclidean distance between S1,P and S2,m is small, then we tend to apply the disparity of S2,m to estimate the disparity of S1,P. That is, we first determine the score function as follows:
S C ( S 1 , P , , S 2 , m ) = N ( S 2 , m ) × s i m ( S 1 , P , S 2 , m )
where N(S2,m) is the number of matched SIFT keypoints in S2,m, and sim(S1,P, S2,m) means the similarity of S1,P and S2,m. For example, we can choose:
s i m ( S 1 , P , S 2 , m ) = R 0 R ( S 1 , P , S 2 , m )
where:
R 0 = M a x m = 1 , 2 , , M ( R ( S 1 , P , S 2 , m ) )
with this definition, if the distance of the centers of gravity of S1,P and S2,m is small, then sim(S1,P, S2,m) is large, which means that S2,m is definitely a valuable reference. Thus, one can use the following equation to assign the disparity value d for the superpixel S1,P:
d [ S 1 , P ] = d [ S 2 , v ] i f v = arg M a x m = 1 , 2 , , M { S C ( S 1 , P , S 2 , m ) }

3.3.3. Disparity Refinement

The process of disparity refinement is almost the same as that of disparity fusion. The difference is that it is placed after disparity fusion and disparity extension. It further updates the previously determined disparity value of each superpixel iteratively to achieve a lower disparity estimation error.

4. Evaluation and Discussion

The proposed algorithm was evaluated on the Middlebury stereo dataset [1]. The images from this dataset are piecewise planar scenes. The Middlebury stereo dataset contains stereoscopic test images and the ground-truth disparity map for each image whose raw data is to be scaled or divided by a factor of 4 if it is translated into the disparity values.
Then, the Venus image from [1] was applied to show the intermediate results of the proposed disparity estimation algorithm.
First, the results of SIFT keypoint extraction and local matching are shown in the left part of Figure 2. The superpixel segmentation result using the ERS is shown in the right part of Figure 2.
Figure 2. (Left) Results of extracting SIFT keypoints, which are marked by blue x. (Right) Superpixel generation result for the stereoscopic image of Venus.
In the steps of disparity generation and assignment in Figure 1, exploiting SIFT keypoint matching and the ERS in stereo images can obtain pixel- and superpixel-based disparity values, respectively. The result of using these techniques for disparity map generation is shown in the left part of Figure 3. In this step, most superpixels were assigned a disparity value. However, some superpixels (i.e., non-SIFT superpixels) are not yet assigned a disparity value. The problem was later solved by using the technique of disparity extension.
Figure 3. (Left) Disparity calculated using the SIFT keypoints and the superpixel-based assignment method. (Right) The disparity error map determined from the left subfigure.
In the right subfigure of Figure 3, the disparity estimation error (denoted by derror[i]) of the left subfigure is shown:
d e r r o r [ i ] = | d e [ i ] d g t [ i ] |
where dgt[i] is the disparity value in the ground truth, de[i] is the estimated disparity of the left subfigure in Figure 3, and i denotes the ith pixel in the disparity map.
Then, to further improve the accuracy of disparity estimation, the technique of disparity fusion was applied; its result is shown in Figure 4.
Figure 4. (Left) The result after applying disparity fusion, where the symbol o with red color indicates the gravity of each superpixel. (Right) The corresponding disparity error.
To assign the disparity value of the non-SIFT superpixel, the technique of disparity extension was applied. Its result is shown in Figure 5.
Figure 5. (Left) The result after applying disparity extension. (Right) The corresponding disparity error.
Note that the process of disparity fusion can much improve the accuracy of disparity estimation for SIFT-superpixels, as in Figure 4. By contrast, disparity extension much improves the performance of disparity estimation for non-SIFT superpixels, as in Figure 5.
Then, the technique of disparity refinement was applied to achieve better accuracy and a lower disparity estimation error. In this step, both the disparity values of SIFT superpixels and non-SIFT superpixels were updated and refined. The disparity error maps after the application of the disparity refinement technique and the final data processing step are given in Figure 6. Moreover, the disparity map of the Venus image in the ground truth and that generated by the proposed algorithm are shown in Figure 7. Furthermore, more results obtained by utilizing the proposed disparity estimation algorithm for other stereo images in the Middlebury dataset are also provided. The disparity map comparisons between the ground truth and the disparity estimation results for Tsukuba, Cones, and Teddy images are given in Figure 8, Figure 9 and Figure 10, respectively.
Figure 6. (Left) The disparity error map after applying disparity refinement. (Right) The final disparity error map using all of the processes in the proposed algorithm for Venus.
Figure 7. (Left) The disparity map in the ground truth and (Right) that generated from the proposed disparity estimation algorithm for Venus.
Figure 8. (Left) The disparity map in the ground truth. (Right) The disparity map generated by the proposed disparity estimation algorithm for Tsukuba.
Figure 9. (Left) The disparity map in the ground truth. (Right) The disparity map generated by the proposed disparity estimation algorithm for Cones.
Figure 10. (Left) The disparity map in the ground truth. (Right) The disparity map generated by the proposed disparity estimation algorithm for Teddy.
To evaluate the performance of the disparity estimation method quantitatively, the metric of the MER (matching error rate) was applied. The MER (unit: %) is also called the error matching rate or the matching error percentage. Its formula is as follows:
M E R = j = 1 T y [ j ] T × 100 y [ j ] = { 1 , i f | d e [ j ] d g t [ j ] | > d t h r e s 0 , o t h e r w i s e ,
where de is the estimated disparity value, and dgt is the ground truth, respectively; T is the total number of pixels in the stereo image, and i is the pixel index. Moreover, dthres is the error matching threshold for disparity estimation. It was chosen as 1 for the evaluation results of the proposed algorithm.
An accurate disparity estimation algorithm should have a lower value of the MER than other disparity estimation algorithms. In Table 2, the disparity estimation results of the proposed algorithm and other existing algorithms are shown. The stereoscopic color images Venus, Tsukuba, Cones, and Teddy, as well as their depth maps in the ground truth, were obtained from the Middlebury stereo dataset [1]. The evaluation results in Table 2 show that the proposed algorithm can achieve a lower disparity error and a more accurate disparity estimation result than other methods. It may be due to the fact that in the proposed algorithm, the local features extracted by SIFT keypoints and the semi-global information of superpixels are both adopted and that the hierarchical disparity correction mechanism based on the weighted similarity and the Euclidean distance of neighboring superpixels is applied to adjust the disparity for the superpixels with no or few keypoints.
Table 2. Comparison of the disparity estimation performance of the proposed algorithm and other methods in terms of the matching error rate (MER; unit: %) using the Middlebury stereo dataset.
Moreover, an ablation study was performed (its results are presented in Table 3) to show the effect of each of the proposed techniques. The 1st line in Table 3 shows that, when using SIFT feature extraction (pixelwise information) and ERS segmentation (superpixel-based information), the MERs of the four images were 2.36%, 8.81%, 7.93%, and 10.53%, respectively. In the following three lines, we present the results of the application of the techniques of disparity fusion, disparity extension, and disparity refinement, which were applied to further improve the disparity estimation results. The 2nd line shows that, with disparity fusion, the MERs were reduced by 35.6%, 19.6%, 16.9%, and 16.9% for Venus, Tsukuba, Cones, and Teddy images, respectively. The 3rd line shows that, with disparity extension, the MERs of the four images were reduced by 21.7%, 6.9%, 18.8%, and 18.9%, respectively. The last line shows that, with disparity refinement, the MERs were reduced by 18.5%, 38.2%, 41.8%, and 24.8%, respectively. The results in Table 3 show that all the proposed techniques are helpful for improving the performance of disparity estimation.
Table 3. Ablation study of each of the proposed techniques in the proposed disparity estimation algorithm tested on the Middlebury stereo dataset in terms of the MER (unit: %).
The proposed algorithm was implemented by MATLAB with Intel CPU i5 1.8 GHz and a 4 GB RAM. The computation time is summarized in Table 4. The complexity of the proposed method is O(NM), where M × N is the size of the input image.
Table 4. Computation time (using MATLAB with Intel CPU i5 1.8 GHz and 4 GB RAM).

5. Conclusions

A precise disparity estimation algorithm based on keypoint feature extraction, semi-global superpixel information, and post-adjusting mechanisms is proposed in this paper. In addition to SIFT keypoints, it also adopts the technique of ERS segmentation and several disparity adjustment techniques, including the integration of keypoint and superpixel information, the disparity extension scheme based on the Euclidean distance of superpixels, and the iterative refinement mechanism. The experimental results show that the proposed algorithm performs better than other methods. The ablation study also shows that all the proposed techniques are beneficial for improving the disparity estimation performance. With the proposed algorithm, a very accurate disparity estimation result can be obtained, which is helpful for depth estimation, virtual reality, image stitching, and stereoscopic image processing. Regarding future work, to further improve the proposed algorithm, we are looking forward to applying machine learning techniques to train the models for disparity extension. With the proper choice of superpixel features, the disparity can be determined more precisely with a support vector machine or a deep learning model. Moreover, proper polling mechanisms can also be applied for disparity fusion and to integrate the disparity of the keypoints within a superpixel well. Furthermore, with some modifications, the proposed idea of using superpixel-based information can also be applied in other disparity estimation problems, including the disparity estimation in the single-image scenario or the multi-view scenario.

Author Contributions

Methodology, C.-W.H. and J.-J.D.; software, C.-W.H.; writing—original draft, C.-W.H.; writing—review and editing, J.-J.D.; supervision, J.-J.D.; project administration, J.-J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by the Ministry of Science and Technology, Taiwan, grant number MOST 108-2221-E-002-041-MY3.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Scharstein, D.; Szeliski, R. High-accuracy stereo depth maps using structured light. In Proceedings of the IEEE Computer Society Conference in Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 1, pp. 195–202. [Google Scholar]
  2. Birchfield, S.; Tomasi, C. Depth discontinuities by pixel-to-pixel stereo. Int. J. Comput. Vis. 1999, 35, 269–293. [Google Scholar] [CrossRef]
  3. Anandan, P. A computational framework and an algorithm for the measurement of visual motion. Int. J. Comput. Vis. 1989, 2, 283–310. [Google Scholar] [CrossRef]
  4. Black, M.J.; Anandan, P. A framework for the robust estimation of optical flow. In Proceedings of the International Conference Computer Vision, Berlin, Germany, 11–14 May 1993; pp. 231–236. [Google Scholar]
  5. Shabanian, H.; Balasubramanian, M. A new hybrid stereo disparity estimation algorithm with guided image filtering-based cost aggregation. Symp. Electron. Imaging 2021, 33, 59-1–59-7. [Google Scholar] [CrossRef]
  6. Ma, N.; Men, Y.; Men, C.; Li, X. Accurate dense stereo matching based on image segmentation using an adaptive multi-cost approach. Symmetry 2016, 8, 159. [Google Scholar] [CrossRef]
  7. Baker, H.H. Edge based stereo correlation. In Proceedings of the DARPA Image Understanding Workshop, Palo Alto, CA, USA, April 1980; pp. 168–175. [Google Scholar]
  8. Baker, H.; Binford, T. Depth from Edge and Intensity Based Stereo; University of Illinois at Urbana-Champaign: Champaign, IL, USA, 1981. [Google Scholar]
  9. Birchfield, S.; Tomasi, C. A pixel dissimilarity measure that is insensitive to image sampling. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 401–406. [Google Scholar] [CrossRef] [Green Version]
  10. Birchfield, S.; Tomasi, C. Multiway cut for stereo and motion with slanted surfaces. In Proceedings of the International Conference Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 1, pp. 489–495. [Google Scholar]
  11. Baker, S.; Szeliski, R.; Anandan, P. A layered approach to stereo reconstruction. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, Santa Barbara, CA, USA, 23–25 June 1998; pp. 434–444. [Google Scholar]
  12. Barnard, S.T. Stochastic stereo matching over scale. Int. J. Comput. Vis. 1989, 3, 17–32. [Google Scholar] [CrossRef]
  13. Nguyen, P.H.; Ahn, C.W. Stereo matching methods for imperfectly rectified stereo images. Symmetry 2019, 11, 570. [Google Scholar] [CrossRef] [Green Version]
  14. Ttofis, C.; Kyrkou, C.; Theocharides, T. A low-cost real-time embedded stereo vision system for accurate disparity estimation based on guided image filtering. IEEE Trans. Comput. 2016, 65, 2678–2693. [Google Scholar] [CrossRef]
  15. Arnold, R.D. Automated Stereo Perception. Ph.D. Thesis, Artificial Intelligence Laboratory; Stanford University, Stanford, CA, USA, 1983. [Google Scholar]
  16. San, T.T.; War, N. Feature based disparity estimation using hill-climbing algorithm. In Proceedings of the International Conference Software Engineering Research, London, UK, 7–9 June 2017; pp. 129–133. [Google Scholar]
  17. Yao, C.; Jia, Y.; Di, H.; Li, P.; Wu, Y. A decomposition model for stereo matching. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 6091–6100. [Google Scholar]
  18. Zhang, Z.; Qiao, J.; Lin, S. A semi-supervised monocular stereo matching method. Symmetry 2019, 11, 690. [Google Scholar] [CrossRef] [Green Version]
  19. Wang, Y.; Wang, L.; Wu, G.; Yang, J.; An, W.; Yu, J.; Guo, Y. Disentangling light fields for super-resolution and disparity estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 1, 1. [Google Scholar] [CrossRef] [PubMed]
  20. Chen, L.; Sun, J.; Xie, Y.; Zhang, S.; Shuai, Q.; Jiang, Q.; Zhang, G.; Bao, H.; Zhou, X. Shape Prior Guided Instance Disparity Estimation for 3D Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 1, 1. [Google Scholar] [CrossRef] [PubMed]
  21. Xu, J.; He, W.; Tian, Z. Stereo matching based on improved matching cost calculation and weighted guided filtering. In Proceedings of the International Conference Communications and Networking, Hangzhou, China, 20–21 November 2020; pp. 490–504. [Google Scholar]
  22. Zhang, K.; Fang, Y.; Min, D.; Sun, L.; Yang, S.; Yan, S.; Tian, Q. Cross-scale cost aggregation for stereo matching. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1590–1597. [Google Scholar]
  23. Hirschmuller, H. Accurate and efficient stereo processing by semi-global matching and mutual information. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 807–814. [Google Scholar]
  24. Ouyang, Q.; Du, L.; Cao, C. A new disparity selection method based on extreme-point extraction. In Proceedings of the Chinese Automation Congress, Shanghai, China, 6–8 November 2020; pp. 847–852. [Google Scholar]
  25. Humenberger, M.; Engelke, T.; Kubinger, W. A census-based stereo vision algorithm using modified semi-global matching and plane fitting to improve matching quality. In Proceedings of the IEEE Computer Society Conference in Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 77–84. [Google Scholar]
  26. Stankiewicz, O.; Wegner, K. Depth map estimation software version 2. In Proceedings of the ISO/IEC MPEG Meeting, Archamps, France, 27 April 2008. [Google Scholar]
  27. Zhong, Y.; Dai, Y.; Li, H. Self-supervised learning for stereo matching with self-improving ability. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1–13. [Google Scholar]
  28. Yamaguchi, K.; McAllester, D.; Urtasun, R. Efficient joint segmentation, occlusion labeling, stereo and flow estimation. In Proceedings of the European Conference Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 756–771. [Google Scholar]
  29. Zbontar, J.; LeCun, Y. Stereo matching by training a convolutional neural network to compare image patches. J. Mech. Learn. Res. 2016, 17, 2287–2318. [Google Scholar]
  30. Zhang, C.; Li, Z.; Cheng, Y.; Cai, R.; Chao, H.; Rui, Y. Meshstereo: A global stereo model with mesh alignment regularization for view interpolation. In Proceedings of the International Conference Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2057–2065. [Google Scholar]
  31. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the IEEE Computer Society Conference Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  32. Huang, C.W.; Ding, J.J. Superpixel and plane information based disparity estimation algorithm in stereo matching. In Proceedings of the IEEE Eurasia Conference IOT, Communication and Engineering, Taiwan, China, 29–31 October 2021; pp. 397–400. [Google Scholar]
  33. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the IEEE International Conference Computer Vision, Corfu, Greece, 20–25 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.