Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation
Abstract
1. Introduction
2. Related Work
2.1. 3D Scanning Technologies
2.2. Main Features of Stereo Matching Technology
2.3. Datasets
2.3.1. Datasets with Dense Ground Truth
2.3.2. Datasets with Sparse Ground Truth
2.3.3. Synthetically Rendered Datasets
2.3.4. Other Datasets
3. Materials and Methods
- Estimating the quality of calibration.
- Determining the extent of occluded areas.
- Verifying the completeness of ground truth.
3.1. Estimating the Quality of Calibration
- Identifying keypoints in both the left and right images.
- Matching keypoints.
- Selecting keypoints.
- Calculating an average discrepancy.
3.2. The POA Metric
3.3. CPL Metric
4. Results
4.1. Evaluation of the Middlebury Dataset
4.2. Evaluation of ETH3D
4.3. Evaluation of Booster
4.4. Evaluation of PlantStereo
4.5. Evaluation of KITTI
4.6. Evaluation of MS2
4.7. Evaluation of Synthetic Datasets
4.8. Evaluation Summary
5. Discussion
6. Conclusions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
| BRIEF | Binary Robust Independent Elementary Features |
| BRISK | Binary Robust Invariant Scalable Keypoints |
| GT | ground truth |
| LIDAR | Light Detection and Ranging |
| NIR | Stereo Near Infrared |
| NLP | Natural Language Processing |
| SIFT | Scale-Invariant Feature Transform |
| SLAM | Simultaneous Localization and Mapping |
| SURF | Speeded Up Robust Features |
References
- Guo, X.; Yang, K.; Yang, W.; Wang, X.; Li, H. Group-wise Correlation Stereo Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3273–3282. [Google Scholar]
- Knöbelreiter, P.; Sormann, C.; Shekhovtsov, A.; Fraundorfer, F.; Pock, T. Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems. In Proceedings of the The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
- Zbontar, J.; LeCun, Y. Stereo matching by training a convolutional neural network to compare image patches. J. Mach. Learn. Res. 2016, 17, 2287–2318. [Google Scholar]
- Scharstein, D.; Szeliski, R. A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms. Int. J. Comput. Vis. 2002, 47, 7–42. [Google Scholar] [CrossRef]
- Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar] [CrossRef]
- Kaczmarek, A.L. A Method for Adapting Stereo Matching Algorithms to Real Environments. Appl. Sci. 2025, 15, 4070. [Google Scholar] [CrossRef]
- Kaczmarek, A.L.; Blaschitz, B. Equal Baseline Camera Array—Calibration, Testbed and Applications. Appl. Sci. 2021, 11, 8464. [Google Scholar] [CrossRef]
- Błaszczak-Bąk, W.; Kamiński, W.; Bednarczyk, M.; Suchocki, C.; Masiero, A. Real-Time DTM Generation with Sequential Estimation and OptD Method. Appl. Sci. 2025, 15, 4068. [Google Scholar] [CrossRef]
- Zhang, Z.; Wei, C.; Wu, G.; Barth, M.J. Vulnerable Road User Detection for Roadside-Assisted Safety Protection: A Comprehensive Survey. Appl. Sci. 2025, 15, 3797. [Google Scholar] [CrossRef]
- Gupta, M.; Yin, Q.; Nayar, S.K. Structured Light in Sunlight. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 545–552. [Google Scholar] [CrossRef]
- Seitz, S.M.; Curless, B.; Diebel, J.; Scharstein, D.; Szeliski, R. A Comparison and Evaluation of Multi-View Stereo Reconstruction Algorithms. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 1, pp. 519–528. [Google Scholar] [CrossRef]
- Middlebury Stereo Evaluation-Version 3. Available online: https://vision.middlebury.edu/stereo/eval3/ (accessed on 17 October 2025).
- KITTI Vision Benchmark Suite. Available online: https://www.cvlibs.net/datasets/kitti/ (accessed on 17 October 2025).
- Kaczmarek, A.L. 3D Vision System for a Robotic Arm Based on Equal Baseline Camera Array. J. Intell. Robot. Syst. 2019, 99, 13–28. [Google Scholar] [CrossRef]
- Bradski, D.G.R.; Kaehler, A. Learning Opencv, 1st ed.; O’Reilly Media, Inc.: Newton, MA, USA, 2008. [Google Scholar]
- Scharstein, D.; Hirschmüller, H.; Kitajima, Y.; Krathwohl, G.; Nešić, N.; Wang, X.; Westling, P. High-Resolution Stereo Datasets with Subpixel-Accurate Ground Truth. In Proceedings of the Pattern Recognition, GCPR 2014, Münster, Germany, 2–5 September 2014; Jiang, X., Hornegger, J., Koch, R., Eds.; Springer: Cham, Switzerland, 2014; pp. 31–42. [Google Scholar]
- Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets Robotics: The KITTI Dataset. Int. J. Robot. Res. (IJRR) 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
- Schöps, T.; Schönberger, J.L.; Galliani, S.; Sattler, T.; Schindler, K.; Pollefeys, M.; Geiger, A. A Multi-View Stereo Benchmark with High-Resolution Images and Multi-Camera Videos. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Zama Ramirez, P.; Tosi, F.; Poggi, M.; Salti, S.; Di Stefano, L.; Mattoccia, S. Open Challenges in Deep Stereo: The Booster Dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. CVPR. [Google Scholar]
- Ramirez, P.Z.; Costanzino, A.; Tosi, F.; Poggi, M.; Salti, S.; Mattoccia, S.; Stefano, L.D. Booster: A Benchmark for Depth from Images of Specular and Transparent Surfaces. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 85–102. [Google Scholar] [CrossRef] [PubMed]
- NTIRE 2025: HR Depth from Images of Specular and Transparent Surfaces. Available online: https://cvlab-unibo.github.io/booster-web/ntire25.html (accessed on 17 October 2025).
- Wang, Q.; Wu, D.; Liu, W.; Lou, M.; Jiang, H.; Ying, Y.; Zhou, M. PlantStereo: A High Quality Stereo Matching Dataset for Plant Reconstruction. Agriculture 2023, 13, 330. [Google Scholar] [CrossRef]
- Bao, W.; Wang, W.; Xu, Y.; Guo, Y.; Hong, S.; Zhang, X. InStereo2K: A large real dataset for stereo matching in indoor scenes. Sci. China Inf. Sci. 2020, 63, 212101. [Google Scholar] [CrossRef]
- Hua, Y.; Kohli, P.; Uplavikar, P.; Ravi, A.; Gunaseelan, S.; Orozco, J.; Li, E. Holopix50k: A Large-Scale In-the-wild Stereo Image Dataset. arXiv 2020, arXiv:2003.11172. [Google Scholar]
- Wang, Y.; Wang, L.; Yang, J.; An, W.; Guo, Y. Flickr1024: A Large-Scale Dataset for Stereo Image Super-Resolution. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 3852–3857. [Google Scholar] [CrossRef]
- Wang, C.; Lucey, S.; Perazzi, F.; Wang, O. Web Stereo Video Supervision for Depth Prediction from Dynamic Scenes. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019; pp. 348–357. [Google Scholar] [CrossRef]
- Shin, U.; Park, J.; Kweon, I.S. Deep Depth Estimation From Thermal Image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1043–1053. [Google Scholar]
- Zhu, A.Z.; Thakur, D.; Özaslan, T.; Pfrommer, B.; Kumar, V.; Daniilidis, K. The Multivehicle Stereo Event Camera Dataset: An Event Camera Dataset for 3D Perception. IEEE Robot. Autom. Lett. 2018, 3, 2032–2039. [Google Scholar] [CrossRef]
- Yang, G.; Song, X.; Huang, C.; Deng, Z.; Shi, J.; Zhou, B. DrivingStereo: A Large-Scale Dataset for Stereo Matching in Autonomous Driving Scenarios. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 899–908. [Google Scholar] [CrossRef]
- Chang, M.F.; Lambert, J.W.; Sangkloy, P.; Singh, J.; Bak, S.; Hartnett, A.; Wang, D.; Carr, P.; Lucey, S.; Ramanan, D.; et al. Argoverse: 3D Tracking and Forecasting with Rich Maps. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Wilson, B.; Qi, W.; Agarwal, T.; Lambert, J.; Singh, J.; Khandelwal, S.; Pan, B.; Kumar, R.; Hartnett, A.; Pontes, J.K.; et al. Argoverse 2: Next Generation Datasets for Self-driving Perception and Forecasting. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks (NeurIPS Datasets and Benchmarks 2021), Virtual Conference, 5–14 December 2021. [Google Scholar]
- Lambert, J.; Hays, J. Trust, but Verify: Cross-Modality Fusion for HD Map Change Detection. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks (NeurIPS Datasets and Benchmarks 2021), Virtual Conference, 5–14 December 2021. [Google Scholar]
- Sivaprakasam, M.; Maheshwari, P.; Castro, M.G.; Triest, S.; Nye, M.; Willits, S.; Saba, A.; Wang, W.; Scherer, S. TartanDrive 2.0: More Modalities and Better Infrastructure to Further Self-Supervised Learning Research in Off-Road Driving Tasks. arXiv 2024, arXiv:2402.01913. [Google Scholar] [CrossRef]
- Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 Year, 1000km: The Oxford RobotCar Dataset. Int. J. Robot. Res. (IJRR) 2017, 36, 3–15. [Google Scholar] [CrossRef]
- Maddern, W.; Pascoe, G.; Gadd, M.; Barnes, D.; Yeomans, B.; Newman, P. Real-time Kinematic Ground Truth for the Oxford RobotCar Dataset. arXiv 2020, arXiv:2002.10152. [Google Scholar]
- Burri, M.; Nikolic, J.; Gohl, P.; Schneider, T.; Rehder, J.; Omari, S.; Achtelik, M.W.; Siegwart, R. The EuRoC micro aerial vehicle datasets. Int. J. Robot. Res. 2016, 35, 1157–1163. [Google Scholar] [CrossRef]
- Zhao, T.; He, J.; Lv, J.; Min, D.; Wei, Y. A Comprehensive Implementation of Road Surface Classification for Vehicle Driving Assistance: Dataset, Models, and Deployment. IEEE Trans. Intell. Transp. Syst. 2023, 24, 8361–8370. [Google Scholar] [CrossRef]
- Moravec, J.; Šára, R. High-recall calibration monitoring for stereo cameras. Pattern Anal. Appl. 2024, 27, 41. [Google Scholar] [CrossRef]
- Wen, B.; Trepte, M.; Aribido, J.; Kautz, J.; Gallo, O.; Birchfield, S. FoundationStereo: Zero-Shot Stereo Matching. arXiv 2025, arXiv:2501.09898. [Google Scholar]
- Tremblay, J.; To, T.; Birchfield, S. Falling Things: A Synthetic Dataset for 3D Object Detection and Pose Estimation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Los Alamitos, CA, USA, 18–22 June 2018; pp. 2119–21193. [Google Scholar] [CrossRef]
- Mayer, N.; Ilg, E.; Häusser, P.; Fischer, P.; Cremers, D.; Dosovitskiy, A.; Brox, T. A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Mehl, L.; Schmalfuss, J.; Jahedi, A.; Nalivayko, Y.; Bruhn, A. Spring: A High-Resolution High-Detail Dataset and Benchmark for Scene Flow, Optical Flow and Stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
- Schmalfuss, J.; Oei, V.; Mehl, L.; Bartsch, M.; Agnihotri, S.; Keuper, M.; Bruhn, A. RobustSpring: Benchmarking Robustness to Image Corruptions for Optical Flow, Scene Flow and Stereo. arXiv 2025, arXiv:2505.09368. [Google Scholar] [CrossRef]
- Butler, D.J.; Wulff, J.; Stanley, G.B.; Black, M.J. A Naturalistic Open Source Movie for Optical Flow Evaluation. In Proceedings of the Computer Vision–ECCV 2012, Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 611–625. [Google Scholar]
- Cabon, Y.; Murray, N.; Humenberger, M. Virtual KITTI 2. arXiv 2020, arXiv:2001.10773. [Google Scholar] [CrossRef]
- Gaidon, A.; Wang, Q.; Cabon, Y.; Vig, E. Virtual worlds as proxy for multi-object tracking analysis. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4340–4349. [Google Scholar]
- Deschaud, J. KITTI-CARLA: A KITTI-like dataset generated by CARLA Simulator. arXiv 2021, arXiv:2109.00892. [Google Scholar] [CrossRef]
- Dosovitskiy, A.; Ros, G.; Codevilla, F.; López, A.M.; Koltun, V. CARLA: An Open Urban Driving Simulator. arXiv 2017, arXiv:1711.03938. [Google Scholar] [CrossRef]
- Wang, Q.; Zheng, S.; Yan, Q.; Deng, F.; Zhao, K.; Chu, X. IRS: A Large Naturalistic Indoor Robotics Stereo Dataset to Train Deep Models for Disparity and Surface Normal Estimation. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Karaev, N.; Rocco, I.; Graham, B.; Neverova, N.; Vedaldi, A.; Rupprecht, C. DynamicStereo: Consistent Dynamic Depth from Stereo Videos. arXiv 2023, arXiv:2305.02296. [Google Scholar] [CrossRef]
- Straub, J.; Whelan, T.; Ma, L.; Chen, Y.; Wijmans, E.; Green, S.; Engel, J.J.; Mur-Artal, R.; Ren, C.Y.; Verma, S.; et al. The Replica Dataset: A Digital Replica of Indoor Spaces. arXiv 2019, arXiv:1906.05797. [Google Scholar] [CrossRef]
- Fisher, R. CVonline: Image Databases. Available online: http://homepages.inf.ed.ac.uk/rbf/CVonline/Imagedbase.htm (accessed on 17 October 2025).
- Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor Segmentation and Support Inference from RGBD Images. In Proceedings of the Computer Vision–ECCV 2012, Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
- Huang, X.; Wang, P.; Cheng, X.; Zhou, D.; Geng, Q.; Yang, R. The ApolloScape Open Dataset for Autonomous Driving and Its Application. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2702–2719. [Google Scholar] [CrossRef]
- Wang, S.; Bai, M.; Máttyus, G.; Chu, H.; Luo, W.; Yang, B.; Liang, J.; Cheverie, J.; Fidler, S.; Urtasun, R. TorontoCity: Seeing the World with a Million Eyes. arXiv 2016, arXiv:1612.00423. [Google Scholar] [CrossRef]
- Kondermann, D.; Nair, R.; Honauer, K.; Krispin, K.; Andrulis, J.; Brock, A.; Güssefeld, B.; Rahimimoghaddam, M.; Hofmann, S.; Brenner, C.; et al. The HCI Benchmark Suite: Stereo and Flow Ground Truth with Uncertainties for Urban Autonomous Driving. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 19–28. [Google Scholar] [CrossRef]
- Li, J.; Wang, P.; Xiong, P.; Cai, T.; Yan, Z.; Yang, L.; Liu, J.; Fan, H.; Liu, S. Practical Stereo Matching via Cascaded Recurrent Network with Adaptive Correlation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16242–16251. [Google Scholar] [CrossRef]
- Yang, G.; Manela, J.; Happold, M.; Ramanan, D. Hierarchical Deep Stereo Matching on High-Resolution Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5510–5519. [Google Scholar] [CrossRef]
- Camera Calibration and 3D Reconstruction-OpenCV 4.13.0 Documentation. Available online: https://docs.opencv.org/4.x/d9/d0c/group__calib3d.html (accessed on 17 October 2025).
- Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
- Bay, H.; Ess, A.; Tuytelaars, T.; Gool, L.V. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
- Liu, R.; Zhang, H.; Liu, M.; Xia, X.; Hu, T. Stereo Cameras Self-Calibration Based on SIFT. In Proceedings of the 2009 International Conference on Measuring Technology and Mechatronics Automation, Zhangjiajie, China, 11–12 April 2009; Volume 1, pp. 352–355. [Google Scholar] [CrossRef]
- Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary Robust invariant scalable keypoints. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar] [CrossRef]
- Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P. BRIEF: Binary Robust Independent Elementary Features. In Proceedings of the Computer Vision–ECCV 2010, Crete, Greece, 5–11 September 2010; Daniilidis, K., Maragos, P., Paragios, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 778–792. [Google Scholar]
- Alcantarilla, P.F.; Nuevo, J.; Bartoli, A. Fast Explicit Diffusion for Accelerated Features in Nonlinear Scale Spaces. In Proceedings of the British Machine Vision Conference, Bristol, UK, 9–13 September 2013. [Google Scholar]
- Karami, E.; Prasad, S.; Shehata, M. Image Matching Using SIFT, SURF, BRIEF and ORB: Performance Comparison for Distorted Images. arXiv 2017, arXiv:1710.02726. [Google Scholar] [CrossRef]
- OpenCV 4.13.0-dev: Basics of Brute-Force Matcher. Available online: https://docs.opencv.org/4.x/dc/dc3/tutorial_py_matcher.html (accessed on 23 November 2025).








| Pepper | Pumpkin | Spinach | Tomato | |
|---|---|---|---|---|
| training | 180 | 100 | 900 | 100 |
| testing | 32 | 50 | 100 | 50 |
| Dataset | Amount | Content | Date |
|---|---|---|---|
| KITTI [13] | 394 | streets in Germany | 2015 |
| MS2 [27] | 180,000 | streets in China | 2023 |
| MVSEC [28] | n.a. | streets and indoor areas | 2018 |
| DrivingStereo [29] | 174,437 | streets in China | 2019 |
| Argoverse [30,31,32] | over 1000 | streets in U.S. | 2019 |
| TartanDrive 2.0 [33] | 7 h of videos | off-road areas | 2024 |
| Oxford RobotCar [34,35] | 100 videos | streets in UK | 2016 |
| EuRoC MAV [36] | n.a. | data from a drone, ground truth in the form of a 3D scan | 2016 |
| Road Surface Dataset [37] | 300 | views of bumps on roads | 2023 |
| Dataset | Amount | Content | Technology | Release Date |
|---|---|---|---|---|
| FoundationStereo [39] | 1 million | large diversity and high photorealism | NVIDIA Omniverse | 2025 |
| Falling Things [40] | 60,000 | household objects | Unreal Engine 4 | 2025 |
| FlyingThings3D, Driving, Monkaa [41] | 39,000 | animated objects | Blender | 2016 |
| Spring [42,43] | 6000 | open-source animated movie “Spring” | Blender | 2023 |
| MPI Sintel [44] | 1000 | animated movie | Blender | 2012 |
| Virtual KITTI 2 [45,46] | over 10,000 | photo-realistic street views inspired by KITTI | Unity | 2020 |
| KITTI-CARLA [47,48] | 5000 | street views resembling data from KITTI | CARLA simulator | 2021 |
| IRS [49] | 100,000 | naturally looking indoor spaces | Unreal Engine 4 | 2021 |
| Dynamic Replica [50,51] | 524 videos | humans and animals in everyday scenes | Facebook Replica | 2023 |
| Dataset | Size | DCB | SDC | min Y disc | max Y disc |
|---|---|---|---|---|---|
| pepper testing | 32 | 0.71 | 0.83 | −0.038 | 0.767 |
| pepper training | 180 | 0.79 | 0.79 | −0.758 | 0.933 |
| pumpkin testing | 50 | 0.93 | 0.83 | −0.260 | 0.452 |
| pumpkin training | 100 | 0.89 | 0.85 | −0.238 | 0.614 |
| spinach testing | 100 | 0.99 | 0.88 | −0.320 | 0.278 |
| spinach training | 900 | 0.99 | 0.83 | −0.607 | 0.730 |
| tomato testing | 50 | 0.9 | 0.87 | −0.171 | 0.539 |
| tomato training | 100 | 0.88 | 0.85 | −0.3 | 0.561 |
| total | 1512 | 0.93 | 0.79 | −0.748 | 0.933 |
| Datasets | DCB | SDC | POA | CPL | Notes |
|---|---|---|---|---|---|
| Middlebury | 0.96 | 0.85 | 1 | 0.96 | |
| ETH3D | 0.99 | 0.87 | 1 | 0.6 | |
| Booster | 0.83 | 0.24 | 1 | 0.95 | |
| PlantStereo | 0.93 | 0.79 | 0.51 | 0.86 | |
| KITTI | 0.93 | 0.88 | 1 | 0.19 | sparse GT |
| MS2 | 0.72 | 0.89 | 1 | 0.17 | sparse GT |
| Scene Flow (Mayer et al. [41]) | 1 | 1 | 0 | 1 | synthetic |
| FoundationStereo | 1 | 1 | 0 | 1 | synthetic |
| MPI Sintel | 1 | 1 | 1 | 1 | synthetic |
| Other synthetic datasets | 1 | 1 | either 1 or 0 | 1 | synthetic |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kaczmarek, A.L. Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation. Appl. Sci. 2025, 15, 12663. https://doi.org/10.3390/app152312663
Kaczmarek AL. Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation. Applied Sciences. 2025; 15(23):12663. https://doi.org/10.3390/app152312663
Chicago/Turabian StyleKaczmarek, Adam L. 2025. "Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation" Applied Sciences 15, no. 23: 12663. https://doi.org/10.3390/app152312663
APA StyleKaczmarek, A. L. (2025). Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation. Applied Sciences, 15(23), 12663. https://doi.org/10.3390/app152312663

