Comparative Analysis of Correction Methods for Multi-Camera 3D Image Processing System and Its Application Design in Safety Improvement on Hot-Working Production Line
Abstract
Featured Application
Abstract
1. Introduction
2. State of the Art-Methods of Geometric Distortion Compensation and VR Content Streaming
- GoogLeNet/Inception models, from Inception V1 to V4 [39].
- (1)
- (2)
- (3)
- Dynamic Omni-Directional Stereo Imaging—synchronized multi-camera array systems proposed in [9,58] and applying interpolation for the view adjustment. This category of solutions is contemporarily most common. It is also analyzed in this article. The biggest challenge related to these systems, the massive amount of data with high processing requirements (Facebook’s 360 cameras record 30 frames per second, generating 17 Gb/s of raw data streamed for offline processing), was tackled in the current work, with noticeable improvements compared to the results from reported works. The confirmation of the improvement achieved by utilizing the chosen method with special adjustments and modifications is presented in the images recorded during real-time video transmission of an event (paragliders competition).
3. The Test and Analysis of Different Approaches to Distortion Correction Considering Effectiveness and Data Processing Efficiency
- Sony A7 III (ILCE-7M3) + adapter Metabones MB-EF-E-BT5 without optics + Canon EF 8–15 mm f/4L Fisheye USM.
- Panasonic Lumix DC-GH5S + adapter Metabones Canon EF to Micro Four Thirds T Speed Booster XL 0.64z with additional optics + Canon EF 8–15 mm f/4L Fisheye USM.
- Panasonic Lumix DC-GH5S + Fujinon FE185C086HA-1 with additional passive adapter.
- Pixelink PL-D755CU-BL-08051 + Sunex DSL-315B with additional passive adapter.
- Third-order polynomial correction (ABCD polynomial);
- Nearest-neighbor interpolation (polyline method) with interpolation between experimentally derived points.
- LUT—Precomputed ABCD polynomial values stored in a temporary texture; tested at nine resolutions.
- realtime-lutcorr—Frame-by-frame correction using polyline data stored in a 1D texture.
- realtime-abcd—Polynomial ABCD computed for each frame independently.
- realtime—No correction applied.
- (1)
- Using a precise calculation method from the family concurrency: precise_math, described further as the precise method.
- (2)
- Vector normalization in the appropriate place, significantly reducing artifacts, described as the fast + fix method.
- (3)
- Method WITHOUT any changes is described as the fast method.
- 3840 × 2160 resolution (1920 × 2160 per eye), 30 FPS, 13 Mbps bitrate—suitable for standard high-definition VR transmission.
- 1920 × 1080 resolution (960 × 1080 per eye), 60 FPS, 20 Mbps bitrate—optimized for dynamic content such as sports events.
- 1920 × 1080 resolution, 25 FPS, 10 Mbps bitrate—used in low-light conditions requiring longer exposure times.
4. The Design of a VR Application for Safety Improvement Training System
- Accurate positioning of the forklift near the main furnace;
- Loading the ladle with molten aluminum alloy;
- Transporting the load to one of the 48 smaller furnaces spread across three production lines;
- Approaching the assigned workstation;
- Depositing the alloy into the smaller furnace.
- Stereoscopic 3D Video Training Module—This module will deliver live-streamed, high-resolution 3D video focusing on tasks prone to procedural errors, such as step omissions, incorrect sequences, or lack of concentration. The video feed will provide multi-angle views and include visual annotations (e.g., arrows, numbering, and safety tips) to enhance instructional clarity. The processed recordings will also be used for on-demand training and procedural reviews.
- VR Simulation Module—A fully interactive simulator replicating the workstation environment will be used for practical training and skill assessment. This will allow new hires to practice and demonstrate competence before being deployed on the factory floor.
5. Discussion and Conclusions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Konrad, R.; Dansereau, D.G.; Masood, A.; Wetzstein, G. SpinVR: Towards Live-Streaming 3D Virtual Reality Video. ACM Trans. Graph. 2017, 36, 209. [Google Scholar] [CrossRef]
- Luhmann, T.; Robson, S.; Kyle, S.; Boehm, J.; de Gruyter, W. Close-Range Photogrammetry and 3D Imaging; Gruyter GmbH & Co KG: Berlin, Germany, 2020. [Google Scholar]
- Fanini, B.; Ferdani, D.; Demetrescu, E. Temporal Lensing: An Interactive and Scalable Technique for Web3D/WebXR Applications in Cultural Heritage. YOCOCU2020 Hands Herit. Exp. Conserv. Mastering Manag. Herit. 2021, 4, 710–724. [Google Scholar] [CrossRef]
- Vieri, C.; Lee, G.; Balram, N.; Jung, S.; Yang, J.; Yoon, S.; Kang, I. An 18 megapixel 4.3" 1443 ppi 120 Hz OLED display for wide field of view high acuity head mounted displays. J. Soc. Inf. Disp. 2018, 26, 314–324. [Google Scholar] [CrossRef]
- Gao, Z.; Hwang, A.; Zhai, G.; Peli, E. Correcting geometric distortions in stereoscopic 3D imaging. PLoS ONE 2018, 13, e0205032. [Google Scholar] [CrossRef] [PubMed]
- Dansereau, D.G.; Schuster, G.; Ford, J.; Wetzstein, G. A Wide-Field-of-View Monocentric Light Field Camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5048–5057. [Google Scholar]
- Hoffman, D.; Meraz, Z.; Turner, E. Limits of peripheral acuity and implications for VR system design. J. Soc. Inf. Disp. 2018, 26, 483–495. [Google Scholar] [CrossRef]
- Hoffman, D.M.; Lee, G. Temporal Requirements for VR Displays to Create a More Comfortable and Immersive Visual Experience. Inf. Disp. 2020, 35, 9–39. [Google Scholar] [CrossRef]
- Anderson, R.; Gallup, D.; Barron, J.T.; Kontkanen, J.; Snavely, N.; Hernandez, C.; Agarwal, S.; Seitz, S.M. Jump: Virtual Reality Video. ACM Trans. Graph. (SIGGRAPH Asia) 2016, 35, 397–407. [Google Scholar] [CrossRef]
- Arza-García, M.; Núñez-Temes, C.; Lorenzana, J.A. Evaluation of a low-cost approach to 2-D digital image correlation vs. a commercial stereo-DIC system in Brazilian testing of soil specimens. Arch. Civ. Mech. Eng. 2022, 22, 4. [Google Scholar] [CrossRef]
- Hughes, C.; Jones, E.; Glavin, M.; Denny, P. Validation of Polynomial-based Equidistance Fish-Eye Models. In IET Irish Signals and Systems Conference (ISSC); IET: Stevenage, UK, 2009. [Google Scholar]
- Kumlera, J.; Martin, B. Fisheye lens designs and their relative performance. Proc. SPIE 2000, 4093, 360–369. [Google Scholar]
- Kannala, J.; Brandt, S.S. A generic camera model and calibration method for conventional, wide-angle, and fish-eye lenses. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1335–1340. [Google Scholar] [CrossRef]
- Hughes, C.J.; Glavin, M.; Jones, E.; Trivedi, M.M. Review of geometric distortion compensation in automotive camera systems. IEEE Trans. Intell. Transp. Syst. 2010, 11, 377–384. [Google Scholar]
- Scaramuzza, D.; Martinelli, A.; Siegwart, R. A flexible technique for accurate omnidirectional camera calibration and structure from motion. In Proceedings of the IEEE International Conference on Computer Vision Systems, New York, NY, USA, 4–7 January 2006; pp. 45–51. [Google Scholar]
- Rong, Y.; Su, S.; Yan, S. Radial distortion correction with deep convolutional neural networks. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 1067–1071. [Google Scholar]
- Yu, X.; Ramalingam, S. New calibration methods and datasets for omnidirectional cameras. Int. J. Comput. Vis. 2018, 126, 1113–1135. [Google Scholar]
- Madejski, G.; Zbytniewski, S.; Kurowski, M.; Gradolewski, D.; Kaoka, W.; Kulesza, W.J. Lens Distortion Measurement and Correction for Stereovision Multi-Camera System. Eng. Proc. 2024, 82, 85. [Google Scholar] [CrossRef]
- Hughes, C.; Glavin, M.; Jones, E.; Denny, P. Review of Geometric Distortion Compensation in Fish-Eye Cameras. In Proceedings of the IET Irish Signals and Systems Conference (ISSC), Galway, Ireland, 18–19 June 2008. [Google Scholar]
- Slama, C. (Ed.) Manual of Photogrammetry. In American Society of Photogrammetry, 4th ed.; American Society of Photogrammetry: Falls Church, VA, USA, 1980. [Google Scholar]
- Tsai, R. A versatile camera calibration technique for high accuracy 3d machine vision metrology using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 4, 323–344. [Google Scholar] [CrossRef]
- Mallon, J.; Whelan, P.F. Precise radial un-distortion of images. In Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, UK, 26–26 August 2004; Volume 1, pp. 18–21. [Google Scholar]
- Ahmed, M.; Farag, A. Nonmetric calibration of camera lens distortion: Differential methods and robust estimation. IEEE Trans. Image Process. 2005, 14, 1215–1230. [Google Scholar] [CrossRef]
- Fitzgibbon, A.W. Simultaneous linear estimation of multiple view geometry and lens distortion. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001; Volume 1, pp. 125–132. [Google Scholar]
- Asari, K.V. Design of an efficient VLSI architecture for non-linear spatial warping of wide-angle camera images. J. Sys. Arch. 2004, 50, 743–755. [Google Scholar] [CrossRef]
- Shah, S.; Aggarwal, J.K. A simple calibration procedure for fisheye (high distortion) lens camera. In Proceedings of the 1994 IEEE International Conference on Robotics and Automation, San Diego, CA, USA, 8–13 May 1994; Volume 4, pp. 3422–3427. [Google Scholar]
- Basu, A.; Licardie, S. Alternative models for fish-eye lenses. Elsevier Pattern Recognit. Lett. 1995, 16, 433–441. [Google Scholar] [CrossRef]
- Devernay, F.; Faugeras, O. Straight lines have to be straight: Automatic calibration and removal of distortion from scenes of structured environments. Springer-Verl. J. Mach. Vis. Appl. 2001, 13, 14–24. [Google Scholar] [CrossRef]
- Ishii, C.; Sudo, Y.; Hashimoto, H. An image conversion algorithm from fisheye image to perspective image for human eyes. In Proceedings of the IEEE/ASME International Conference on Advanced Intelligent Mechatronics, Kobe, Japan, 20–24 July 2003; Volume 2, pp. 1009–1014. [Google Scholar]
- Chen, L.; Li, S.; Bai, Q.; Yang, J.; Jiang, S.; Miao, Y. Review of Image Classification Algorithms Based on Convolutional Neural Networks. Remote Sens. 2021, 13, 4712. [Google Scholar] [CrossRef]
- Hu, F.; Xia, G.S.; Hu, J.; Zhang, L. Transferring deep convolutional neural networks for the scene classification of high-resolution remote sensing imagery. Remote Sens. 2015, 7, 14680–14707. [Google Scholar] [CrossRef]
- Minetto, R.; Segundo, M.P.; Sarkar, S. Hydra: An ensemble of convolutional neural networks for geospatial land classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6530–6541. [Google Scholar] [CrossRef]
- Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene classification with recurrent attention of VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1155–1167. [Google Scholar] [CrossRef]
- Chen, G.; Zhang, X.; Tan, X.; Cheng, Y.; Dai, F.; Zhu, K.; Gong, Y.; Wang, Q. Training small networks for scene classification of remote sensing images via knowledge distillation. Remote Sens. 2018, 10, 719. [Google Scholar] [CrossRef]
- Mangasarian, O.L.; Musicant, D.R. Data Discrimination via Nonlinear Generalized Support Vector Machines. In Complementarity: Applications, Algorithms and Extensions; Springer: Boston, MA, USA, 2001; pp. 233–251. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Bengio, Y.; Courville, A.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 15 October 2015; pp. 1–9. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Identity Mappings in Deep Residual Networks. In Computer Vision—ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 630–645. [Google Scholar]
- Zagoruyko, S.; Komodakis, N. Wide Residual Networks. arXiv 2016, arXiv:1605.07146. [Google Scholar] [CrossRef]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
- Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G.; Albanie, S. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Park, J.; Woo, S.; Lee, J.Y.; Kweon, I.S. BAM: Bottleneck Attention Module. arXiv 2018, arXiv:1807.06514. [Google Scholar] [CrossRef]
- Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Vedaldi, A. Gather-Excite: Exploiting Feature Context in Convolutional Neural Networks; NIPS’18; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 9423–9433. [Google Scholar]
- Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
- Ionescu, C.; Vantzos, O.; Sminchisescu, C. Matrix Backpropagation for Deep Networks with Structured Layers. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2965–2973. [Google Scholar]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. arXiv 2019, arXiv:1910.03151. [Google Scholar]
- Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. arXiv 2021, arXiv:2103.02907. [Google Scholar] [CrossRef]
- Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
- Howard, A.; Pang, R.; Adam, H.; Le, Q.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.C.; Tan, M.; Chu, G.; et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
- Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
- Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-Aware Neural Architecture Search for Mobile. arXiv 2018, arXiv:1807.11626. [Google Scholar]
- Brown, M.; Lowe, D.G. Automatic Panoramic Image Stitching Using Invariant Features. Int. J. Comput. Vis. 2007, 74, 59–73. [Google Scholar] [CrossRef]
- Matzen, K.; Cohen, M.F.; Evans, B.; Kopf, J.; Szeliski, R. Low-cost 360 Stereo Photography and Video Capture. ACM Trans. Graph. 2017, 36, 148:1–148:12. [Google Scholar] [CrossRef]
- Richardt, C.; Pritch, Y.; Zimmer, H.; Sorkine-Hornung, A. Megastereo: Constructing High-Resolution Stereo Panoramas. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 1256–1263. [Google Scholar]
- Bourke, P. Capturing omni-directional stereoscopic spherical projections with a single camera. In Proceedings of the 2010 16th International Conference on Virtual Systems and Multimedia, Seoul, Republic of Korea, 20–23 October 2010; pp. 179–183. [Google Scholar]
- Jones, A.; McDowall, I.; Yamada, H.; Bolas, M.; Debevec, P. Rendering for an Interactive 360 Light Field Display. ACM Trans. Graph. (SIGGRAPH) 2007, 26, 3. [Google Scholar] [CrossRef]
- Peleg, S.; Ben-Ezra, M.; Pritch, Y. Omnistereo: Panoramic stereo imaging. IEEE Trans. PAMI 2001, 23, 279–290. [Google Scholar] [CrossRef]
- Tanaka, K.; Hayashi, J.; Inami, M.; Tachi, S. TWISTER: An immersive autostereoscopic display. In Proceedings of the IEEE Virtual Reality 2004, Chicago, IL, USA, 27–31 March 2004; pp. 59–66. [Google Scholar]
- Gąbka, J. Edge computing technologies as a crucial factor of successful Industry 4.0 growth. The case of live video data streaming. In Advances in Manufacturing II Volume 1; Solutions for Industry 4.0. Lecture Notes in Mechanical Engineering; Justyna, T., Olaf, C., José, M.M., Pavlenko, I., Eds.; Springer International Publishing: Cham, Switzerland, 2019. [Google Scholar]
- Zhu, X.; Yu, S. A review of distortion correction in wide-angle vision systems: Challenges and opportunities. Appl. Sci. 2021, 11, 3001. [Google Scholar]
Aberration Values | |||
---|---|---|---|
Focal Length | 8 mm | 12 mm | 15 mm |
f/4 | 0.058 | 0.047 | 0.039 |
f/4,5 | 0.056 | 0.047 | 0.041 |
f/5 | 0.052 | 0.049 | 0.044 |
f/5,6 | 0.055 | 0.046 | 0.046 |
f/6,3 | 0.052 | 0.041 | 0.041 |
f/7,1 | 0.053 | 0.049 | 0.049 |
f/8 | 0.052 | 0.047 | 0.047 |
f/9 | 0.052 | 0.049 | 0.049 |
f/10 | 0.055 | 0.044 | 0.053 |
f/11 | 0.056 | 0.044 | 0.049 |
f/13 | 0.051 | 0.048 | 0.05 |
f/14 | 0.055 | 0.048 | 0.049 |
f/16 | 0.057 | 0.046 | 0.049 |
f/18 | 0.055 | 0.048 | 0.049 |
f/20 | 0.057 | 0.049 | 0.052 |
f/22 | 0.056 | 0.05 | 0.053 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gąbka, J. Comparative Analysis of Correction Methods for Multi-Camera 3D Image Processing System and Its Application Design in Safety Improvement on Hot-Working Production Line. Appl. Sci. 2025, 15, 9136. https://doi.org/10.3390/app15169136
Gąbka J. Comparative Analysis of Correction Methods for Multi-Camera 3D Image Processing System and Its Application Design in Safety Improvement on Hot-Working Production Line. Applied Sciences. 2025; 15(16):9136. https://doi.org/10.3390/app15169136
Chicago/Turabian StyleGąbka, Joanna. 2025. "Comparative Analysis of Correction Methods for Multi-Camera 3D Image Processing System and Its Application Design in Safety Improvement on Hot-Working Production Line" Applied Sciences 15, no. 16: 9136. https://doi.org/10.3390/app15169136
APA StyleGąbka, J. (2025). Comparative Analysis of Correction Methods for Multi-Camera 3D Image Processing System and Its Application Design in Safety Improvement on Hot-Working Production Line. Applied Sciences, 15(16), 9136. https://doi.org/10.3390/app15169136