Previous Article in Journal
Methodology of Object Reconstruction by Photogrammetry and Structured-Light Scanning for Industrial 3D Visualisation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Monocular Visual Measurement System Uncertainty Analysis and One-Step End–End Estimation Upgrade

by
Kuai Zhou
1,2,*,
Wenmin Chu
1 and
Peng Zhao
2
1
School of Aeronautical Engineering, Nanjing University of Industry Technology, Nanjing 210023, China
2
College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(23), 7179; https://doi.org/10.3390/s25237179 (registering DOI)
Submission received: 24 October 2025 / Revised: 18 November 2025 / Accepted: 21 November 2025 / Published: 24 November 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Monocular visual measurement and vision-guided robotics technology find extensive application in modern automated manufacturing, particularly in aerospace assembly. However, during assembly pose measurement and guidance, the propagation and accumulation of multi-source errors—including those from visual measurement, hand–eye calibration, and robot calibration—impact final assembly accuracy. To address this issue, this study first proposes an uncertainty analysis method for monocular visual measurement systems in assembly pose, encompassing the determination of uncertainty propagation paths and input uncertainty values. Building on this foundation, the system’s uncertainty is analyzed. Inspired by the uncertainty analysis results, this study further proposes a direct one-step solution to a series of problems in robot calibration and hand–eye calibration using a nonlinear mapping estimation method. Through experiments and discussion, a high-performance, one-step, end-to-end pose estimation convolutional neural network (OECNN) is constructed. The OECNN achieves direct mapping from the pose variation of the target object to the drive volume variation of the positioner. The uncertainty analysis conducted in this study yields a series of conclusions that are significant for further enhancing the precision of assembly pose estimation. The proposed uncertainty analysis methodology may also serve as a reference for uncertainty analysis in complex systems. Experimental validation demonstrates that the proposed one-step end-to-end pose estimation method exhibits high accuracy. It can be applied to automated assembly tasks involving various vision-guided robots, including those with typical configurations, and it is particularly suitable for high-precision assembly scenarios, such as aircraft assembly.
Keywords: monocular vision; aircraft assembly; uncertainty analysis; one-step end-to-end; pose estimation monocular vision; aircraft assembly; uncertainty analysis; one-step end-to-end; pose estimation

Share and Cite

MDPI and ACS Style

Zhou, K.; Chu, W.; Zhao, P. Monocular Visual Measurement System Uncertainty Analysis and One-Step End–End Estimation Upgrade. Sensors 2025, 25, 7179. https://doi.org/10.3390/s25237179

AMA Style

Zhou K, Chu W, Zhao P. Monocular Visual Measurement System Uncertainty Analysis and One-Step End–End Estimation Upgrade. Sensors. 2025; 25(23):7179. https://doi.org/10.3390/s25237179

Chicago/Turabian Style

Zhou, Kuai, Wenmin Chu, and Peng Zhao. 2025. "Monocular Visual Measurement System Uncertainty Analysis and One-Step End–End Estimation Upgrade" Sensors 25, no. 23: 7179. https://doi.org/10.3390/s25237179

APA Style

Zhou, K., Chu, W., & Zhao, P. (2025). Monocular Visual Measurement System Uncertainty Analysis and One-Step End–End Estimation Upgrade. Sensors, 25(23), 7179. https://doi.org/10.3390/s25237179

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop