Open AccessThis article is
- freely available
A Visual Servoing-Based Method for ProCam Systems Calibration
DREAM, Insitut Pascal, 24 avenue des Landais, Aubiere 63170, France
* Author to whom correspondence should be addressed.
Received: 3 July 2013; in revised form: 27 August 2013 / Accepted: 23 September 2013 / Published: 1 October 2013
Abstract: Projector-camera systems are currently used in a wide field of applications, such as 3D reconstruction and augmented reality, and can provide accurate measurements, depending on the configuration and calibration. Frequently, the calibration task is divided into two steps: camera calibration followed by projector calibration. The latter still poses certain problems that are not easy to solve, such as the difficulty in obtaining a set of 2D–3D points to compute the projection matrix between the projector and the world. Existing methods are either not sufficiently accurate or not flexible. We propose an easy and automatic method to calibrate such systems that consists in projecting a calibration pattern and superimposing it automatically on a known printed pattern. The projected pattern is provided by a virtual camera observing a virtual pattern in an OpenGL model. The projector displays what the virtual camera visualizes. Thus, the projected pattern can be controlled and superimposed on the printed one with the aid of visual servoing. Our experimental results compare favorably with those of other methods considering both usability and accuracy.
Keywords: projector calibration; structured light system; OpenGL modeling; visual servoing
Citations to this Article
Cite This Article
MDPI and ACS Style
Berry, F.; Aider, O.A.; Mosnier, J. A Visual Servoing-Based Method for ProCam Systems Calibration. Sensors 2013, 13, 13318-13333.
Berry F, Aider OA, Mosnier J. A Visual Servoing-Based Method for ProCam Systems Calibration. Sensors. 2013; 13(10):13318-13333.
Berry, Francois; Aider, Omar A.; Mosnier, Jeremie. 2013. "A Visual Servoing-Based Method for ProCam Systems Calibration." Sensors 13, no. 10: 13318-13333.