Next Article in Journal
Interval State Estimation in Active Distribution Systems Considering Multiple Uncertainties
Next Article in Special Issue
Machine Learning for Sensorless Temperature Estimation of a BLDC Motor
Previous Article in Journal
Non-Contact Measurement of Motion Sickness Using Pupillary Rhythms from an Infrared Camera
Previous Article in Special Issue
Piston Error Measurement for Segmented Telescopes with an Artificial Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Constrained Multiple Planar Reconstruction for Automatic Camera Calibration of Intelligent Vehicles

1
Division of Electronic Engineering, Jeonbuk National University, 567 Baekje-daero, Deokjin-gu, Jeonju-si 54896, Jeollabuk-do, Korea
2
Samsung Advanced Institute of Technology (SAIT), 130 Samsung-ro, Yeongtong-gu, Suwon-si 16678, Gyeonggi-do, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(14), 4643; https://doi.org/10.3390/s21144643
Submission received: 1 June 2021 / Revised: 1 July 2021 / Accepted: 4 July 2021 / Published: 6 July 2021
(This article belongs to the Special Issue Machine Learning in Sensors and Imaging)

Abstract

:
In intelligent vehicles, extrinsic camera calibration is preferable to be conducted on a regular basis to deal with unpredictable mechanical changes or variations on weight load distribution. Specifically, high-precision extrinsic parameters between the camera coordinate and the world coordinate are essential to implement high-level functions in intelligent vehicles such as distance estimation and lane departure warning. However, conventional calibration methods, which solve a Perspective-n-Point problem, require laborious work to measure the positions of 3D points in the world coordinate. To reduce this inconvenience, this paper proposes an automatic camera calibration method based on 3D reconstruction. The main contribution of this paper is a novel reconstruction method to recover 3D points on planes perpendicular to the ground. The proposed method jointly optimizes reprojection errors of image features projected from multiple planar surfaces, and finally, it significantly reduces errors in camera extrinsic parameters. Experiments were conducted in synthetic simulation and real calibration environments to demonstrate the effectiveness of the proposed method.

1. Introduction

Recovering the positions of 3D points from 2D-2D correspondences is a fundamental building block in geometric computer vision. This is called triangulation, and it is an essential procedure for many applications including structure-from-motion (SfM) [1,2,3], simultaneous localization and mapping (SLAM) [4,5,6], and visual odometry [7,8]. Triangulation is conducted based on displacements between image correspondences obtained from stereo cameras or a moving camera. In an ideal case, back-projected rays from an image correspondence intersect at a point in three dimensional space, and it can be simply formulated by a direct linear transformation. However, in practice, the rays do not necessarily intersect due to measurement noise involved in image features, and these features do not in general satisfy the epipolar geometry [9]. Therefore, recovering 3D information is not a trivial problem even in a two-view case.
A standard approach for addressing the problem of noisy measurements is to estimate 2D corrected correspondences which satisfy the epipolar geometry with the minimum geometric cost [9,10,11]. These 2D corrected points are maximum likelihood estimates under the assumption of zero-mean isotropic Gaussian noise on the measurements [10]. Triangulation is conducted for the corrected correspondences, and it is equivalent to estimate 3D points which minimize the reprojection error. This procedure is called optimal triangulation. In case that all the 3D points are on a plane, their projected points in two views are associated with a projective transformation which is called homography [12]. Chum et al. proposed a method to find optimal 2D correspondences projected from 3D points on a plane, and it is called optimal planar reconstruction [13]. Kanatani et al. further derived an optimal solution for planar scene triangulation in case that plane and camera parameters are unknown [14]. Planar constraint in general reduces a significant amount of reconstruction error by associating multiple image features for correcting individual 2D measurements. This motivates us to associate image features projected from multiple planar surfaces to improve the precision of 3D reconstruction.
In this paper, we propose a multiple planar reconstruction method which can be applicable in a man-made environment: planes of interest are perpendicular to the ground. It is worth noting that this scenario is quite reasonable in environments for end-of-line calibration and indoor camera calibration. This assumption implies that the normal vectors of planes are coplanar. Whereas previous methods reconstruct individual planes, we seek to jointly optimize the structures of multiple planes simultaneously by introducing coplanarity constraints of their normal vectors.
Another main work of this paper is a novel extrinsic camera calibration method. Among various types of extrinsic parameters, our goal is to estimate extrinsic parameters between the camera coordinate and the world coordinate, which are essential prerequisite for high-level functions in intelligent vehicles such as distance estimation and lane departure warning. Extrinsic parameters consist of relative rotation and translation between the camera and world coordinates, and the world coordinate is also called the vehicle coordinate. In recent intelligent vehicles, due to unpredictable mechanical changes or variations on weight distribution, it is desirable to perform extrinsic camera calibration on a regular basis [15], e.g., at the start of every driving. However, camera calibration requires correspondences between 2D image projections and their 3D points [16] to solve a variant of Perspective-n-Point (PnP) problems [17,18,19,20,21], and the procedure for measuring accurate 3D points in the world coordinate is the biggest bottleneck in conventional methods.
Recently, deep learning methods have been utilized in automatic camera calibration for intelligent vehicles [22,23,24]. However, many original equipment manufacturers (OEMs) and Tier 1 component companies require conventional computer vision methods to guarantee the safety and reliability of the camera calibration function. Moreover, even OEMs, which employ deep learning techniques in recognition and planning algorithms, parallelly utilize conventional methods to acquire a satisfactory level of functional safety for several essential functions such as camera calibration. This paper proposes an extrinsic camera calibration method to reduce manual work in conventional approaches. The proposed method uses 3D cues in the camera coordinate to infer 3D information in the world coordinate by utilizing 3D points on a planar chessboard which is perpendicular to the ground. The main advantage of the method is that it is able to estimate extrinsic parameters without measuring 3D points in the world coordinate.
The contributions of this paper are twofold: (1) multiple planar reconstruction method to jointly recover 3D structures of multiple planar surfaces and (2) extrinsic camera calibration method based on the reconstructed points. This paper is organized as follows. Section 2 presents the multiple planar reconstruction method, and Section 3 explains the extrinsic camera calibration method based on 3D reconstruction. Section 4 presents experimental results in both synthetic and real environments to demonstrate the effectiveness of the proposed methods.

2. Optimal Multiple-Planar Reconstruction

This section presents the proposed method for joint reconstruction of multiple planar surfaces. We assume that planar chessboards are installed perpendicular to the ground, and a vehicle with a camera moves along the ground with a planar motion. Under these assumptions, the plane normal vectors and camera motion vector are orthogonal to the ground normal vector, and therefore, the plane normal and camera motion vectors are coplanar. Although these assumptions are not easy to satisfy in a road situation, it is worth utilizing the proposed method to improve the precision and robustness of end-of-line and indoor calibrations, which are usually conducted in a man-made environment. In Figure 1, a synthetic configuration containing four planar boards in three dimensional space is projected onto an image plane by using a camera matrix and arbitrary extrinsic parameters. The camera motion vector is indicated by m , and the tetragons filled with a reddish color and the tetragons filled with 2D points depict the projections of the four planar boards from the first and second views, respectively.
The outline of the proposed method is as follows. First, normal vectors of the planar chessboards are jointly optimized to minimize the algebraic error of homographies subject to the coplanarity constraints. Then, these homographies are used to correct 2D measurements, and 3D structures of the planar surfaces are recovered via triangulation of the corrected 2D correspondences. Finally, extrinsic parameters are estimated by using the method presented in Section 3.

2.1. Plane Induced Homographies

Suppose that planar surfaces are imaged by a calibrated camera in two views under planar motion of a vehicle. In three dimensional space, the k-th plane can be represented as its scaled normal vector n k so that n k X C + 1 = 0 , where X C is a 3D point in the camera coordinate. Suppose that the essential matrix between the two-view is given by E = [ t ] × R , where R is rotation matrix, t is translation vector, and [ · ] × is a 3 × 3 skew-symmetric matrix for representing cross product as a matrix multiplication. Then, it is well-known that a homography induced by the k-th plane can be represented as
H k = R t n k .
The essential matrix, R , and t can be computed via ego-motion estimation. There are several methods to estimate ego-motion of a vehicle, and many previous methods utilize optical flow of background features [25,26].
Suppose that the i-th point on the k-th plane in three dimensional space is projected onto two image planes, and x i , k and x i , k are homogeneous representations of the 2D projections in the first and second views, respectively. Then the 2D correspondence satisfies x i , k H k x i , k = ( R t n k ) x i , k , and it can be reformulated as
x i , k n k = ( x i , k × R x i , k ) ( x i , k × t ) ( x i , k × t ) ( x i , k × t ) = b i , k ,
when x i , k and t are not parallel. The operation × indicates cross-product between two vectors, and ≃ implies that two vectors are equal up to scale. The scaled normal vector n k of the k-th plane can be obtained by solving the unconstrained optimization problem:
minimize n k A k n k b k 2 ,
where b k = [ b 1 , k , , b N k , k ] R N k and A k = [ x 1 , k , , x N k , k ] R N k × 3 . Homography optimization in the previous work [14] can be reformulated as a two-step process: optimization of a scaled normal vector by (3) and homography recovery by (1).

2.2. Multiple Planar Reconstruction

This section presents the main idea of the proposed method which introduces coplanarity constraints of plane normal vectors and camera motion vector to jointly reconstruct multiple planar surfaces. In three dimensional space, coplanarity of the normal vectors n i ( i = 1 , , K ) and camera motion m = R t can be represented as
( n 1 × n k ) · m = 0 , k = 2 , , K .
To formulate the quadratic constraints in terms of optimization variables, all of the constraints are associated with the camera motion vector m , which is constant in the optimization problem. The number of equations for constraining the coplanarity of K normal vectors and m is K ( K + 1 ) / 2 . However, these equations have redundancy, because, for example, the coplanarities of ( n i , n j , m ) and ( n i , n k , m ) ensure the coplanarity of ( n j , n k , m ) . Therefore, the minimum number of equality constraints for the coplanarity of K normal vectors is K 1 .
Let w be a concatenated normal vector such that w = [ n 1 , , n K ] , then the coplanarity (4) can be reformulated as
w C k w = 0 , k = 2 , , K ,
where C k is a 3 K × 3 K symmetric block matrix, of which partitions are 3 × 3 zero matrices except that C 1 k = [ m ] × and C k 1 = [ m ] × ; C i j is a 3 × 3 submatrix corresponding to the i-th row and j-th column block. By introducing an auxiliary dimension to w so that w ˜ = [ w , 1 ] , the optimization problem for minimizing the objective function of (3) subject to the coplanarity constraints (4) can be formulated as
minimize w ˜ w ˜ Q w ˜ subject to w ˜ C ˜ k w ˜ = 0 , k = 2 , , K , w ˜ C 0 w ˜ = 1 ,
where Q = A ˜ A ˜ A ˜ b b A ˜ 0 , C ˜ k = C k 0 0 0 , C 0 = 0 3 K 0 0 1 , A ˜ R ( k = 1 K N k ) × 3 is the block diagonal matrix, of which the k-th diagonal submatrix is A k R N k × 3 and off-diagonal blocks are zero matrices, and 0 3 K is a 3 K × 3 K zero matrix. This optimization problem (6) is a quadratically constrained quadratic program (QCQP); the objective is a quadratic function with a positive semidefinite matrix, and the constraints are quadratic with symmetric matrices. Becuase (6) is an NP-hard optimization problem, we reformulate it as a semidefinite program (SDP) by applying the parameterization of S = w ˜ w ˜ and relaxation of a rank constraint.

2.3. SDP Relaxation

With the parametrization of S = w ˜ w ˜ S + , where S + is the set of positive semidefinite matrices, the QCQP (6) can be reformulated in terms of inner products of matrices as
minimize S S + Q , S subject to C ˜ k , S = 0 , k = 2 , , K , C 0 , S = 1 , r a n k ( S ) = 1 .
By eliminating the rank constraint in (7), we can obtain the semidefinite relaxation:
minimize S S + Q , S subject to C ˜ , S = 0 , k = 2 , , K , C 0 , S = 1 .
Becuase the SDP (8) is a convex optimization, we can find the global optimum reliably. Zhao proved the tightness between a primal QCQP and its rank relaxation at noise-free observations, and further showed the stability of rank relaxation at noisy observations [27]. We indeed observe that the solution of rank-relaxed problem (8) always satisfies the rank-1 in both synthetic simulation and real calibration environments.

2.4. Recovering 3D Points

Once the optimal S of the SDP (8) is obtained, w ˜ can be recovered by computing the eigenvector of S corresponding to the largest eigenvalue. By utilizing w ˜ and ego-motion of the vehicle, individual homographies are computed by (1). Based on the plane induced homographies, optimal corrections x ^ i , k and x ^ i , k can be obtained for each 2D measurements x i , k and x i , k , by solving a polynomial of degree 8 [13] or by using Sampson’s method [28]. The positions of 3D points can be recovered by applying triangulation to the corrected 2D points x ^ i , k and x ^ i , k . Figure 2 shows 3D points on planar surfaces reconstructed by naïve triangulation [28], optimal planar reconstruction [13], and our method.

3. Camera Calibration Based on 3D Reconstruction

3.1. Vehicle Modeling

In this section, we propose a monocular camera calibration method based on 3D reconstruction. Figure 3 shows our vehicle model. Our world coordinate is defined so that its origin is the perpendicular projection of the camera centre to the ground, and the direction of Z W axis is identical to the normal vector of the world coordinate so that it follows ISO 8855. Under the definition of the world coordinate, fixed values of longitudinal and lateral offsets between the world origin and the foremost point of a vehicle can be compensated at the process of generating signals such as distances to frontal vehicles and time to collision.
The relation between the world and camera coordinates can be formulized in terms of Euler angles (pitch θ , yaw ψ , roll ϕ ) and camera height (h) as
X C = R ( θ , ϕ ; ψ ) X W + t ( θ , ϕ , h ; ψ ) ,
where X C = [ X C , Y C , Z C ] is a 3D point in the camera coordinate, X W = [ X W , Y W , Z W ] is a 3D point in the world coordinate, and the rotation matrix R ( θ , ϕ ; ψ ) is defined as (10).
R ( θ , ϕ ; ψ ) = cos θ sin ψ cos ϕ + sin θ sin ϕ cos ψ cos ϕ sin θ sin ψ cos ϕ + cos θ sin ϕ cos θ sin ψ sin ϕ sin θ cos ϕ cos ψ sin ϕ sin θ sin ψ sin ϕ cos θ cos ϕ cos θ cos ψ sin ψ sin θ cos ψ .
Since 0 = R ( θ , ϕ ; ψ ) · [ 0 , 0 , h ] + t ( θ , ϕ , h ; ψ ) , the translation can be represented as
t ( θ , ϕ , h ; ψ ) = r 3 h ,
where r i is the i-th column vector of R ( θ , ϕ ; ψ ) .
Suppose that 3D world points of interest are on rectangular planar boards, which are perpendicular to the ground, and their Z W components (height) are measured beforehand in the world coordinate. Image features projected from these 3D points are detected while a vehicle with a camera moves along the ground, and yaw angle of the camera is estimated by computing a focus of expansion as presented in [29]. The 3D positions corresponding to these image features are recovered in the camera coordinate by using the multiple planar reconstruction method which is explained in Section 2. The objective of the automatic calibration algorithm is to estimate pitch ( θ ), roll ( ϕ ), and camera height (h) to recover relative rotation and translation between the camera and world coordinates.

3.2. Estimation of Extrinsic Parameters

Let X W i and X W j be the 3D points on a vertical line which is perpendicular to the ground. Since X W i X W j = 0 and Y W i Y W j = 0 , component-wise differences between X W i and X W j can be simplified as (12).
X C i X C j = ( sin θ sin ψ cos ϕ cos θ sin ϕ ) ( Z W i Z W j ) , Y C i Y C j = ( sin θ sin ψ sin ϕ + cos θ cos ϕ ) ( Z W i Z W j ) , Z C i Z C j = sin θ cos ψ ( Z W i Z W j ) .
Based on (12), pitch angle ( θ ) can be estimated by
sin θ = Z C i Z C j cos ψ ( Z W i Z W j ) .
By solving X C i and Y C i in terms of sin ϕ , roll angle ( ϕ ) can be estimated by
sin ϕ = cos θ ( X C i X C j ) sin θ sin ψ ( Y C i Y C j ) ( sin 2 θ sin 2 ψ + cos 2 θ ) ( Z W i Z W j ) .
After the computation of θ and ϕ , camera height (h) can be obtained by solving the following equation with respect to X W , Y W , and h:
[ r 1 , r 2 , r 3 ] [ X W , Y W , h ] = X C r 3 Z W .
Finally, camera extrinsic parameters can be recovered by using (10) and (11).

4. Experimental Results

The proposed method is composed of constrained multiple planar reconstruction and automatic extrinsic camera calibration. To demonstrate the effectiveness of each method, we synthesized a simulation environment, and the reconstruction and calibration errors were evaluated step by step. In both simulation and real experiments, Naïve triangulation [28] and optimal planar reconstruction method [13] were compared with the proposed method. To analyze the effect of the coplanarity constraint, we evaluate the proposed method with two experimental setups: the coplanarity of two plane normal vectors ( K = 2 ) and the coplanarity of four plane normal vectors ( K = 4 ) . For fairness, we used all of the 3D points on the four planar surfaces in every reconstruction method. For example, in the case of K = 2 , two SDPs were optimized to use all of the image features projected from the four planar surfaces. The reconstruction and camera height errors were measured in millimetre (mm), and rotation errors were measured in degree.

4.1. Synthetic Environment

To generate a simulation environment, camera extrinsic parameters were randomly sampled under the normal distributions: θ , ψ , ϕ N ( 0 , 1 2 ) and h N ( 1300 , 50 2 ) , where N is normal distribution with a given mean and variance. This synthetic environment reflects the variations of real extrinsic parameters in our vehicle model, and degree and mm units are utilized for representing angles and camera height, respectively. In the simulation environment, known 3D world points on planar surfaces were projected onto two-view images with the size of 1920 × 1200 by using similar intrinsic parameters to the real case, and Gaussian noise with zero mean and standard deviation of σ was added to the 2D projected image points. To generate the synthetic two-view images, we utilized the vehicle motion when the vehicle moves 1000 mm in forward direction as presented in Figure 4. From the 2D noisy correspondences, reconstruction methods were utilized to recover their 3D points in the camera coordinate, and the proposed calibration method was applied to estimate extrinsic parameters. Each experiment was conducted 100 times, and averaged absolute errors were measured for both reconstructed 3D points and estimated extrinsic parameters.
To evaluate reconstruction accuracy, root-mean-square errors between 3D estimates and their true positions were measured in three dimensional space. Table 1 presents reconstruction errors with respect to various amounts of noise on 2D image projections; the standard deviation σ of the Gaussian pixel-noise was varied from 0.1 to 3.0 . We present two cases of simulation results: In one case, virtual planar boards are located at the longitudinal distance of around 8 m from the camera at the first view, and in the other case, those are located around 10 m. With an identical amount of pixel-noise, reconstruction error increases as the distance to the planar boards increases. Although reconstruction error increases as the amount of pixel-noise increases, the proposed reconstruction method consistently shows higher accuracies compared to the other methods. Furthermore, Table 1 demonstrates that joint optimization of one SDP for the four planes is more advantageous than separate optimization of two independent SDPs for upper two planes and lower two planes. It is because normal vectors of planes in upper and lower groups are not associated with a coplanarity constraint in the case of K = 2 . This result implies that joint reconstruction of entire planar surfaces is effective to reduce the reconstruction error.
Figure 4 shows the reconstruction and calibration errors in the case that distances to targets were around 8 m and the standard deviation of pixel noise was σ = 0.5 . By reducing the reconstruction error, calibration error was significantly decreased especially for pitch angle and camera height. In the results of the proposed method, the reconstruction and calibration errors of K = 4 case were lower than those of K = 2 case. It implies that increasing the number of planes was beneficial to reduce the amount of errors. However, it was not practical to setup more than four planes in real experiments, because 2D image features projected from planes which were located far from the vehicle caused a large amount of pixel noise. Therefore, we utilized four planar surfaces for extrinsic calibration in real experiments.
In the proposed reconstruction method, ego-motion was assumed as a planar motion to formulate a coplanarity constraint with plane normal vectors. To analyze the effect of vehicle motion noise to the performance of the proposed method, we conducted simulation experiments with and without vehicle motion noise. The motion noise was modeled as a Gaussian distribution, and we utilized the standard deviation of ego-motions measured in real driving scenarios to generate Gaussian motion noise in the simulation environment. Table 2 presents calibration accuracies with and without vehicle motion noise under various amounts of pixel noise. Although calibration errors were increased by the ego-motion noise, experimental results show that the proposed method was robust compared to previous methods even under the motion noise.

4.2. Real Calibration Environment

This section presents experimental results in a real calibration environment to demonstrate the effectiveness of the proposed method. In our garage, chessboards were installed so that they are perpendicular to the ground as shown in Figure 5. While a vehicle moved smoothly, images were collected with the size of 1920 × 1200 by utilizing an in-vehicle frontal camera, FLIR Point Grey Grasshopper 3. Intrinsic parameters of the camera were computed in advance by using the method presented in [30]. Background features were extracted and tracked by grid-based feature detection and Lucas–Kanade method [31], and the essential matrix was computed by the five-point method [32] with RANSAC [33] to estimate camera motion. The correspondences of chessboard features were independently detected, and yaw angle of the camera with respect to moving direction was calculated based on focus of expansion, which was computed from the chessboard features. The multiple planar reconstruction method was applied to recover 3D structures of chessboard features, and finally, camera extrinsic parameters were estimated by using the proposed calibration method. This calibration process was performed multiple times while a vehicle was passing the chessboards, and these estimates were averaged to compute a final calibration parameters. In our experiment, the vehicle moved about 5 km/h to obtain enough number of image pairs, and the averaged values of 10 estimates were utilized as final extrinsic parameters. The number of calibration trials could be affected by vehicle speed, field of view of the camera, and distances between chessboards.
To evaluate the accuracy of the proposed method, we collected reference values of extrinsic parameters from an identical experiment environment. In the procedure for generating reference parameters, we manually measured 3D locations of multiple feature points with respect to the world coordinate using a laser range finder, and corresponding 2D projections in the image domain were labelled. 2D-3D correspondences were used to solve a Perspective-n-Point (PnP) problem to compute extrinsic parameters. All the procedures took around 30 min, and it was repeated eight times to obtain averaged extrinsic parameters; the reference values for camera height, pitch angle, roll angle are 1195.48 mm, 0.2413 degree, 0.3663 degree, respectively. In the real experiment, we measured absolute errors between the reference parameters and estimated extrinsic parameters.
To demonstrate the effectiveness of the proposed method, we conducted experiments with four different reconstruction methods: Naïve triangulation [28] and optimal planar reconstruction method [13], and the proposed constrained multiple planar reconstructions ( K = 2 and K = 4 ). Calibration errors in the real calibration scenario are presented in Figure 6, and the proposed reconstruction method gives much lower calibration errors compared to the conventional methods. Similar to experimental results in synthetic simulation, calibration accuracy was improved by utilizing a greater number of planar surfaces in a SDP. Compared to the previous planar reconstruction method [13], height error of the proposed method ( K = 4 ) was reduced from 110.1 mm to 23.9 mm, and pitch angle and roll angle errors were reduced from 0.2764 degree to 0.0470 degree and from 1.1098 degree to 0.0859 degree, respectively; about 78% and 87% of height and angle errors were reduced by using the coplanarity constraint. Because angle errors less than 0.1 degree and height error less than 30 mm were not significant to perform high-level functions such as distance estimation and lane departure warning, the proposed method was able to be utilized in intelligent vehicle industries for computing extrinsic parameters between the camera coordinate and the world coordinate.

5. Conclusions

In this paper, we propose a method for automatic camera calibration of intelligent vehicles. The proposed method is based on 3D reconstruction of a man-made environment, and the key contribution of this paper is novel multiple planar reconstruction method to reduce errors in camera extrinsic parameters. We first formulate a QCQP with the coplanarity constraints between plane normal vectors and camera motion vector. The QCQP is reformulated into an SDP, and the optimal solution is obtained using rank-1 relaxation. From the optimal solution of the relaxed SDP, normal vectors are computed for 3D reconstruction of planar surfaces.
We also propose a method to compute camera extrinsic parameters by utilizing planar surfaces which are perpendicular to the ground. This man-made environment is quite reasonable for end-of-line calibration and indoor camera calibration. Main benefit of the proposed method is that it does not require 3D measurements of image features, and thus, extrinsic calibration can be conducted automatically at the start of every driving. In both synthetic simulation and real calibration environment, the proposed reconstruction method significantly outperformed the previous 3D reconstruction methods, and thus errors in extrinsic parameters were dramatically reduced.

Author Contributions

Conceptualization, S.J.L.; methodology, S.J.L., J.-W.L. and W.L.; software, S.J.L. and J.-W.L.; validation, J.-W.L.; formal analysis, S.J.L. and W.L.; data curation, J.-W.L.; writing—original draft preparation, S.J.L.; writing—review and editing, J.-W.L. and W.L.; visualization, C.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1G1A1009792). This paper was supported by research funds for newly appointed professors of Jeonbuk National University in 2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1G1A1009792). This paper was supported by research funds for newly appointed professors of Jeonbuk National University in 2020. We deeply appreciate Daehyun Ji (SAIT) and Jahoo Koo (SAIT) for their technical support and project management.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Song, S.; Chandraker, M. Robust scale estimation in real-time monocular SFM for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1566–1573. [Google Scholar] [CrossRef] [Green Version]
  2. Schonberger, J.L.; Frahm, J.M. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  3. Zhou, Z.; Jin, H.; Ma, Y. Robust plane-based structure from motion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1482–1489. [Google Scholar] [CrossRef]
  4. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  6. Bustos, A.P.; Chin, T.J.; Eriksson, A.; Reid, I. Visual SLAM: Why Bundle Adjust? In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 2385–2391. [Google Scholar] [CrossRef] [Green Version]
  7. Mouragnon, E.; Lhuillier, M.; Dhome, M.; Dekeyser, F.; Sayd, P. Real time localization and 3d reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; pp. 363–370. [Google Scholar] [CrossRef] [Green Version]
  8. Li, Y.; Ushiku, Y.; Harada, T. Pose Graph Optimization for Unsupervised Monocular Visual Odometry. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 5439–5445. [Google Scholar] [CrossRef] [Green Version]
  9. Hartley, R.I.; Sturm, P. Triangulation. Comput. Vis. Image Underst. 1997, 68, 146–157. [Google Scholar] [CrossRef]
  10. Kanatani, K.; Sugaya, Y.; Niitsuma, H. Triangulation from two views revisited: Hartley-Sturm vs.optimal correction. In Proceedings of the British Machine Vision Conference, Leeds, UK, 1–4 September 2008; pp. 18.1–18.10. [Google Scholar] [CrossRef] [Green Version]
  11. Lindstrom, P. Triangulation made easy. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1554–1561. [Google Scholar] [CrossRef]
  12. Kanatani, K.; Sugaya, Y.; Kanazawa, Y. Guide to 3D Vision Computation; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  13. Chum, O.; Pajdla, T.; Sturm, P. The geometric error for homographies. Comput. Vis. Image Underst. 2005, 97, 86–102. [Google Scholar] [CrossRef] [Green Version]
  14. Kanatani, K.; Niitsuma, H. Optimal two-view planar scene triangulation. IPSJ Trans. Comput. Vis. Appl. 2011, 3, 67–79. [Google Scholar] [CrossRef]
  15. Miksch, M.; Yang, B.; Zimmermann, K. Automatic extrinsic camera self-calibration based on homography and epipolar geometry. In Proceedings of the 2010 IEEE Intelligent Vehicles Symposium, La Jolla, CA, USA, 21–24 June 2010; pp. 832–839. [Google Scholar] [CrossRef]
  16. Bazargani, H.; Laganière, R. Camera calibration and pose estimation from planes. IEEE Instrum. Meas. Mag. 2015, 18, 20–27. [Google Scholar] [CrossRef]
  17. Haralick, R.M.; Lee, C.N.; Ottenburg, K.; Nölle, M. Analysis and solutions of the three point perspective pose estimation problem. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Maui, HI, USA, 3–6 June 1991; pp. 592–598. [Google Scholar] [CrossRef]
  18. Horaud, R.; Conio, B.; Leboulleux, O.; Lacolle, B. An analytic solution for the perspective 4-point problem. Comput. Vis. Graph. Image Process. 1989, 47, 33–44. [Google Scholar] [CrossRef] [Green Version]
  19. Quan, L.; Lan, Z. Linear n-point camera pose determination. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 774–780. [Google Scholar] [CrossRef] [Green Version]
  20. Gao, X.S.; Hou, X.R.; Tang, J.; Cheng, H.F. Complete solution classification for the perspective-three-point problem. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 930–943. [Google Scholar] [CrossRef]
  21. Lepetit, V.; Moreno-Noguer, F.; Fua, P. Epnp: An accurate o(n) solution to the pnp problem. Int. J. Comput. Vis. 2009, 81, 155–160. [Google Scholar] [CrossRef] [Green Version]
  22. Lee, J.; Sung, M.; Lee, H.; Kim, J. Neural Geometric Parser for Single Image Camera Calibration. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 541–557. [Google Scholar] [CrossRef]
  23. Itu, R.; Borza, D.; Danescu, R. Automatic extrinsic camera parameters calibration using Convolutional Neural Networks. In Proceedings of the 2017 13th IEEE International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 7–9 September 2017; pp. 273–278. [Google Scholar] [CrossRef]
  24. Antunes, M.; Barreto, J.P.; Aouada, D.; Ottersten, B. Unsupervised vanishing point detection and camera calibration from a single manhattan image with radial distortion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4288–4296. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, X.; Ban, K.; Ishii, K. Estimation of mobile robot ego-motion and obstacle depth detection by using optical flow. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, San Antonio, TX, USA, 11–14 October 2009; pp. 1770–1775. [Google Scholar] [CrossRef]
  26. Yamaguchi, K.; Kato, T.; Ninomiya, Y. Vehicle ego-motion estimation and moving object detection using a monocular camera. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR), Hong Kong, China, 20–24 August 2006; pp. 610–613. [Google Scholar] [CrossRef]
  27. Zhao, J. An efficient solution to non-minimal case essential matrix estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef] [PubMed]
  28. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge Univertity Press: Cambridge, UK, 2003. [Google Scholar]
  29. Kolupaev, A.V.; Metelyov, A.P.; Prozorov, D.E.; Kurbatova, E.E.; Kharina, N.L. Speed increasing of FOE calculation in autonomous vehicle control systems. In Proceedings of the 2018 IEEE East-West Design & Test Symposium (EWDTS), Kazan, Russia, 14–17 September 2018; pp. 1–4. [Google Scholar] [CrossRef]
  30. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  31. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence, Vancouver, BC, Canada, 24–28 August 1981; pp. 121–130. Available online: https://www.researchgate.net/publication/215458777 (accessed on 6 July 2021).
  32. Nistér, D. An efficient solution to the five-point relative pose problem. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 756–770. [Google Scholar] [CrossRef] [PubMed]
  33. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
Figure 1. Synthetic simulation environment for multiple planar reconstruction and camera extrinsic calibration.
Figure 1. Synthetic simulation environment for multiple planar reconstruction and camera extrinsic calibration.
Sensors 21 04643 g001
Figure 2. Reconstructed 3D points by using Naïve triangulation (top row), optimal planar reconstruction (middle row), and proposed reconstruction method (bottom row) in different viewpoints.
Figure 2. Reconstructed 3D points by using Naïve triangulation (top row), optimal planar reconstruction (middle row), and proposed reconstruction method (bottom row) in different viewpoints.
Sensors 21 04643 g002
Figure 3. The definition of camera and world coordinates in our vehicle model.
Figure 3. The definition of camera and world coordinates in our vehicle model.
Sensors 21 04643 g003
Figure 4. Reconstruction and calibration errors in simulation experiments.
Figure 4. Reconstruction and calibration errors in simulation experiments.
Sensors 21 04643 g004
Figure 5. Real calibration environment: (a) optical flow of background features for estimating ego-motion of the vehicle. (b) Optical flow of chessboard features for 3d reconstruction.
Figure 5. Real calibration environment: (a) optical flow of background features for estimating ego-motion of the vehicle. (b) Optical flow of chessboard features for 3d reconstruction.
Sensors 21 04643 g005
Figure 6. Calibration errors in real calibration scenario.
Figure 6. Calibration errors in real calibration scenario.
Sensors 21 04643 g006
Table 1. Reconstruction errors (mm) with respect to various amounts of pixel noise.
Table 1. Reconstruction errors (mm) with respect to various amounts of pixel noise.
MethodDistance σ
0.10.20.30.50.71.01.52.03.0
Naïve triangulation8 m15.730.846.678.0110.4156.0236.6318.2474.8
Optimal triangulation15.730.846.778.0110.4156.0236.5317.9473.8
Planar reconstruction3.77.311.417.926.537.758.077.3119.5
Propoased method (K = 2)3.26.29.815.222.432.451.270.6110.5
Propoased method (K = 4)3.05.89.114.120.530.148.167.9106.3
Naïve triangulation10 m30.160.790.3152.4210.6301.0455.4610.2938.5
Optimal triangulation30.160.790.3152.4210.5300.9455.1609.1935.4
Planar reconstruction6.913.921.034.749.370.2108.8151.9252.2
Propoased method (K = 2)6.011.618.330.043.062.399.4140.8240.1
Propoased method (K = 4)5.610.917.228.040.160.695.8138.0237.4
Table 2. Calibration accuracy in simulation environment with respect to various amounts of pixel noise.
Table 2. Calibration accuracy in simulation environment with respect to various amounts of pixel noise.
Method σ = 0.3 σ = 0.5 σ = 0.7 σ = 1.0
PitchRollHeightPitchRollHeightPitchRollHeightPitchRollHeight
Without camera-motion noise
Naïve triangulation0.98290.2859141.31.37050.4759197.32.43380.9065349.83.37272.2364483.3
Planar reconstruction0.63830.201791.71.09720.3225158.01.63170.3994234.82.22170.7261319.5
Propoased method (K = 2)0.03720.20485.30.07510.329510.90.16470.402123.60.33940.699048.8
Propoased method (K = 4)0.03120.15874.50.05570.25498.00.06830.28909.80.13300.515319.1
With camera-motion noise
Naïve triangulation3.00580.3214432.73.45680.5304498.94.51590.9182646.84.92081.7227704.0
Planar reconstruction2.86680.2067412.82.97810.3346429.93.81550.4555547.04.00490.6995574.8
Propoased method (K = 2)0.07030.215210.20.12900.348618.30.27410.479739.40.45050.718165.2
Propoased method (K = 4)0.05870.15019.20.10760.253815.10.22000.327131.70.30970.502344.2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, S.J.; Lee, J.-W.; Lee, W.; Jang, C. Constrained Multiple Planar Reconstruction for Automatic Camera Calibration of Intelligent Vehicles. Sensors 2021, 21, 4643. https://doi.org/10.3390/s21144643

AMA Style

Lee SJ, Lee J-W, Lee W, Jang C. Constrained Multiple Planar Reconstruction for Automatic Camera Calibration of Intelligent Vehicles. Sensors. 2021; 21(14):4643. https://doi.org/10.3390/s21144643

Chicago/Turabian Style

Lee, Sang Jun, Jae-Woo Lee, Wonju Lee, and Cheolhun Jang. 2021. "Constrained Multiple Planar Reconstruction for Automatic Camera Calibration of Intelligent Vehicles" Sensors 21, no. 14: 4643. https://doi.org/10.3390/s21144643

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop