You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

19 July 2018

Calibration and Noise Identification of a Rolling Shutter Camera and a Low-Cost Inertial Measurement Unit

,
and
1
School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology (GIST), Gwangju 61005, Korea
2
Korea Electronics Technology Institute (KETI), Seongnam-si 13509, Korea
3
Department of Mechanical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Gyroscopes and Accelerometers

Abstract

A low-cost inertial measurement unit (IMU) and a rolling shutter camera form a conventional device configuration for localization of a mobile platform due to their complementary properties and low costs. This paper proposes a new calibration method that jointly estimates calibration and noise parameters of the low-cost IMU and the rolling shutter camera for effective sensor fusion in which accurate sensor calibration is very critical. Based on the graybox system identification, the proposed method estimates unknown noise density so that we can minimize calibration error and its covariance by using the unscented Kalman filter. Then, we refine the estimated calibration parameters with the estimated noise density in batch manner. Experimental results on synthetic and real data demonstrate the accuracy and stability of the proposed method and show that the proposed method provides consistent results even with unknown noise density of the IMU. Furthermore, a real experiment using a commercial smartphone validates the performance of the proposed calibration method in off-the-shelf devices.

1. Introduction

An inertial measurement unit (IMU) and a camera form a widely used sensor configuration for mobile platform localization. For example, visual-inertial SLAM [1,2,3,4,5,6] is an alternative navigation method in GPS-denied environments such as tunnels and indoor areas where GPS signals are not available. In particular, a MEMS-based IMU and a rolling shutter camera, which capture images line-by-line, are commonly used owing to their low-cost sensing capability.
To use inertial measurements and images simultaneously, calibration of these two heterogeneous sensors is very critical. The calibration of a low-cost IMU and a rolling shutter camera has been studied for last decades in robotics and computer vision community [7,8,9,10,11,12,13,14,15,16,17,18,19]. Previous studies [7,8,9,10,11,12,13,14,15] have separately estimated the IMU intrinsics (i.e., bias, scale factor, and misalignment) and rolling shutter camera intrinsics (i.e., focal length, principle point, and rolling shutter readout time), and then their extrinsics (i.e., relative rotation and translation and time delay) are obtained using the given intrinsic parameters. However, separate calibration is a cumbersome and time-consuming task because each calibration has different procedures. Moreover, intrinsic calibration errors are negatively propagated to extrinsic parameters, which are estimated using pre-computed intrinsics. For these reasons, joint intrinsic and extrinsic calibration of the IMU and the camera has been proposed [16,17,18,19]. These approaches simultaneously minimize the calibration errors of intrinsics and extrinsics.
Most IMU–camera calibration methods estimate the calibration parameters with fixed noise densities of the IMU and camera measurements, which are typically handled as tuning parameters. To achieve accurate calibration, we need to know such noise densities of the IMU and camera. Fortunately, for cameras, the variation in noise densities is not large because calibration patterns (e.g., a checkerboard) are used for corner extraction. Therefore, it is dependent on feature extraction algorithms, whose accuracy in terms of pixels can be easily evaluated and is well-known [20]. In contrast, setting the IMU noise density parameters is difficult and heuristic, because IMUs have various noise densities according to the types and costs of IMUs, and the noise density is not intuitive. Furthermore, noise information of low-cost IMUs is not provided in general.
To address these problems, we jointly estimate the noise density of the low-cost IMU and the intrinsic/extrinsic calibration parameters through the graybox system identification, which estimates the unknown parameters describing the prediction of the dynamic system well [21,22,23]. In general, the graybox method is implemented on a nonlinear optimization problem that minimizes the residual between the predicted and observed measurements of filtering (i.e., the Kalman filter). The proposed framework is composed of two types of graybox methods, which include filtering and optimization, for calibration and noise identification, respectively.
For calibration, rather than finding the calibration parameters in a single optimization step, we divide the calibration process into the filtering step for initialization and the optimization step for refinement. As a result, the convergence time of our framework becomes much faster than the single optimization method. This is because it finds sub-optimal calibration parameters in the filtering step, whose computation time is more efficient than the optimization, and then uses the parameters as an initial value of optimization. Hence, the good initial estimates from filtering reduces the convergence time of optimization for the calibration problem. In fact, it is difficult to design both the calibration and noise parameter estimation problems together in the single optimization step. In the filtering step, we utilize the unscented Kalman filter (UKF) to address the strong nonlinearity of our calibration problem, and its state vector consists of the intrinsic/extrinsic calibration parameters as well as the IMU motion parameters including the position, velocity, and orientation. In the optimization step, we refine the calibration parameters with the noise density estimated from the previous filtering step, while the UKF only estimates the IMU motion including the position, velocity, and orientation.
For noise identification, the optimizer estimates the noise density that attempts to make the UKF converge properly while enabling the estimation of calibration parameters by UKF. Unlike the previous graybox method [23], we define both the prediction errors and state error covariances, which is a constraint, as the cost function of optimization. The IMU noise density is parameterized in terms of system noise covariances in UKF, and both the prediction errors and state error covariances in the cost function involve the system noise covariance in each frame. Without the constraint, the optimization solver forces the system noise covariance to become very large, because it estimates the noise density, which is over-fitted to measurements minimizing the errors (i.e., residuals) in UKF. To avoid such divergence, the proposed method not only minimizes the residuals but also constrains the state error covariances. The initialization of the noise parameters for this optimization is performed using a grid search strategy with the cost function for noise estimation.
We demonstrate the overall procedure of the proposed method with the input and output in Figure 1, and summarize the outputs as follows.
Figure 1. Overall framework of the proposed algorithm. x 0 is the initial state vector and θ i , c , e indicates IMU intrinsics (i), camera intrinsics (c), and extrinsics (e). u and z denote inertial and visual measurements, respectively. R is a known measurement noise covariance. The notation · ^ denotes the final estimates and subscript 0 denotes an initial step.
  • IMU intrinsics: biases, scale factors, and misalignments
  • Camera intrinsics: focal length, radial distortion coefficients, and rolling shutter readout time
  • IMU–camera extrinsics: relative rotation, translation, and time delay between an IMU and a camera
  • IMU noise parameters: noise density of a gyroscope and an accelerometer
The remainder of this paper is organized as follows. Section 2 briefly reviews the previous studies related to IMU–camera calibration. Section 3 illustrates the problem formulation for our calibration method. A detailed description about the calibration system model and its measurement model is provided in Section 4. Section 5 describes the framework and the implementation of our calibration method. Section 6 analyzes the experimental results to validate the advantages of the proposed calibration method. Finally, we conclude the paper in Section 7.

3. Problem Formulation

Our system comprises a low-cost IMU and a rolling shutter camera, and these two sensors are rigidly connected as in Figure 2. The IMU measures 3-DOF acceleration and 3-DOF angular velocity, and a rolling shutter camera captures an image row-by-row. In this paper, we represent the coordinates of the IMU, the camera, and the world by { I } , { C } , and { W } , respectively. These notations are used as a left superscript to denote a reference coordinate and a left subscript to represent a target coordinate, as shown in Figure 2. For example, the 3D position p of the IMU coordinate { I } from the world coordinate { W } is represented by I W p , and the 3D translation t of the camera { C } coordinate from the IMU coordinate { I } is denoted by C I t .
Figure 2. Schematic overview and the calibration coordinate description for the rigidly connected low-cost IMU and rolling shutter camera. The notations ( { I } , { C } , and { W } ) denote the coordinates of the IMU, the rolling shutter camera, and the world, respectively.
To generate calibration data, we sequentially obtain inertial measurements and images by capturing a checkerboard along three-axis linear and angular motion. The inertial measurements u from the IMU are composed of acceleration a R 3 and angular velocity w R 3 . Visual measurements z R 2 × M comprise M corners of the checkerboard in the image coordinate. Given the visual and inertial measurements, we formulate the calibration and noise identification as a minimization problem as follows.
θ ^ = min θ E ( θ ; u , z ) ,
where the cost function E ( · ) illustrates the alignment error between motions of the IMU and the camera, which is computed from the visual and inertial measurements u , z , with the calibration and noise parameters denoted by θ = { θ i , θ c , θ e , θ n } . The intrinsic parameter θ i of the IMU describes the inherent difference between the raw measurements of sensors and the real-world value. The intrinsic parameter θ c of the rolling shutter camera explains the mapping from the camera coordinate to the image coordinate. The extrinsic parameter θ e between the IMU and the camera describes the relative geometric difference between the IMU coordinate and the camera coordinate. The noise parameter θ n of the IMU represents noise characteristics of inertial measurements.
u = a ( u , θ i ) : IMU intrinsics z = b ( C X , θ c ) : Camera intrinsics I X = c ( C X , θ e ) : IMU-camera extrinsics θ n = d ( u ) : IMU noises
where u indicates the intrinsically calibrated inertial measurements, and X C R 3 indicates the position of a three-dimensional point in a camera coordinate. a ( · ) , b ( · ) , c ( · ) , and d ( · ) are functions for describing each parameter. The calibration and noise parameters are composed of several variables. Table 2 shows a detailed description of the unknown variables in each parameter.
Table 2. The parameter description.

4. Model

In this section, we describe the system and measurement models for calibration as follows.
x k = F x k 1 , u k 1 + w k 1 z k = H x k + v k ,
where the state vector x represents the sensor motion state and calibration parameters, and are propagated with the inertial measurements u based on the system model F · . The nonlinear measurement model H · describes the relation between the state vector and the measurement z . w is a process noise and v is a measurement noise. Both are assumed to be white Gaussian noise, w N ( 0 , Q ) and v N ( 0 , G ) with the noise covariance Q and G , respectively. Subscripts k 1 and k denote time steps. We explain the aforementioned notations in detail as follows.
We include the sensor motion parameters in the state vector. Therefore, the state vector x R 40 is composed of a motion state ρ , extrinsic parameters θ e , IMU intrinsic parameters θ i , and camera intrinsic parameters θ c .
x = ρ θ e θ i θ c .
Here, the motion state ρ R 10 is represented by
ρ = I W p I W v I W q ,
where I W p R 3 and I W v R 3 indicate the position and velocity of the IMU coordinate from the world coordinate, respectively; and I W q R 4 is an orientation expressed as a unit quaternion of the IMU coordinate from the world coordinate.
The extrinsic parameter θ e R 8 is denoted by
θ e = I C t I C q λ d ,
where I C t R 3 is a translation from the IMU coordinate to the camera coordinate, and I C q R 4 is a rotation represented as a unit quaternion from the IMU coordinate to the camera coordinate. λ d is a temporal delay between the IMU and the camera.
The IMU intrinsic parameter θ i R 18 is represented by
θ i = b a b g m a m g s a s g ,
where b a R 3 , m a R 3 , and s a R 3 indicate a bias, a misalignment, and a scale factor of the accelerometer, respectively. b g R 3 , m g R 3 , and s g R 3 are the parameters of the gyroscope.
The camera’s intrinsic parameter θ c R 4 is represented by
θ c = f d λ r s ,
where f R is the focal length, d R 2 is radial distortion, and λ r s R is the rolling shutter readout time of the rolling shutter camera.

4.1. System Model: Low-Cost IMU

We describe the system model F ( · ) in Equation (3) performing the transition of the state vector x . The visual and inertial measurements are obtained at different frame rates and their time intervals have some variation. An example of this time interval is demonstrated in Figure 3.
Figure 3. Time diagram for visual and inertial measurements. The green and red vertical lines indicate the timestamps of the IMU and camera measurements, respectively.
The state transition (or prediction) is sequentially performed with the inertial measurements u s until a new visual measurement is obtained, as demonstrated in Figure 3. The number of predictions becomes # p r e d i c t i o n = # u s + 1 . We define the time interval Δ t for each prediction as follows.
Δ t = t i m u , 1 ( t i m g , k 1 + λ d ) : First t i m u , s t i m u , s 1 : Middle ( t i m g , k + λ d ) t i m u , e n d : Last
where k is the time-step of the visual measurements, while s is that of the inertial measurements.
The nonlinear system model in Equation (3) is defined as follows.
F x k 1 , u s = F m o t ( ρ k 1 , u s ) θ e , k 1 θ i , k 1 θ c , k 1 .
Here, the motion state is sequentially predicted with the IMU measurement u s and the time interval Δt, and the other calibration parameters are considered as constant because they are static and two sensors are rigidly connected. Note that we do not regard the bias states as smoothly time varying parameters, unlike [19] because the input inertial measurements for calibration are obtained for a short time (about 1–2 min).
F m o t ρ k 1 , u s = I W p k 1 + I W v k 1 Δ t I W v k 1 + I W a s Δ t I W q k 1 + Ω w s Δ t 2 I W q k 1 ,
where the quaternion kinematic function Ω : R 3 R 4 × 4 and the angular velocity w s explain the time variation of IMU orientation. The linear acceleration I W a s and angular velocity w s are converted from the raw measurements a m and w m via the intrinsic parameters of the IMU as follows.
I W a s w s = I W R S a M a a m + b a g S g M g w m + b w ,
where I W R R 3 × 3 is a rotation matrix expressing the orientation of the IMU, which is converted from I W q , and g is the gravity vector in the world coordinate. The scale factor matrices S a and S g are generated by the scale factor states s a and s g as
S a = s a , x 0 0 0 s a , y 0 0 0 s a , z , S w = s g , x 0 0 0 s g , y 0 0 0 s g , z
and the misalignment matrices (i.e., M a and M g ) between the orthogonal IMU coordinate and the accelerometer/gyroscope coordinates are represented as below
M a = 1 m a , x m a , y 0 1 m a , z 0 0 1 , M g = 1 m g , x m g , y 0 1 m g , z 0 0 1 .
The lower triangular elements become zero because we assume that the x-axes of the accelerometer and the gyroscope coincide to one of the orthogonal IMU coordinates as in [13,19].

4.2. Measurement Model: Rolling Shutter Camera

The measurement model in Equation (3) describes the relation between the state vector x and the measurements z . The visual measurement z is a set of calibration pattern corners obtained from an image. Unlike a global shutter camera, a low-cost rolling shutter camera captures an image row-by-row. The projective geometry of a rolling shutter camera is formulated as follows.
u v K C v C h 2 R C v C h 2 t r s W C R W C t 0 1 W X ,
where W X P 3 denotes a 3D point in the world coordinate. The intrinsic matrix is represented by K R 3 × 3 . The notation ∼ denotes normalization of a projected point. The rotation matrix and translation vector of the world coordinate from a camera coordinate are represented by W C R SO ( 3 ) and W C t R 3 , respectively. They are computed from our state vector, which is the position I W p and orientation I W q of the IMU from the world coordinate and transformation C I t , C I q between the IMU and the camera. C v C h 2 R and C v C h 2 t denote rolling shutter transformation from the center row ( h 2 ) to a row ( v ) of the projected point because we use the center row as the reference row of an image when defining the rolling shutter distortion. Here, h is the height of the image. To compute the rolling shutter transformation, we use the Cayley transform model described in [40].
C v C v 2 R R h 2 v λ r s C w , C v C v 2 t = h 2 v λ r s C v ,
where λ r s is the rolling shutter readout time included in the state vector and C w and C v are angular and linear velocities of the camera. The angular and linear velocities are computed from gyroscope measurements and states.
C w = I C R I w s , C v W C R I W v k | k 1 ,
where I C R is the rotation of the camera from the IMU, I w s is an intrinsically calibrated angular velocity of the IMU in Equation (12), and I W v k | k 1 is a predicted velocity of the IMU. We use them at the closest point to the time step t i m g + λ d , as follows.
Finally, each corner z i z in the visual measurement is obtained by considering radial distortion.
z i = u v = H ( x ; W X ) = ( 1 + d 1 r 2 + d 2 r 4 ) u v ,
where the subscript i is an index of a projected point and { d 1 , d 2 } d is the radial distortion state. To utilize multiple points, we concatenate the corner positions in the image coordinate as follows.
z = z 1 z M ,
where M is the number of corner points.

5. Proposed Method

Our calibration and noise identification is based on the graybox system identification [22,41], which is a nonlinear optimization method minimizing the residuals of the Kalman filter (KF). Figure 1 demonstrates the overall framework of the proposed method. We first initialize the noise parameters by grid search because we do not have any prior knowledge on the noise. Then, we estimate the calibration and noise parameters with our novel graybox method, whose cost function considers not only residuals but also state error covariances of KF. Here, the noise parameters are estimated by nonlinear optimization, while the UKF estimates the motion state and the calibration parameters together. Finally, in the refinement step, the optimization module estimates the calibration parameters using the conventional graybox-based calibration [23], whose cost function considers only residuals, while the UKF module only estimates the motion state.
We define the cost function in Equation (1) as a square form of the state error covariances of the UKF and the residuals between the predicted and observed measurements.
E 1 ( θ n ) = k = 1 M z k z k | k 1 2 + d i a g ( S k | k ) 2 , S k | k = H P k | k H
where d i a g ( · ) is the diagonal of a square matrix and M is the number of visual measurements. The covariance S k | k is computed in filtering with the updated state error covariance P k | k and a Jacobian matrix of the measurement function H . The computation of the predicted measurements and the state error covariances in UKF is described in Equation (22).
For the initialization of the noise parameters, we use the same cost function and a grid search strategy. The search range is 0 < θ n < 1 , and the grid size is 5 with log-scale intervals. Figure 4 demonstrates the cost map computed by the grid search. Then, we estimate the noise and calibration parameters using the graybox method. For the optimization, we exploit the Levenberg–Marquardt algorithm, which is a nonlinear least square solver.
arg min θ n E 1 ( θ n )
Figure 4. Cost map for noise parameter estimation. E 1 ( θ n ) ¯ is the normalized cost function defined in Equation (19).
At this time, the calibration parameters are simultaneously estimated by the UKF. Therefore, we estimate the noise parameters that result in the optimal convergence of the UKF for calibration.

5.1. Noise and Calibration Parameters in UKF

We adopt the UKF [42] owing to its efficient performance on the nonlinear system and measurement model. The prediction of the state x k 1 | k 1 and its error covariance P k 1 | k 1 are formulated as below.
χ k 1 | k 1 i = s p ( x k 1 | k 1 , P k 1 | k 1 , Q ( θ n ) ) χ k | k 1 i = F ( χ k 1 | k 1 i , u k 1 , θ i ) x k | k 1 = i W s i χ k | k 1 i P k | k 1 = i W c i [ χ k | k 1 i x k | k 1 ] [ χ k | k 1 i x k | k 1 ]
where s p is the sigma point generation function, χ is the generated sigma points of the state, the superscript i is the index of the sigma point ( i = 1 , , N ), and W s , W c are weights of sigma points for the state and covariance, respectively. The sigma points χ k 1 | k 1 i are generated with the state x k 1 | k 1 , the state error covariance P k 1 | k 1 , and the system noise covariance Q . Each sigma point is predicted using system model F ( · ) with the inertial measurement u k 1 and the IMU intrinsic parameter θ i . The system noise covariance Q ρ for motion states is defined from the IMU noise parameter { σ a , σ g } θ n . The system noise covariances for other states are set to zero because they are static. As mentioned before, we regard the bias states as constant parameters, unlike [19].
Q ρ = σ a Δ t 2 2 I 3 0 3 0 3 × 4 0 3 σ a Δ t 2 I 3 0 3 × 4 0 4 × 3 0 4 × 3 σ g Δ t 2 I 4
The update of the predicted state x k | k 1 and the predicted state error covariance P k | k 1 are formulated as follows.
χ k | k 1 i = s p ( x k | k 1 , P k | k 1 , R ) γ k | k 1 i = H ( χ k | k 1 i , θ c , θ e ) z k | k 1 = i W s i γ k | k 1 i P z k z k = i W c i [ γ k i z k | k 1 ] [ γ k i z k | k 1 ] P x k z k = i W c i [ χ k | k 1 i x k | k 1 ] [ γ k i z k | k 1 ] K k = P x k z k P z k z k 1 x k | k = x k | k 1 + K k ( z k z k | k 1 ) P k | k = P k | k 1 K k P k | k 1 K k .
The predicted sigma point χ k | k 1 i is transformed through the measurement model H ( · ) with the camera intrinsic parameter θ c and the IMU–camera extrinsic parameter θ i . Then, the Kalman gain K k is computed with the predicted measurement covariance P ^ z k z k and state-measurement cross-covariance matrix P ^ x k z k . The predicted state and state error covariance are updated with the Kalman gain. The measurement noise covariance G is defined as a block-diagonal matrix.
G = σ z 2 I 2 0 2 0 2 0 2 0 2 0 2 0 2 σ z 2 I 2

5.2. Refinement

We refine the estimated calibration parameters with the conventional graybox method. Here, we only use the residuals between the predicted measurement and the observed measurement in the cost function, and the filter module only estimates the IMU motion state. As a result, the fixed system noise covariance Q ρ R 10 , instead of Q R 40 , for motion states is used in UKF. However, unlike [23], which estimated only extrinsic parameters, we estimate the intrinsic parameters of the IMU and the camera as well as the extrinsic parameters. Therefore, the cost function in Equation (1) is defined as follows:
arg min θ i , c , e E 2 ( θ i , c , e ) , E 2 ( θ i , c , e ) = k = 1 N z k z k | k 1 2 .
Since this optimization is performed in batch manner, unlike calibration parameter estimation by filtering in the previous step, outliers or disturbances such as abrupt motion or illumination changes can be corrected in this step.

5.3. UKF Initialization

In this section, we describe the initialization of the states, the state error covariance, and the measurement noise covariance for the UKF. The initial position and orientation of the IMU are computed by transforming the position and the orientation of the camera at the first frame through the initial relative rotation and translation between the IMU and the camera. The velocity of the IMU is initialized to zero. The initial focal length f is set to the width of the image. The relative rotation between the IMU and the camera I C q is initialized by the angular velocities of the IMU and estimated rotation from the camera [15]. The IMU provides the raw angular velocity measurements. To compute angular velocities of the camera, we estimate the camera orientations with respect to the checkerboard pattern using homography [24]. Then, the angular velocity is obtained from the derivative of the orientation of the camera. The other parameters—the relative translation I C t and time delay λ d between the IMU and the camera, biases b a , b g , misalignments m a , m g , distortion coefficient d , and rolling shutter readout time λ r s —are initially set to zero, and scale factors s a , s g are set to one. The initial state error covariance matrix is empirically set as shown in Table 3. Based on the pixel localization error of the corner extraction algorithm, the standard deviation σ z for the measurement noise covariance is set to 1. The gravity g in the world coordinate is set to [ 0 , 0 , 9.81 ] , and the z-axis of the world coordinate is the vertical direction. The initial calibration parameters for the refinement step are obtained from the estimated calibration parameter from the UKF in the calibration and noise identification step.
Table 3. Initial state error covariance. The elements in a variable are set to equal values.

6. Experimental Results

We evaluate the proposed method on the synthetic and real data. The synthetic data contain synthetically generated motion information of an IMU–camera system and corresponding calibration pattern points. Note that the two sensors are rigidly connected. Using the synthetic data, we validate the accuracy and stability of the proposed method by running 100 Monte-Carlo simulations with fixed noise parameters. Furthermore, we compare the calibration results with and without noise identification using random noise parameters to demonstrate the effect of noise estimation. In real-data experiments, we utilize two experimental setups: a self-designed system, which is equipped with two rolling shutter cameras and one low-cost IMU, and a commercial smartphone, which has a rolling shutter camera and a low-cost IMU. The first setup is specially designed to evaluate the extrinsic calibration accuracy. Since the ground truth cannot be obtained for real-data experiments, we indirectly measure the performance based on loop-closing errors between the two cameras and the IMU. With these two setups, we first analyze the performance of the proposed method based on the standard deviation of the estimate. We also compare the proposed method with hand-measure estimates in terms of extrinsics and existing calibration methods in terms of intrinsics and extrinsics for reference. These comparisons show that the proposed method produces comparable results on the camera intrinsic calibration. In addition, the smartphone experiments show that the proposed method is more robust to measurement noises than the existing calibration method.

6.1. Synthetic Data

We generate the synthetic measurements for a frame length of 60 s. For this, we synthetically make 20 corners of a checkerboard pattern in the world coordinate and a smooth trajectory of the mobile platform by applying a sinusoidal function, as shown in Figure 5a. The points in the world coordinate are projected to the image coordinate as in [40], and are used as synthetic visual measurements for IMU–camera calibration. We set the image resolution to 1280 × 960. The focal length is set to 700 and the radial distortions are set to [0.1 −0.1] . The rolling shutter readout time λ r s is set to 41.8 μ s. The standard deviation of Gaussian noises for the visual measurements is set to 1 pixel. The IMU acceleration and angular velocity measurements along the world coordinate are obtained by differentiating the trajectory of the IMU. Then, to generate uncalibrated measurements, intrinsic parameters including a scale factor, a misalignment, a bias, and gravity in the world coordinate are used to perform the inertial measurements based on Equation (12). The scale factors ( s a , s g ), misalignments ( m a , m g ), and biases ( b a , b g ) are set to 1.1, 1.1, 0.03, 0.03, 0.1, and 0.1, respectively. These parameters are represented by a 3D vector and the elements of each vector are set to the same value. For instance, m a = [ 0.03 , 0.03 , 0.03 ] . Finally, we apply additive white Gaussian noise to the inertial measurements. Figure 5b,c shows an example of the synthetic inertial measurements. The frame rates of the IMU and the camera measurements are set to 100 and 25 Hz, respectively. For the extrinsic parameters, the relative rotation, relative translation, and time delay between the IMU and the camera are set to [50 −50 50] , [50 50 −50] mm, and 100 ms, respectively. The calibration parameters are initialized as described in Section 5.3.
Figure 5. Synthetic data for the IMU–camera calibration: (a) trajectory of the rigidly connected IMU and camera; (b) synthesized acceleration measurements; and (c) synthesized angular velocity measurements. The blue, orange, and yellow color lines in (b,c) indicate the x-, y-, and z-axis measurements.
In the first experiment, we evaluate the proposed method with fixed IMU noise parameters, which represent the noise density of the IMU set to σ a = 0.1 m/s 2 / Hz , and σ g = 0.1 /s/ Hz . The purpose of this experiment is to show how accurately and stably the proposed method estimates the calibration and noise parameters. Figure 6 shows the estimates of the calibration parameters and the standard deviation of their errors, which are a square root of diagonals of the state error covariance, over the 60-s frame length in the calibration and noise identification step. In our framework, the calibration parameters and their covariances are estimated by the UKF while estimating the noise parameters using nonlinear optimization. The graphs in Figure 6 show the calibration estimates obtained by the UKF at optimal noise parameter estimates by nonlinear optimization. In Figure 6a, the errors of the states, including the rotation C I R , translation C I t , time delay λ d , biases b a and b g , misalignments m a and m w , scale factors s a and s w , and focal length f, the radial distortion d and the rolling shutter readout time λ r s , are rapidly converged to zero. The errors are computed using the ground truth calibration parameters. The rotation errors are displayed in the Euler angle form for intuitive representation. In Figure 6b, the error standard deviations obtained by the UKF are rapidly converged.
Figure 6. State errors and their covariances: The state contains calibration parameters, which are estimated by the filter of the proposed method on the synthetic data. The noise parameter is set to the mean of the estimates by the proposed method ( σ a = 0.1 m/s 2 / Hz , σ g = 0.1 deg/s/ Hz ). The blue, orange, and yellow color lines of the 3-dimensional states and covariances ( ϕ , t , b a , b g , s a , s g , m a , and m g ) mean the x-, y-, and z-axis estimates. The blue and orange color lines of the 2-dimensional state and covariance d mean d 1 , d 2 . The 1-dimensional states and covariances ( λ d , λ r s , and f) are represented by the blue color lines: (a) state error; and (b) state error covariance.
Table 4 summarizes the final calibration and noise parameter estimates of the proposed method by running 100 Monte-Carlo simulations. We describe the mean and standard deviation of the estimates to show statistically meaningful results. Table 4d shows the noise parameter estimates obtained by the proposed method. We find that the ground truth of the noise parameter is within the error bound of the noise parameter estimates. This result indicates that the proposed noise identification reasonably estimates the noise parameters. Therefore, the estimated noise parameters can be used to further refine the calibration parameters in the refinement step. Table 4a–c shows the extrinsic and intrinsic parameter estimates of the IMU and the camera. The differences between the estimates and the ground truth are smaller than their standard deviations. This validates that the proposed method accurately estimates all calibration parameters and noise parameters.
In the second experiment, we generate 100 sequences based on different IMU noise parameters, and, using these sequences, we analyze the proposed calibration method with and without the noise density estimation. The purpose of this experiment is to show that how the proposed noise identification robustly estimates the calibration parameters under unknown IMU noise density. The noise parameters σ a , σ w are randomly selected from 0.001 to 0.2, and the generated noises are added to the inertial measurements. These values cover most of the noise characteristics from low- to high-cost IMUs. We run the proposed method with noise estimation and fixed noise parameters ( σ a / σ w = 0.2/0.2 and 0.001/0.001). Table 5 shows the calibration parameter estimates of the proposed method with and without noise estimation. With noise estimation, the proposed calibration method estimates accurate calibration parameters. The mean of errors between the estimated results and the ground truth is smaller than the standard deviation of the estimates. On the contrary, fixed noise densities cause inaccurate and unstable estimates. The mean of the translation estimates without noise estimation is out of the error bound, and their standard deviation is about two times larger than that of the proposed method with noise estimation. For bias estimates, the mean of the estimates is not within the error bound and their standard deviation is larger. Besides, the estimates of the focal length and rolling shutter readout time are largely affected by the IMU noise parameters, whereas the rotation, misalignments, scale factors, and radial distortions are not dependent on the IMU noise parameters. These results validate that the proposed noise identification approach improves the accuracy and stability of the calibration parameter estimation, especially under unknown IMU noise density.

6.2. Real Data

For the real-data experiments, we use two setups: a self-designed system and a commercial smartphone. Since it is difficult to obtain ground truth for real data, we indirectly evaluate the performance of the proposed method using the results of existing methods and the hand-measured values. In addition, we use the standard deviation of estimates as an evaluation metric to analyze the algorithm stability. Moreover, the loop closing error metric with two cameras and the IMU is utilized to evaluate extrinsic calibration performance.
At first, we evaluate the proposed method using the self-designed system, which consists of two rolling shutter cameras and a low-cost IMU. Figure 7a shows our experimental setup. The rolling shutter cameras are Logitech C80 (Logitech, Lausanne, Switzerland) and the low-cost IMU is Withrobot myAHRS+. The image resolution is 640 × 480, and the frame rates of the camera and the IMU are, respectively, 30 and 100 Hz. We record a set of measurements for approximately 90 s and perform experiments on the 10 datasets.
Figure 7. Two types of the low-cost IMU and the rolling shutter camera setup for real-data experiments: (a) our self-designed system having two rolling shutter cameras and a low-cost IMU; and (b) Samsung Galaxy Alpha (Samsung, Seoul, South Korea) having a rolling shutter camera and a low-cost IMU.
Table 6 demonstrates calibration parameter estimates of camera 1 and the IMU. For extrinsic and IMU intrinsic parameters, we compare our method with the existing calibration method (i.e., the Kalibr-imucam method [19]) and the hand-measured values. The noise density of the IMU for the Kalibr-imucam method is set to the noise density estimated by the proposed method. Table 6a describes the extrinsic parameter estimates of the proposed method, the Kalibr-imucam method, and the hand-measured values. The rotation parameter C I q is parameterized by Euler angle ϕ instead of unit quaternion for intuitive understanding. The average rotation, translation, and time delay estimates of the proposed and Kalibr-imucam methods are close to each other, and the average rotation and translation estimates of both methods are within the error bound of the hand-measured value. In addition, the standard deviation of the extrinsic parameter estimates from both methods is very small. This comparison shows that the proposed method successfully estimates the extrinsic parameters; moreover, the estimated noise parameters from the proposed method lead to the successful calibration of the existing calibration method.
Furthermore, with the two rolling shutter cameras, we evaluate the extrinsic calibration performance using the loop-closing error metric, which is defined as the transformation error δ T = C 1 I T C 2 C 1 T C 2 I T 1 between the low-cost IMU and the two cameras (i.e., C 1 and C 2 ). Here, C 1 I T represents the relative transformation from the IMU to camera 1, C 2 C 1 T represents the relative transformation between the two cameras, and I C 2 T denotes the relative transformation from camera 2 to the IMU. C 2 C 1 T is estimated through the stereo camera calibration algorithm [43], and C 1 I T and C 2 I T 1 are estimated through the IMU–camera calibration algorithm. We run 10 sequences with the left camera–IMU and right camera–IMU. Then, we compute 10 × 10 loop closing errors by using bipartite matching (i.e., total 100 sets). Table 6b shows the mean and standard deviation of the loop closing errors on rotation and translation motions. The proposed and Kalibr-imucam methods produce very small errors on rotation and translation. Since the error bound of C 2 C 1 T is about ±2 mm, the errors in our method and the Kalibr-imucam method are negligible.
Table 6c describes the IMU intrinsic parameter estimates of the proposed and Kalibr-imucam methods. We do not compare classic IMU intrinsic calibration methods [7,8,9] because they require special equipment such as robot arms to estimate intrinsic parameters. Besides, since the Kalibr-imucam method regards biases as time-varying parameters, unlike the proposed method, we do not compare it in this table. In both methods, the standard deviations of the misalignments and the scale factors are small enough, which indicates that the estimates are reliable. The results show that the z-axis bias of the accelerometer is larger than the x,y-axis biases due to the effect of the gravity vector located on the z-axis of the IMU coordinate (i.e., z-axis values of inertial measurements are much larger than those of x,y-axis). Interestingly, the misalignments are close to zero and the scale factors are close to one. This means that the IMU is somewhat intrinsically calibrated, although it is a low-cost sensor. The average estimates of the misalignments and scale factors from both methods have small differences because the Kalibr-imucam method also considers the effects of linear accelerations on gyroscopes. However, there are negligible numerical errors in the misalignments and scale factors.
For camera intrinsic parameters, we compare our method with the Kalibr-rs method [12] and the MATLAB calibration toolbox, as shown in Table 6d. The Kalibr-rs method uses image sequences, which are the same measurements as those used for the proposed method, and the MATLAB toolbox uses 20 still images of the same checker board. The focal length and radial distortion estimates of the three methods are close to each other. In addition, the average rolling shutter readout time estimates of the proposed method are close to the average estimates of the Kalibr-rs method. Although there are small differences between the estimates obtained from the three methods, they are negligible. Besides, the smaller standard deviation of the estimates from the proposed method indicates that the proposed method is more reliable than the Kalibr-rs method. This comparison of the intrinsic parameter estimates from the proposed and the two existing methods, which use various measurements, validates the camera intrinsic calibration of the proposed method.
Table 6e describes the estimated noise parameters. The standard deviation of IMU noise density estimates is small enough and their mean can be used for other sensor fusion algorithms such as IMU–camera calibration and visual-inertial SLAM methods. Although we do not compare the noise density provided from the other algorithms or manufacturers, the noise estimation performance of the proposed method is indirectly verified through comparison with extrinsic parameters of the hand-measured value and camera intrinsic parameters from the MATLAB Toolbox. In addition, the successful calibration of the Kalibr-imucam method with the noise density estimates obtained from the proposed method validates the noise estimation of the proposed method.
Second, we evaluate the proposed method using a smartphone “Samsung Galaxy Alpha”, which is equipped with a rolling shutter camera and a low-cost IMU; its configuration is shown in Figure 7b. The image resolution is 720 × 480, and the frame rates of the camera and the IMU are, respectively, 30 and 50 Hz. We record 10 sets of measurements whose frame length is about 90 s.
Table 7 shows the calibration parameter estimates of the smartphone. We compare the proposed method with the existing calibration method (i.e., Kalibr-imucam [19]) and hand-measured values, similarly to the above experiment. The average rotation estimates of the proposed method and the Kalibr-imucam method are close to those of the hand-measured value. The standard deviation of the rotation estimates from the proposed method is smaller than that obtained from the Kalibr-imucam method, and this result indicates that the proposed method is more reliable than the Kalibr-imucam method. Besides, the translation estimates of the Kalibr-imucam method are converged to unrealistic values that are zero, but our estimates are close to the hand-measured value. The average time delay estimates of the proposed method are close to those of [19]; however, the standard deviation is large because an irregular time delay occurs due to the camera module activation of the smartphone. In summary, in this experiment, the extrinsic calibration of the proposed method outperforms the Kalibri-imucam method unlike the first experiment, which uses the self-designed setup. We argue that the proposed method is more robust than the noisy measurements because our framework is based on the graybox method, which internally uses the Kalman filter. In practice, the rolling shutter camera and the IMU of the commercial smartphone contain larger noises than those of the self-designed setup, as described in Table 7d.
Table 7d demonstrates the IMU intrinsic estimates of the proposed method. Similar to the results of the first experiment, the z-axis bias of the accelerometer is larger than those of the other axes. The misalignment and scale factor estimates of the accelerometer from the proposed and Kalibri-imucam methods are, respectively, close to zero and one, and the biases and misalignments of the gyroscope are close to zero. This result indicates that they were already intrinsically calibrated. However, the scale factors of the gyroscope are calibrated through the proposed and Kalibr-imucam methods, as shown in the table. Although the scale factor estimates from two methods are slightly different, the low standard deviation of our estimates indicates that the proposed method is more reliable.
For camera intrinsic parameters, we compare our estimates to the Kalibr-rs method [12] and the MATLAB calibration toolbox, as shown in Table 7c. The focal length, radial distortion, and rolling shutter readout time estimates of the proposed method are close to the estimates of two existing methods. Although the estimates of the second coefficient of radial distortion are different for each method, the first coefficient is more important than the second. Besides, the proposed method provides the lower standard deviation of the rolling shutter readout time estimates, as in the first experiment.
Table 7d describes the estimated noise densities. They are two times larger than the noise density estimated on the smartphone setup. This means that the IMU noise of the smartphone setup is more severe than that of the self-designed setup. This experiment shows that a full calibration of off-the-shelf devices is possible with the proposed method.
Furthermore, we compare the operating time of the existing method (Kalibr [12,19]) and our method. The experimental environment for the comparison is Intel i7-4790K CPU running at a single core 4.00 GHz. Table 8 demonstrates the operating time of the Kalibri [12,19] and the proposed method. The timing statistics are measured on the 10 datasets used for the above “Samsung Galaxy Alpha” experiment. In case of the existing method, it is required to use both the Kalibr-rs [12] and the Kalibr-imucam [19] together because they separately estimate the intrinsic and extrinsic parameters, whereas, the proposed method does not. The total operating time of the Kalibr method is about 2 h in average for the full calibration. In particular, the operating time of Kalibri-rs [12] is longer than 1 h because of the iterative batch optimization for adaptive knot placement. However, the proposed method estimates noise parameters as well as calibration parameters, and it is 1.733 times faster than the Kalibr method in average.
Table 8. Timing comparison of the existing and proposed methods. Mean and standard deviation of the operating times are described together.
We also compare the prediction errors of the UKF for motion estimation with calibration parameters estimated by the Kalibr method and the proposed method. In this experiment, the states of the UKF are a position, velocity, and orientation of the IMU, and they are predicted by inertial measurements and corrected by visual measurements, which are corners of a checkerboard pattern. We record a set of inertial and visual measurements using the smartphone “Samsung Galaxy Alpha” for 60 s. The prediction errors using the Kalibr are accumulated due to the IMU noise and they result in the motion drift. However, Figure 8 shows that the calibration parameters of the proposed method reduce the mean and variance of prediction errors. The average RMSE in terms of pixel for all frames decreases from 3.03 to 2.45 (about 19.2%).
Figure 8. Prediction errors in the UKF for motion estimation with the calibration parameters of the Kalibr and the proposed method. The prediction errors of visual measurements (corners) for all frames are recorded, and a cross represents a prediction error. The color of crosses means the index of a frame (i.e., blue indicates the start of the sequence and red does the end of the sequences): (a) Kalibr-imucam [19] and Kalibr-rs [12] methods, with average RMSE of 3.03 pixel; and (b) proposed method, with average RMSE of 2.45 pixel.
Finally, we compare the localization accuracy with calibration parameters estimated by the Kalibr method and the proposed method. In this experiment, we capture image sequences and inertial measurements using the smartphone “Samsung Galaxy Alpha” in an outdoor vehicle driving environment. The trajectories are estimated by VINS-mono [44], which is an open source visual-inertial SLAM algorithm. In this experiment, we do not use loop closure correction and online calibration on the extrinsic parameter and temporal delay. Figure 9 shows the estimated trajectory with the proposed method is close to the ground truth trajectory, whereas the trajectory estimates with the Kalibr method suffers from scale drift as the trajectory becomes longer. The experimental results validate that the proposed method improves the performance of the real-world system as well.
Figure 9. Outdoor localization results with the calibration parameters of the Kalibr and the proposed method. The estimated trajectories are overlaid on Google Map. The visual and inertial measurements are collected by the smartphone “Samsung Galaxy Alpha” under vehicle driving condition. The length of the whole trajectory is about 0.8 km. The ground truth trajectory is obtained from GPS measurements.

7. Conclusions

This paper proposes a robust and accurate calibration method for a low-cost IMU and a rolling shutter camera. The proposed joint calibration estimates not only the intrinsic and extrinsic parameters but also the IMU noise parameters. To improve calibration efficiency including runtime, we divide the framework into two steps. In the first step, we roughly estimate the intrinsic and extrinsic parameters through filtering while estimating the noise parameters in the optimization module. In the second step, we refine the intrinsic and extrinsic parameters via the optimization module while estimating the sensor motion in filtering. The experimental results of the synthetic data demonstrate the superiority of our framework, and, in particular, the experiments on two real-data setups validate the performance of the proposed method in off-the-shelf devices.
As a result, the proposed method enhances the runtime about 73.3% and reduces IMU drift by 19.2% in comparison with those of the Kalibr method [12,19]. In particular, the results of the visual-inertial SLAM on the real-world system demonstrates that the proposed method outperforms the Kalibr method.

Author Contributions

C.-R.L. studied the literature and devised research concepts; C.-R.L. and K.-J.Y. designed the experiments; C.-R.L. conducted the experiments and analyzed the results, and took charge of the entire manuscript; and J.H.Y. and K.-J.Y. provided the crucial intellectual support and intensively revised the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (NRF-2018R1A2B3008640) and the Technology Innovation Program (10083646, Development of Deep Learning based Future Prediction and Risk Assessment Technology considering Inter-Vehicular Interaction in Cut-in Scenario) funded By the Ministry of Trade, Industry & Energy (MOTIE, Korea).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alves, J.; Lobo, J.; Dias, J. Camera-Inertial Sensor Modeling and Alignment for Visual Navigation. Mach. Intell. Robot. Control 2003, 5, 103–112. [Google Scholar]
  2. Kelly, J.; Sukhatme, G.S. Visual-Inertial Sensor Fusion: Localization, Mapping and Sensor-to-Sensor Self-Calibration. Int. J. Robot. Res. 2011, 30, 56–79. [Google Scholar] [CrossRef]
  3. Laurent Kneip, M.C.; Siegwart, R. Robust Real-Time Visual Odometry with a Single Camera and an IMU. In Proceedings of the British Machine Vision Conference, Dundee, UK, 29 August–2 September 2011. [Google Scholar]
  4. Huang, G.; Kaess, M.; Leonard, J.J. Towards Consistent Visual-Inertial Navigation. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  5. Liu, Y.; Chen, Z.; Zheng, W.; Wang, H.; Liu, J. Monocular Visual-Inertial SLAM: Continuous Preintegration and Reliable Initialization. Sensors 2017, 17, 2613. [Google Scholar] [CrossRef] [PubMed]
  6. Caruso, D.; Eudes, A.; Sanfourche, M.; Vissière, D.; Besnerais, G.L. A Robust Indoor/Outdoor Navigation Filter Fusing Data from Vision and Magneto-Inertial Measurement Unit. Sensors 2017, 17, 2795. [Google Scholar] [CrossRef] [PubMed]
  7. Chatfield, A.B. Fundamentals of High Accuracy Inertial Navigation; Reston, V.A., Ed.; American Institute of Aeronautics and Astronautics, Inc.: Reston, VA, USA, 1997. [Google Scholar]
  8. Rogers, R.M. Applied Mathematics in Integrated Navigation Systems; American Institute of Aeronautics and Astronautics: Reston, VA, USA, 2003; Volume 1. [Google Scholar]
  9. Hall, J.J.; Williams, R.L., II; Grass, F. Inertial measurement unit calibration platform. J. Robot. Syst. 2000, 17, 623–632. [Google Scholar] [CrossRef]
  10. Meingast, M.; Geyer, C.; Sastry, S. Geometric models of rolling-shutter cameras. arXiv, 2005; arXiv:cs/0503076. [Google Scholar]
  11. Ringaby, E.; Forssén, P.E. Efficient video rectification and stabilisation for cell-phones. Int. J. Comput. Vis. 2012, 96, 335–352. [Google Scholar] [CrossRef]
  12. Oth, L.; Furgale, P.; Kneip, L.; Siegwart, R. Rolling Shutter Camera Calibration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  13. Tedaldi, D.; Pretto, A.; Menegatti, E. A Robust and Easy to Implement Method for IMU Calibration without External Equipments. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  14. Mirzaei, F.M.; Roumeliotis, S.I. A Kalman Filter-Based Algorithm for IMU-Camera Calibration: Observability Analysis and Performance Evaluation. IEEE Trans. Robot. Autom. 2008, 24, 1143–1156. [Google Scholar] [CrossRef]
  15. Furgale, P.; Rehder, J.; Siegwar, R. Unified Temporal and Spatial Calibration for Multi-Sensor Systems. In Proceedings of the IEEE International Conference on Robotics and Automation, Tokyo, Japan, 3–7 November 2013. [Google Scholar]
  16. Zachariah, D.; Jansson, M. Joint calibration of an inertial measurement unit and coordinate transformation parameters using a monocular camera. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation, Zurich, Switzerland, 15–17 September 2010. [Google Scholar]
  17. Li, M.; Yu, H.; Zheng, X.; Mourikis, A.I. High-fidelity sensor modeling and self-calibration in vision-aided inertial navigation. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  18. Lee, C.R.; Yoon, J.H.; Yoon, K.J. Robust Calibration of an Ultralow-Cost Inertial Measurement Unit and a Camera: Handling of Severe System Uncertainty. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  19. Rehder, J.; Nikolic, J.; Schneider, T.; Hinzmann, T.; Siegwart, R. Extending kalibr: Calibrating the Extrinsics of Multiple IMUs and of Individual Axes. In Proceedings of the IEEE International Conference on Robotics and Automation, Stockholm, Sweden, 16–21 May 2016. [Google Scholar]
  20. Rockett, P.I. Performance assessment of feature detection algorithms: A methodology and case study on corner detectors. IEEE Trans. Image Process. 2003, 12, 1668–1676. [Google Scholar] [CrossRef] [PubMed]
  21. Graebe, S. Theory and Implementation of Gray Box Identification. Ph.D. Thesis, Royal Institute of Technology, Stockholm, Sweden, 1990. [Google Scholar]
  22. Ljung, L. System Identification: Theory for the User; Prentice Hall PTR: Upper Saddle River, NJ, USA, 1999. [Google Scholar]
  23. Hol, J.D.; Schön, T.B.; Gustafsson, F. Modeling and Calibration of Inertial and Vision Sensors. Int. J. Robot. Res. 2010, 29, 231–244. [Google Scholar] [CrossRef]
  24. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  25. O’Sullivan, L.; Corke, P. Empirical modelling of rolling shutter effect. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  26. Nebot, E.; Durrant-Whyte, H. Initial calibration and alignment of low-cost inertial navigation units for land vehicle applications. J. Robot. Syst. 1999, 16, 81–92. [Google Scholar] [CrossRef]
  27. Kim, A.; Golnaraghi, M. Initial calibration of an inertial measurement unit using an optical position tracking system. In Proceedings of the IEEE Position Location and Navigation Symposium (PLANS), Monterey, CA, USA, 26–29 April 2004. [Google Scholar]
  28. Cheuk, C.M.; Lau, T.K.; Lin, K.W.; Liu, Y. Automatic calibration for inertial measurement unit. In Proceedings of the IEEE Control Automation Robotics & Vision (ICARCV), Guangzhou, China, 5–7 December 2012. [Google Scholar]
  29. Dong-Si, T.; Mourikis, A.I. Initialization in Vision-aided Inertial Navigation with Unknown Camera-IMU Calibration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012. [Google Scholar]
  30. Fleps, M.; Mair, E.; Ruepp, O.; Suppa, M.; Burschka, D. Optimization based IMU camera calibration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011. [Google Scholar]
  31. Blázquez, M.; Colomina, I. Relative INS/GNSS aerial control in integrated sensor orientation: Models and performance. ISPRS J. Photogramm. Remote Sens. 2012, 67, 120–133. [Google Scholar] [CrossRef]
  32. Cucci, D.A.; Rehak, M.; Skaloud, J. Bundle adjustment with raw inertial observations in UAV applications. ISPRS J. Photogramm. Remote Sens. 2017, 130, 1–12. [Google Scholar] [CrossRef]
  33. Karpenko, A.; Jacobs, D.; Baek, J.; Levoy, M. Digital Video Stabilization and Rolling Shutter Correction Using Gyroscopes; Technical Report; Stanford University: Stanford, CA, USA, 2011. [Google Scholar]
  34. Li, M.; Mourikis, A.I. Online Temporal Calibration for Camera-IMU Systems: Theory and Algorithms. Int. J. Robot. Res. 2014, 33, 947–964. [Google Scholar] [CrossRef]
  35. Li, M.; Mourikis, A.I. Vision-aided Inertial Navigation with Rolling-Shutter Cameras. Int. J. Robot. Res. 2014, 33, 1490–1507. [Google Scholar] [CrossRef]
  36. Board, I. IEEE Standard Specification Format Guide and Test Procedure for Single-Axis Interferometric Fiber Optic Gyros; IEEE Standard; IEEE: Piscataway, NJ, USA, 1998; pp. 952–1997. [Google Scholar]
  37. Vaccaro, R.J.; Zaki, A.S. Statistical modeling of rate gyros. IEEE Trans. Instrum. Meas. 2012, 61, 673–684. [Google Scholar] [CrossRef]
  38. Mohamed, A.; Schwarz, K. Adaptive Kalman filtering for INS/GPS. J. Geodesy 1999, 73, 193–203. [Google Scholar] [CrossRef]
  39. Nikolic, J.; Furgale, P.; Melzer, A.; Siegwart, R. Maximum likelihood identification of inertial sensor noise model parameters. IEEE Sens. J. 2016, 16, 163–176. [Google Scholar] [CrossRef]
  40. Albl, C.; Kukelova, Z.; Pajdla, T. R6P-Rolling Shutter Absolute Camera Pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  41. Ljung, L.; Soderstrom, T. Theory and Practice of Recursive Identification; MIT Press: Cambridge, MA, USA, 1987. [Google Scholar]
  42. Van der Merwe, R. Sigma-Point Kalman Filters for Probabilistic Inference in Dynamic State-Space Models. Ph.D. Thesis, Oregon Health & Science University, Portland, OR, USA, 2004. [Google Scholar]
  43. Maye, J.; Furgale, P.; Siegwart, R. Self-supervised Calibration for Robotic Systems. In Proceedings of the IEEE Intelligent Vehicles Symposium, Gold Coast, Australia, 23–26 June 2013. [Google Scholar]
  44. Lin, Y.; Gao, F.; Qin, T.; Gao, W.; Liu, T.; Wu, W.; Yang, Z.; Shen, S. Autonomous aerial navigation using monocular visual-inertial fusion. J. Field Robot. 2018, 35, 23–51. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.