You are currently viewing a new version of our website. To view the old version click .
ISPRS International Journal of Geo-Information
  • Article
  • Open Access

12 May 2018

Comprehensive Analysis of System Calibration between Optical Camera and Range Finder

,
and
1
Department of Civil &Environmental Engineering, Seoul National University, 599 Gwanak-ro, Gwanak-gu, Seoul 151-744, Korea
2
Department of Civil and Environmental Engineering, Myongji University, 116 Myongji-ro, Cheoin-gu, Yongin, Gyeonggi-do 449-728, Korea
*
Author to whom correspondence should be addressed.

Abstract

This paper describes the comprehensive analysis of system calibration between an optical camera and a range finder. The results suggest guidelines for accurate and efficient system calibration enabling high-quality data fusion. First, self-calibration procedures were carried out using a testbed designed for both the optical camera and range finder. The interior orientation parameters of the utilized sensors were precisely computed. Afterwards, 92 system calibration experiments were carried out according to different approaches and data configurations. For comparison of the various experimental results, two measures, namely the matching rate of fusion data and the standard deviation of relative orientation parameters derived after system calibration procedures, were considered. Among the 92 experimental cases, the best result (the matching rate of 99.08%) was shown for the use of the one-step system calibration method and six datasets from multiple columns. Also, the root mean square values of the residuals after the self- and system calibrations were less than 0.8 and 0.6 pixels, respectively. In an overall evaluation, it was confirmed that the one-step system calibration method using four or more datasets provided more stable and accurate relative orientation parameters and data fusion results than the other cases.

1. Introduction

In recent years, 3D modeling has been applied in various areas including public safety, virtual environments, fire and police planning, location-based services, environmental monitoring, intelligent transportation, structural health monitoring, underground construction, motion capturing, and so on [1]. In general, 3D modeling usually can be constructed through several procedures after acquiring the data from a sensor system. The systems for 3D modeling are largely divided into single-modal and multi-modal types. The single-modal system consists of one type of sensor: in most cases, multiple optical cameras. By contrast, the multi-modal system comprises at least two types of sensors, such as laser scanning sensors, optical cameras, and range finders [2]. More specifically, the single-modal system can be built at relatively low cost and has the advantage of data production with high-resolution RGB-D data (color and depth). As a disadvantage, however, when area images of areas with low levels of information on feature points or with repetitive similar textures are considered, the accuracy of the 3D model of the area is markedly lowered.
On the other hand, the multi-modal system generally consists of optical cameras providing color information and other sensors giving depth information, such as light detection and ranging (LiDAR), laser line scanners, and range finders [3]. The system has the advantage of the fact that the shapes of objects are reconstructed more accurately [3]. Also, the quality of data in areas lacking texture or with discontinuous depths is higher than in the single-modal system [4].
Because of the benefits of using the multi-modal system, various types of data acquisition systems with different sensors have recently been developed [3]. The sensor system comprised of optical cameras and range finders is one of the well-known multi-modal systems. An intensity image from the range finder can also be provided in the form of a 2D matrix, similar to an optical camera, and each image pixel has a depth value. Also, as range finders are relatively lower-priced than laser line scanners and LiDAR, the sensor system can be configured at lower cost.
On the other hand, the optical-range-finder system also has disadvantages attributable to the characteristics of range finders. In general, range finders have lower resolutions than LiDAR. Additionally, they enable obtainment of data only within short distances, and thus, when the distance to an object becomes longer than a certain threshold, the quality of 3D models will deteriorate. Moreover, since range finders measure depth information using infrared rays, they produce high levels of noise when encountering natural light (e.g., strong sunlight). In cases of data acquisition in indoor environments, the disadvantages of using the optical-range-finder system can be significantly reduced, since the distance to indoor objects or structures is sufficiently short, and the influences of sunlight are minimized.
In this context, this paper will discuss a multi-modal system comprised of an optical camera and a range finder to be utilized in indoor environments. This research will also focus on an analysis of the factors affecting the quality of fusion data (i.e., RGB-D data) and will provide guidelines for obtaining accurate fusion results.
The remainder of this paper is organized as follows. Section 2 reviews the latest research in self- and system calibrations that are essential for data fusion using a multi-modal system. Section 3 describes the sensor geometric functional models and testbed utilized for the present calibrations. Section 4 outlines the procedures of the calibration and evaluation strategy in detail. Section 5 presents the experimental results and a related discussion. Finally, Section 6 draws conclusions on the findings of this research and looks ahead to upcoming work.

3. Sensor Geometric Functional Models and Calibration Testbed

3.1. Testbed Designed for Calibration Procedure

In this study, a testbed for the calibration procedure was constructed in consideration of the relevant optical camera and range finder characteristics (Figure 1). The testbed’s chessboard configuration and white circular patterns are used as control points (CPs) for the optical camera and the range finder, respectively. The configuration of the circular pattern target is adopted from [10,11,12] and the number of targets are slightly reduced since we have enough number of targets shown on the acquired images at the different distances. The white circle and its calculated centroid are very useful for the range finder, but not for the optical camera. To resolve this problem, additional targets with chessboard pattern were installed on the testbed plane for the optical camera (Figure 1b,c). The testbed dimensions are 4 m (height) by 4.9 m (width). As can be seen in the figure, the chessboard targets were arranged in different depths from the surface of the white circular patterns.
Figure 1. Testbed: (a) front view; (b) side view; (c) chessboard pattern target.

3.2. Sensors and Mathematical Models

The optical-range-finder system used in the experiment was configured through the installation of a Canon 6D DSLR camera (optical sensor) and a SR4000 range finder on a single aluminum frame (Figure 2). Figure 2a,b show the side view and front view of the installed optical-range-finder system, respectively. The produced sensor system provides, for each round of filming, RGB image data from the optical camera along with intensity and range data from the range finder. More detailed specifications of the sensors are shown in Table 1.
Figure 2. Installed optical-range-finder system: (a) side view; (b) front view.
Table 1. Specifications of sensors.
The imaging geometry for the RGB image data (acquired from the optical camera) and the intensity data (acquired from the range finder) are expressed in functional models based on the collinearity condition, as seen in Equations (1)–(3) ([27]). The range data (acquired from the range finder) are expressed in the Euclidean distance model, as seen in Equation (4).
x = x p f m 11 ( X X 0 ) + m 12 ( Y Y 0 ) + m 13 ( Z Z 0 ) m 31 ( X X 0 ) + m 32 ( Y Y 0 ) + m 33 ( Z Z 0 ) + Δ x
y = y p f m 21 ( X X 0 ) + m 22 ( Y Y 0 ) + m 23 ( Z Z 0 ) m 31 ( X X 0 ) + m 32 ( Y Y 0 ) + m 33 ( Z Z 0 ) + Δ y
R o t a t i o n   m a t r i x   M ( ω , φ , κ ) = [ m 11 m 12 m 13 m 21 m 22 m 23 m 31 m 32 m 33 ]
ρ = ( X X 0 ) 2 + ( Y Y 0 ) 2 + ( Z Z 0 ) 2 + Δ ρ
In the equations, x and y are the image coordinates of the sensor data. ∆x and ∆y are the distortions of the image coordinates. xp and yp are the image coordinates of the principal point and f is the focal length. m11 to m33 are the components of the rotation matrix M. ω is the primary rotation along X-axis, φ is the secondary rotation along Y-axis, and κ is the tertiary rotation along Z-axis. X, Y, and Z are the ground coordinates of the control point. X0, Y0, and Z0 are three of the exterior orientation parameters of the sensor (sensor position). ρ is the measured range value, and ∆ρ is the distortion of the range value.
Standard lens distortion models, as seen in Equations (5)–(7), are utilized for the optical camera and range finder (refer to [6,7,8]). Also, the systematic range bias (∆ρ) model for the range finder is seen in Equations (8) (refer to [10,11,12]).
Δ x = x ¯ ( K 1 r 2 + K 2 r 4 + K 3 r 6 ) + P 1 ( r 2 + 2 x ¯ 2 ) + 2 P 2 x ¯ y ¯ + A 1 x ¯ + A 2 y ¯
Δ y = y ¯ ( K 1 r 2 + K 2 r 4 + K 3 r 6 ) + 2 P 1 x ¯ y ¯ + P 2 ( r 2 + 2 y ¯ 2 )
x ¯ = x x p , y ¯ = y y p , r = x ¯ 2 + y ¯ 2
Δ ρ = d 0 + d 1 ρ + k = 1 3 [ d 2 k sin 2 k π U ρ ] + e 1 x ¯ + e 2 y ¯ + e 3 r + e r r 2 + m = 2 3 n = 0 m e 3 m + n 1 x ¯ m n y ¯ n
In the equations, K1, K2, and K3 are radial lens distortion parameters, P1 and P2 are the decentering lens distortion parameters, A1 and A2 are the electronic biases (affinity and shear), d0 is the rangefinder offset, d1 is the scale error, and d2 to d7 are the cyclic errors; U is the unit wavelength, e1 and e2 are clock skew errors, and e3 to e11 are empirical errors.

4. Calibrations and Comparative Evaluation Strategy

The experiments were conducted, as seen in Figure 3, in four phases: data acquisition, self-calibration, system calibration, and data fusion. First, in the data acquisition phase, optical and range-finder data were acquired for use in self-calibration and system calibration, respectively. The IOPs for the individual sensors were calculated through the self-calibration procedure. Afterwards, the estimated parameters were utilized for the system calibration procedure. The system calibration was performed by the one-step and two-step approaches. Finally, the fusion of optical and range-finder data was performed using ROPs calculated through the system calibration procedure. More detailed explanations of the approach are provided in the following sub-sections.
Figure 3. Calibration procedures proposed.

4.1. Self-Calibration

Since the self-calibration approaches for the individual sensors have been thoroughly reviewed and evaluated in existing studies, the present study carried out the self-calibration based on existing methods. The self-calibration of the optical sensor was first performed using the commonly used least square estimation method - bundle block adjustment [6,7,8,9]. The self-calibration of the range finder was also performed based on the one-step integrated method, which was found to be more accurate than other methods in a comparative evaluation [11].
The datasets for the self-calibration procedures were acquired using the optical-range-finder system (shown in Figure 2). They included optical camera images, range-finder intensity images, and range data for all pixels in the intensity images. Fifteen datasets were acquired for the self-calibration procedures; among them, three were obtained after vertically erecting the system to minimize the correlation between the IOPs and EOPs of the individual sensors. The locations of the data acquired for the self-calibration procedures are shown in Figure 4. The datasets were acquired at the center, left and right sides of the testbed in order to keep the images converged. After carrying out the self-calibration procedures to estimate the IOPs, the values were utilized in the system calibration procedures.
Figure 4. Data acquisition locations for self-calibration.

4.2. System Calibration

The system calibration experiments were initially carried out according to the expression of the geometrical relationship between the sensors (i.e., the one-step and two-step approaches).
More specifically, in the one-step approach, the geometric configurations of the sensors are expressed in the EOPs of the reference sensor (the optical sensor in the present study) and the ROPs between the sensors. In other words, when datasets obtained from n different locations are used for this one-step approach, the geometrical locations of the reference sensor are indicated with n different sets of EOPs (i.e., EOP1, EOP2,…, EOPn), and the locations of the range finder can be indicated with each set of EOPs and 1 set of ROPs which are constant between the optical sensor and the range finder (shown in Figure 5). The values of the EOPs and ROPs are derived directly through a bundle block adjustment procedure.
Figure 5. One-step system calibration.
On the other hand, in the two-step approach, the geometric configurations between the sensors are derived according to the EOPs of the individual sensors. More specifically, the geometrical locations of the sensor system are indicated as n sets of EOPs of the optical sensor and n sets of EOPs of the range finder (refer to Figure 6). These EOPs are derived through a bundle block adjustment procedure. Afterwards, n sets of ROPs are calculated using the derived sets of EOPs of the involved sensors. Finally, a single set of ROPs is calculated by averaging the n sets of ROPs.
Figure 6. Two-step system calibration.
The other important factor that should be considered in system calibration is the datasets’ configurations. First, 11 datasets were acquired using the optical-range-finder system, as seen in Figure 7a,b. The data acquisition locations for the system calibration were selected to have enough number of and even distribution of control points over an entire image. Figure 7a shows layout of data acquisition locations, and Figure 7b shows grouping of the locations. The locations were grouped by row (parallel direction to the testbed plane) and column (orthogonal direction to the testbed plane).
Figure 7. Data acquisition configuration for system calibration: (a) data acquisition locations of sensor system; (b) grouping of the locations.
Afterwards, 46 combinations from the 11 datasets were produced (as seen in Table 2). They are utilized to determine the effects of the number and locations of datasets on the accuracy of system calibration. Again, the datasets were selected while considering the number of datasets (i.e., ranging from 2 to 11 sets), the column selection (i.e., ranging from one column or more than one column), and the convergence (or symmetry) of the datasets. For example, in the case of 2-dataset selection from at least two columns, datasets 2 and 9 (in the second row) can be selected instead of datasets 2 and 5 or datasets 5 and 9.
Table 2. Combinations of datasets based on number and location.

4.3. Optical and Range-Finder Data Fusion

Using the IOPs and ROPs acquired from the procedures in Section 4.1 and Section 4.2, and together with the collinearity conditions (Equations (1) and (2)), RGB-D data can be created through the optical and range-finder data fusion procedure. More specifically, Figure 8 shows the conceptual idea of the data fusion method. In this figure, Popt and PRF are the respective data from the optical and range-finder data. The distortion-free 3D coordinates, (Xdf, Ydf, Zdf), can be calculated after removing the range sensor lens distortions, and range distortions from the raw range datasets. Then, the 3-dimentional coordinates w.r.t. the range finder coordinate system can be transformed into the coordinates w.r.t. the optical camera coordinate system using Equation (9). Afterwards, the corresponding image coordinates of the optical camera data can be calculated using Equations (1) and (2) while considering lens distortion effects. Subsequently, the corresponding color (R, G, B) values can be determined for the image coordinates. In this way, RGB-D data can be generated from the optical-range-finder system.
[ X o p t i c Y o p t i c Z o p t i c ] = T r o p + ( R rop ) T [ X d f Y d f Z d f ]
where (Xdf, Ydf, Zdf) are the coordinates of a certain point in the range finder data after removing distortions. Trop and Rrop are translation and rotation matrices of ROPs between the optical camera and range finder coordinate systems. (Xoptic, Yoptic, Zoptic) are the coordinates of the point w.r.t. the optical camera coordinate system.
Figure 8. Geometric relationship for data fusion using optical-range-finder system.
In this research, the authors dealt with 2 different ways of expressing the geometrical relationship between the sensors (i.e., the one-step and two-step approaches), and 46 combinations of datasets. Accordingly, 92 sets of ROPs and RGB-D datasets were produced from 46 cases with the one-step approach and 46 cases with the two-step approach.

4.4. Proposed Evaluation Strategies

The qualities of the 92 sets of ROPs estimation results were analyzed based on the standard deviations of the estimated parameters. For the case of data fusion results, a measure called matching rate is proposed and utilized in this research. In previous studies, the qualities of the data fusion results were mostly evaluated through visual inspection ([20,22,23,28,29,30]). We have 92 sets of fusion results to be compared each other, a qualitative measure is, therefore, critically necessary for the comparisons. Since the inaccurate data fusion results distinctly appear in the areas where different patches meet each other, the same color value is intentionally assigned into the same patches on the optical and range-finder data, and the color values of the corresponding points are compared. The proposed data fusion evaluation strategy can be explained in detail as follows. First, several regions suitable for quality evaluation were manually selected from the original intensity and optical data (Figure 9a,b). Then, the same color value (user-defined) was assigned for all pixels in the same region on the two images (Figure 9c,d). Lastly, data fusion was carried out with the user-defined color images according to the procedures described previously. In this step, the user-defined colors assigned for the corresponding points (i.e., Popt and PRF) were compared. At this stage, inaccurate data fusion will lead to an increased number of mismatching points (i.e., corresponding points with different color values, COLORoptCOLORRF). The white color value was assigned for mismatching points in the visual investigation (see Figure 10). For quantitative assessment, all points in the evaluation regions and matching points were counted. Afterwards, the matching rate was calculated according to Equation (10).
M a t c h i n g   r a t e = i = 1 n o   o f   d a t a s e t s n o   o f   m a t h c i n g   p o i n t s   i n   i t h   d a t a s e t t o t a l   n o   o f   p o i n t s   f o r   e v a l u a t i o n   i n   i t h   d a t a s e t n o   o f   d a t a s e t s
Figure 9. Three datasets for evaluation of data fusion results: (a) raw intensity images; (b) raw optical images; (c) intensity images with user-defined color values; (d) optical images with user-defined color values.
Figure 10. Examples of data fusion results for accuracy evaluation.
As seen in Figure 9 and Figure 10, three datasets were used for the evaluation of data fusion accuracy. These datasets consisted of two datasets in which various objects such as boxes and cylinders were placed and data acquisition locations were differentiated. The third dataset acquired in a corridor was selected for inclusion of indoor elements such as doors, windows, and pillars.

5. Experimental Results and Discussion

Experiments for self-calibration and system calibration for optical and range data fusion were carried out. More detailed explanations of the experimental results are provided in the following sub-sections.

5.1. Self-Calibration Results

Table 3 presents the estimated IOPs from the optical camera and range finder, and Table 4 shows the standard deviations of the estimated IOPs and residual accuracies. The root mean square (RMS) results of the residual values from a bundle adjustment showed 0.38 pixels (0.0025 mm) and 0.76 pixels (0.0305 mm) for the image data from the optical camera and range finder, respectively. The two RMS values for the image data were less than 0.8 pixels. The RMS value for the range data showed 6.84 mm which was better than the nominal value of 1 cm and the results from previous studies that used the same range finder [10,11]. The RMS values of 8.9 mm and 8.1 mm for the cases without and with distortion parameters were shown in [10], respectively. Also, the RMS values ranging from 12.1 mm to 13.0 mm were shown in Ref. [11]. The standard deviations of the estimated IOPs in Table 4 all showed very small amounts compared to the IOPs itself. This means that the precisions of the self-calibration parameters were high. As such, the estimated IOPs from the self-calibration procedure were accurate enough to be utilized in the system calibration.
Table 3. Estimated self-calibration parameters.
Table 4. Standard deviations of the estimated IOPs and residual accuracies.

5.2. System Calibration Results

The analysis of the system calibration and data fusion results were performed in three ways: (i) two ROP calculation methods (i.e., one-step or two-step method); (ii) different geometrical locations of datasets (i.e., single or multiple columns); (iii) different numbers of datasets.
First of all, the mean, highest, and lowest values of matching rates were derived from all of the dataset cases while applying either the one-step or two-step method (in Table 5, where the standard deviations of the estimated ROPs also are shown). Overall, the one-step method showed better results than the two-step one in terms of matching rate and parameter precision (i.e., standard deviations of ROPs).
Table 5. Matching rates of fusion data and standard deviations of ROPs.
Among the one-step cases, the highest matching rate came from the system calibration performed using six datasets in multiple columns (i.e., columns 1, 2, and 3, and rows 3, and 4; datasets 3, 4, 6, 7, 10, and 11 in Figure 7b). Moreover, this case was the most accurate one among all cases (i.e., 92 cases including the one-step and two-step approaches). The lowest rate was observed for the case using two datasets in a single column (i.e., column 2 and rows 3, and 4; datasets 6, and 7 in Figure 7b).
Considering the two-step cases only, the highest matching rate came from the case using five datasets in multiple columns (i.e., columns 1, 2, and 3, and rows 2, 3, and 4; datasets 3, 5, 6, 7, and 10 in Figure 7b). The lowest matching rate was observed for the same one-step case noted above. The most inaccurate result was shown when two-step calibration was performed using two datasets (i.e., column 1 and rows 2, and 3; datasets 2, and 3 in Figure 7b).
Secondly, the effects of the different geometrical locations of datasets on the matching rate were analyzed. In other words, the mean values and standard deviations of the matching rates derived from the single-column cases (i.e., 15 cases) were compared with those derived from the multiple-column cases (i.e., 14 cases). The comparison was carried out using the combinations having only 2 to 4 datasets, because there were no single-column cases having more than 4 datasets, as seen in Table 2.
As confirmed in Table 6 and Figure 11, the selection of datasets from multiple columns led to higher mean values regardless of the ROP calibration (i.e., one-step or two-step) method. Also, the standard deviations of the matching rates from the multiple columns were lower than those from the single columns.
Table 6. Matching rates according to data locations (single or multiple columns).
Figure 11. Comparisons of matching rates according to data acquisition locations.
Meanwhile, the standard deviations of the ROPs were derived from bundle adjustment for both the single-column cases (i.e., 15 cases) and multiple-column cases (i.e., 14 cases). Afterwards, the averaged standard deviations were calculated from the single- and multiple-column cases separately. These values are shown in Figure 12a,b; as is apparent, translation parameters (Figure 12a) and rotation parameters (Figure 12b) decreased when the number of columns was increased.
Figure 12. Comparisons of standard deviations of ROPs (results from one-step calibration) according to data acquisition locations: (a) translation (X0, Y0, and Z0); (b) rotation (ω, ϕ, and κ).
Thirdly, the relationship between the number of datasets used and the data fusion and system calibration accuracies was analyzed. At this stage, the combinations of datasets chosen only from the multiple columns (31 cases seen in Table 2) were used, since they showed better results in the previous analysis (in Table 6, Figure 11 and Figure 12).
The mean values and standard deviations of the matching rates according to the different numbers of datasets are shown in Table 7 and Figure 13. For example, 4 combination cases were used to produce a mean value and a standard deviation of matching rate for 2-dataset analysis. As seen in the table and the figure, the relationship between the number of datasets and the mean values of matching rates was not clear in the case of the one-step method. In other words, all of the cases with different numbers of datasets had similar mean values. On the other hand, the deviations of the matching rates showed a different tendency. More specifically, in most cases, the deviation values decreased as the number of datasets was increased, as seen in Table 7 and Figure 13. Using more than four datasets showed a clear trend of reducing standard deviations of matching rates.
Table 7. Mean values and standard deviations of matching rates according to number of datasets.
Figure 13. Comparison of matching rates according to number of datasets.
Meanwhile, firstly, the standard deviations of the ROPs were derived from bundle adjustment for all combination cases of multiple columns (i.e., 31 cases as seen in Table 2); afterwards, the averaged standard deviations were calculated for analysis of different numbers of datasets. For example, 4 combination cases were used to produce an averaged standard deviation for 2-dataset analysis. Figure 14a,b show such values according to the different number of datasets after application of the one-step system calibration approach. As seen in these figure, the increase of the number of datasets used led to a corresponding increase in precision (i.e., a decrease of the standard deviations of ROPs). A particularly notable point in Figure 14 is that the use of at least four datasets largely reduced the range of the standard deviations for the same number of datasets when compared with the use of two or three datasets.
Figure 14. Standard deviations of ROPs according to number of datasets (one-step approach): (a) translation (X0, Y0, and Z0); (b) rotation (ω, ϕ, and κ).
In the case of applying the two-step approach, the use of more than four datasets provided high and stable mean values of matching rate, as seen in Table 7 and Figure 13. Also, the standard deviation analysis of the ROPs (in Figure 15a,b) found that four or more datasets provided more stable standard deviations of ROPs. On the other hand, two and three dataset cases showed unstable results, as seen in Table 7, and Figure 13 and Figure 15. In the table and Figure 13, the mean of matching rates for the case of two dataset (i.e., 97.07%) was higher than the value for the case of three dataset (i.e., 95.76%), However, the standard deviations of the matching rates for these two cases were the other way around. In case of the standard deviations of ROPs as seen in Figure 15, there is discrepancy between the values for two and three dataset cases. The standard deviations for two dataset case were mostly higher than the values for three dataset case. Overall, the accuracy (i.e., matching rate) of the two dataset case is higher than the three dataset case; however, the precision (i.e., standard deviation) is vice versa. Such unstable results might come from low number of datasets and low number of combinations as well (i.e., two datasets with four combinations and three datasets with three combinations). Note that four or more dataset cases have either enough number of datasets or combinations compared to two and three dataset cases.
Figure 15. Standard deviations of ROPs according to number of datasets (two-step approach): (a) translation (X0, Y0, and Z0); (b) rotation (ω, ϕ, and κ).
Comparing Figure 14 and Figure 15, one can see that employing more datasets with the one-step approach has a more obvious positive effect on the standard deviations of ROPs than is the case with the two-step approach.
Additionally, four data fusion results among ninety-two cases were selected and illustrated in Figure 16. Table 5 already includes their matching rates and standard deviations of ROPs. The cases with the highest rate (99.08%) using one-step approach and 6 datasets in multiple columns, lowest rate (94.70%) using one-step approach and 2 datasets in single column, highest rate (98.86%) using two-step approach and 5 datasets in multiple columns, and lowest rate (92.80%) using two-step approach and 2 datasets in single column are shown in the figure. As expected, large white regions indicating mismatching points are shown in Figure 16d–f,j–l for the cases with the lowest matching rates. On the other hand, few numbers of white points are shown in Figure 16a–c,g–i for the cases with the highest matching rates.
Figure 16. Four data fusion results among ninety-two cases: (ac) highest rate using one-step approach and 6 datasets in multiple columns; (df) lowest rate using one-step approach and 2 datasets in a single column; (gi) highest rate using two-step approach and 5 datasets in multiple columns; (jl) lowest rate using two-step approach and 2 datasets in a single column.

6. Conclusions and Recommendations

This study provided and compared various system calibration results comprehensively according to the calibration approach (i.e., one-step or two-step) and the characteristics of the used datasets (i.e., different geometrical locations and number of datasets). Prerequisite accurate self-calibration procedures were performed before system calibration. Afterwards, 46 combinations of datasets were made with 11 datasets, and 92 sets of ROPs were calculated through the system calibration procedures. These sets of ROPs and their corresponding fusion data results were compared comprehensively. First of all, the matching rate and precision derived from the one-step system calibration were generally higher than those from the two-step one. Secondly, the datasets acquired from the converging configuration (i.e., multiple columns) provided better performance than did the other case (i.e., single column). Thirdly, we found that the number of datasets used was highly correlated with the precision of the calculated ROPs in cases where the one-step method was applied. In other words, the deviations of ROPs were decreased when the number of datasets used was increased. Additionally, for either the one- or two-step approach, the use of more than four datasets (from multiple columns) provided high and stable mean values of matching rates and stable standard deviations of ROPs.
Based on a comprehensive analysis of the various experimental results, we can conclude that highly accurate system calibration and data fusion results can be obtained through (1) the use of more than four datasets acquired in converging multiple directions and (2) the execution of one-step system calibration. We can also recommend that increasing the number of datasets (at least four datasets from converging multiple directions) will be advantageous to achieve relatively accurate fusion results.
Overall, this study was conducted in consideration of self-calibration, system calibration, and data acquisition configurations. Based on a comprehensive analysis of the experimental results, guidelines for accurate and efficient derivation of ROP and data fusion results are presented herein. In terms of the applications, the proposed methodology can be utilized for virtual reality, indoor modeling, indoor positioning, motion detection and reconstruction, and so on. Follow-up studies will be conducted to alleviate the range data problems arising from low resolution, noisy range data, multi-path effects, the ambiguity range, scattering problem, and other issues.

Author Contributions

K.H.C. and C.K. designed the experiments and wrote paper; K.H.C. performed the experiments; Y.K. made valuable suggestions to improve the quality of the paper.

Acknowledgments

The authors would like to acknowledge a grant from the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and future Planning (No. NRF-2015R1C1A1A02037511). The authors also are thankful to Derek Lichti at the University of Calgary, who allowed us to use his automatic circular target detection software.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chow, J.C.; Lichti, D.D.; Hol, J.D.; Bellusci, G.; Luinge, H. Imu and multiple RGB-D camera fusion for assisting indoor stop-and-go 3D terrestrial laser scanning. Robotics 2014, 3, 247–280. [Google Scholar] [CrossRef]
  2. Salinas, C.; Fernández, R.; Montes, H.; Armada, M. A New Approach for Combining Time-of-Flight and RGB Cameras Based on Depth-Dependent Planar Projective Transformations. Sensors 2015, 15, 24615–24643. [Google Scholar] [CrossRef] [PubMed]
  3. Shim, H.; Adelsberger, R.; Kim, J.D.; Rhee, S.-M.; Rhee, T.; Sim, J.-Y.; Gross, M.; Kim, C. Time-of-flight sensor and color camera calibration for multi-view acquisition. Visual Comput. 2012, 28, 1139–1151. [Google Scholar] [CrossRef]
  4. Nair, R.; Ruhl, K.; Lenzen, F.; Meister, S.; Schäfer, H.; Garbe, C.S.; Eisemann, M.; Magnor, M.; Kondermann, D. A survey on time-of-flight stereo fusion. In Time-of-Flight and Depth Imaging. Sensors, Algorithms, and Applications; Springer: Berlin/Heidelberg, Germany, 2013; pp. 105–127. [Google Scholar]
  5. Zhang, L.; Dong, H.; Saddik, A.E. From 3D sensing to printing: A survey. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2016, 12, 27. [Google Scholar] [CrossRef]
  6. Beyer, H.A. Geometric and Radiometric Analysis of a CCD-Camera Based Photogrammetric Close-Range System; ETH Zurich: Zurich, Germany, 1992. [Google Scholar]
  7. Brown, D.C. Decentering distortion of lenses. Photogramm. Eng. Remote Sens. 1966, 3, 444–462. [Google Scholar]
  8. Fraser, C.S. Digital camera self-calibration. ISPRS J. Photogramm. Remote sens. 1997, 52, 149–159. [Google Scholar] [CrossRef]
  9. Remondino, F.; Fraser, C. Digital camera calibration methods: considerations and comparisons. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 266–272. [Google Scholar]
  10. Lichti, D.D.; Kim, C.; Jamtsho, S. An integrated bundle adjustment approach to range camera geometric self-calibration. ISPRS J. Photogramm. Remote Sens. 2010, 65, 360–368. [Google Scholar] [CrossRef]
  11. Lichti, D.D.; Kim, C. A comparison of three geometric self-calibration methods for range cameras. Remote Sens. 2011, 3, 1014–1028. [Google Scholar] [CrossRef]
  12. Lichti, D.D.; Qi, X. Range camera self-calibration with independent object space scale observations. J. Spat. Sci. 2012, 57, 247–257. [Google Scholar] [CrossRef]
  13. Jamtsho, S. Geometric Modelling of 3D Range Cameras and Their Application for Structural Deformation Measurement. Master’s Thesis, University of Calgary, Calgary, AB‎, Canada, 2010. [Google Scholar]
  14. Westfeld, P.; Mulsow, C.; Schulze, M. Photogrammetric calibration of range imaging sensors using intensity and range information simultaneously. Proc. Opt. 3-D Meas. Tech. IX 2009, 2, 1–10. [Google Scholar]
  15. Lindner, M.; Kolb, A. Lateral and depth calibration of PMD-distance sensors. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 14–16 December 2006; Springer: Berlin‎/‎Heidelberg, Germany, 2006; pp. 524–533. [Google Scholar]
  16. Zhu, J.; Wang, L.; Yang, R.; Davis, J.E. Reliability fusion of time-of-flight depth and stereo geometry for high quality depth maps. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 1400–1414. [Google Scholar] [PubMed]
  17. Mikhelson, I.V.; Lee, P.G.; Sahakian, A.V.; Wu, Y.; Katsaggelos, A.K. Automatic, fast, online calibration between depth and color cameras. J. Vis. Commun. Image Represent. 2014, 25, 218–226. [Google Scholar] [CrossRef]
  18. Van den Bergh, M.; Van Gool, L. Combining RGB and ToF cameras for real-time 3D hand gesture interaction. In Proceedings of the 2011 IEEE Workshop on Applications of Computer Vision (WACV), Kona, HI, USA, 5–7 January 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 66–72. [Google Scholar]
  19. Hansard, M.; Evangelidis, G.; Pelorson, Q.; Horaud, R. Cross-calibration of time-of-flight and colour cameras. Comput. Vis. Image Underst. 2015, 134, 105–115. [Google Scholar] [CrossRef]
  20. Herrera, D.; Kannala, J.; Heikkilä, J. Joint depth and color camera calibration with distortion correction. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2058–2064. [Google Scholar] [CrossRef] [PubMed]
  21. Wu, J.; Zhou, Y.; Yu, H.; Zhang, Z. Improved 3D depth image estimation algorithm for visual camera. In Proceedings of the 2nd International Congress on Image and Signal Processing, CISP′09, Tianjin, China, 17–19 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1–4. [Google Scholar]
  22. Jung, J.; Lee, J.-Y.; Jeong, Y.; Kweon, I.S. Time-of-flight sensor calibration for a color and depth camera pair. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1501–1513. [Google Scholar] [CrossRef] [PubMed]
  23. Vidas, S.; Moghadam, P.; Bosse, M. 3D thermal mapping of building interiors using an RGB-D and thermal camera. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 2311–2318. [Google Scholar]
  24. Lindner, M.; Kolb, A.; Hartmann, K. Data-fusion of PMD-based distance-information and high-resolution RGB-images. In Proceedings of the International Symposium on Signals, Circuits and Systems, ISSCS 2007, Iasi, Romania, 13–14 July 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1–4. [Google Scholar]
  25. Borrmann, D.; Afzal, H.; Elseberg, J.; Nüchter, A. Mutual calibration for 3D thermal mapping. IFAC Proc. Vol. 2012, 45, 605–610. [Google Scholar] [CrossRef]
  26. Munaro, M.; Basso, F.; Menegatti, E. OpenPTrack: Open source multi-camera calibration and people tracking for RGB-D camera networks. Robot. Autonomous Syst. 2016, 75, 525–538. [Google Scholar] [CrossRef]
  27. Mikhail, E.M.; Bethel, J.S.; McGlone, J.C. Introduction to Modern Photogrammetry; John Wiley and Sons, Inc.: New York, NY, USA, 2001. [Google Scholar]
  28. Owens, J.L.; Osteen, P.R.; Daniilidis, K. MSG-cal: Multi-sensor graph-based calibration. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 3660–3667. [Google Scholar]
  29. Shin, Y.-D.; Park, J.-H.; Bae, J.-H.; Baeg, M.-H. A study on reliability enhancement for laser and camera calibration. Int. J. Control Autom. Syst. 2012, 10, 109–116. [Google Scholar] [CrossRef]
  30. Wang, W.; Yamakawa, K.; Hiroi, K.; Kaji, K.; Kawaguchi, N. A Mobile System for 3D Indoor Mapping Using LiDAR and Panoramic Camera. Spec. Interest Group Tech. Rep. IPSJ 2015, 1, 1–7. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.